[ 
https://issues.apache.org/jira/browse/SPARK-18421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15661520#comment-15661520
 ] 

Sean Owen commented on SPARK-18421:
-----------------------------------

Spark doesn't manage storage at all. I don't think this could be in scope 
therefore especially because it could only apply to cloud and would be cloud 
specific.

> Dynamic disk allocation
> -----------------------
>
>                 Key: SPARK-18421
>                 URL: https://issues.apache.org/jira/browse/SPARK-18421
>             Project: Spark
>          Issue Type: New Feature
>          Components: Spark Core
>    Affects Versions: 2.0.1
>            Reporter: Aniket Bhatnagar
>            Priority: Minor
>
> Dynamic allocation feature allows you to add executors and scale computation 
> power. This is great, however, I feel like we also need a way to dynamically 
> scale storage. Currently, if the disk is not able to hold the spilled/shuffle 
> data, the job is aborted (in yarn, the node manager kills the container) 
> causing frustration and loss of time. In deployments like AWS EMR, it is 
> possible to run an agent that add disks on the fly if it sees that the disks 
> are running out of space and it would be great if Spark could immediately 
> start using the added disks just as it does when new executors are added.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to