[
https://issues.apache.org/jira/browse/SPARK-18421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15661520#comment-15661520
]
Sean Owen commented on SPARK-18421:
-----------------------------------
Spark doesn't manage storage at all. I don't think this could be in scope
therefore especially because it could only apply to cloud and would be cloud
specific.
> Dynamic disk allocation
> -----------------------
>
> Key: SPARK-18421
> URL: https://issues.apache.org/jira/browse/SPARK-18421
> Project: Spark
> Issue Type: New Feature
> Components: Spark Core
> Affects Versions: 2.0.1
> Reporter: Aniket Bhatnagar
> Priority: Minor
>
> Dynamic allocation feature allows you to add executors and scale computation
> power. This is great, however, I feel like we also need a way to dynamically
> scale storage. Currently, if the disk is not able to hold the spilled/shuffle
> data, the job is aborted (in yarn, the node manager kills the container)
> causing frustration and loss of time. In deployments like AWS EMR, it is
> possible to run an agent that add disks on the fly if it sees that the disks
> are running out of space and it would be great if Spark could immediately
> start using the added disks just as it does when new executors are added.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]