[
https://issues.apache.org/jira/browse/SPARK-12027?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Hunter Kelly updated SPARK-12027:
---------------------------------
Attachment: doc-emr-patch.txt
A patch to the docs to warn EMR users that they may need to set explicitly set
spark.executor.instances to 0.
> Spark on YARN won't ever ask for more executors than there were containers at
> time of context creation
> ------------------------------------------------------------------------------------------------------
>
> Key: SPARK-12027
> URL: https://issues.apache.org/jira/browse/SPARK-12027
> Project: Spark
> Issue Type: Bug
> Components: Scheduler, YARN
> Affects Versions: 1.5.2
> Reporter: Hunter Kelly
> Labels: scheduler, scheduling, yarn
> Attachments: doc-emr-patch.txt
>
>
> Looking at YarnSchedulerBackend, it appears that totalExpectedExecutors is
> only ever set at startup.
> Based on my experience of running on EMR (and a quick browse through the code
> supports this), Spark will never ask for more executors than what this was
> set to.
> This means that if I add more nodes to my YARN cluster, Spark will never pick
> them up. This is bad. The whole point of using Spark on EMR/YARN is to be
> able to add or remove nodes to the cluster and have Spark "Do The Right
> Thing".
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]