[ 
https://issues.apache.org/jira/browse/SPARK-12027?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hunter Kelly updated SPARK-12027:
---------------------------------
    Attachment: doc-emr-patch.txt

A patch to the docs to warn EMR users that they may need to set explicitly set 
spark.executor.instances to 0.

> Spark on YARN won't ever ask for more executors than there were containers at 
> time of context creation
> ------------------------------------------------------------------------------------------------------
>
>                 Key: SPARK-12027
>                 URL: https://issues.apache.org/jira/browse/SPARK-12027
>             Project: Spark
>          Issue Type: Bug
>          Components: Scheduler, YARN
>    Affects Versions: 1.5.2
>            Reporter: Hunter Kelly
>              Labels: scheduler, scheduling, yarn
>         Attachments: doc-emr-patch.txt
>
>
> Looking at YarnSchedulerBackend, it appears that totalExpectedExecutors is 
> only ever set at startup.
> Based on my experience of running on EMR (and a quick browse through the code 
> supports this), Spark will never ask for more executors than what this was 
> set to.
> This means that if I add more nodes to my YARN cluster, Spark will never pick 
> them up.  This is bad.  The whole point of using Spark on EMR/YARN is to be 
> able to add or remove nodes to the cluster and have Spark "Do The Right 
> Thing".



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to