[ 
https://issues.apache.org/jira/browse/SPARK-16106?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15342926#comment-15342926
 ] 

Apache Spark commented on SPARK-16106:
--------------------------------------

User 'squito' has created a pull request for this issue:
https://github.com/apache/spark/pull/13826

> TaskSchedulerImpl does not correctly handle new executors on existing hosts
> ---------------------------------------------------------------------------
>
>                 Key: SPARK-16106
>                 URL: https://issues.apache.org/jira/browse/SPARK-16106
>             Project: Spark
>          Issue Type: Bug
>          Components: Scheduler
>    Affects Versions: 2.0.0
>            Reporter: Imran Rashid
>            Priority: Trivial
>
> The TaskSchedulerImpl updates the set of executors and hosts in each call to 
> {{resourceOffers}}.  During this call, it also tracks whether there are any 
> new executors observed in {{newExecAvail}}:
> {code}
>       executorIdToHost(o.executorId) = o.host
>       executorIdToTaskCount.getOrElseUpdate(o.executorId, 0)
>       if (!executorsByHost.contains(o.host)) {
>         executorsByHost(o.host) = new HashSet[String]()
>         executorAdded(o.executorId, o.host)
>         newExecAvail = true
>       }
> {code}
> However, this only detects when a new *host* is added, not when an additional 
> executor is added to an existing host (a relatively common event in dynamic 
> allocation).
> The end result is that task locality and {{failedEpochs}} is not updated 
> correctly for new executors.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to