[ https://issues.apache.org/jira/browse/SPARK-16106?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15342926#comment-15342926 ]
Apache Spark commented on SPARK-16106: -------------------------------------- User 'squito' has created a pull request for this issue: https://github.com/apache/spark/pull/13826 > TaskSchedulerImpl does not correctly handle new executors on existing hosts > --------------------------------------------------------------------------- > > Key: SPARK-16106 > URL: https://issues.apache.org/jira/browse/SPARK-16106 > Project: Spark > Issue Type: Bug > Components: Scheduler > Affects Versions: 2.0.0 > Reporter: Imran Rashid > Priority: Trivial > > The TaskSchedulerImpl updates the set of executors and hosts in each call to > {{resourceOffers}}. During this call, it also tracks whether there are any > new executors observed in {{newExecAvail}}: > {code} > executorIdToHost(o.executorId) = o.host > executorIdToTaskCount.getOrElseUpdate(o.executorId, 0) > if (!executorsByHost.contains(o.host)) { > executorsByHost(o.host) = new HashSet[String]() > executorAdded(o.executorId, o.host) > newExecAvail = true > } > {code} > However, this only detects when a new *host* is added, not when an additional > executor is added to an existing host (a relatively common event in dynamic > allocation). > The end result is that task locality and {{failedEpochs}} is not updated > correctly for new executors. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org