[ https://issues.apache.org/jira/browse/SPARK-21876?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16154303#comment-16154303 ]
Imran Rashid commented on SPARK-21876: -------------------------------------- Thanks for reporting this [~julie.zhang]. I got a little stuck at first following how GC would trigger a message to the lost executor. Since the executor never ran any tasks, it wouldn't have any active blocks -- but I guess cleaning a broadcast always results in a message to all executors it thinks are alive. Does that sound correct? > Idling Executors that never handled any tasks are not cleared from > BlockManager after being removed > --------------------------------------------------------------------------------------------------- > > Key: SPARK-21876 > URL: https://issues.apache.org/jira/browse/SPARK-21876 > Project: Spark > Issue Type: Bug > Components: Scheduler, Spark Core > Affects Versions: 1.6.3, 2.2.0 > Reporter: Julie Zhang > > This happens when 'spark.dynamicAllocation.enabled' is set to be 'true'. We > use Yarn as our resource manager. > 1) Executor A is launched, but no task has been submitted to it; > 2) After 'spark.dynamicAllocation.executorIdleTimeout' seconds, executor A > will be removed. (ExecutorAllocationManager.scala schedule(): 294); > 3) The scheduler gets notified that executor A has been lost; (in our case, > YarnSchedulerBackend.scla: 209). > In the TaskschedulerImpl.scala method executorLost(executorId: String, > reason: ExecutorLossReason), the assumption in the None > case(TaskSchedulerImpl.scala: 548) that the executor has already been removed > is not always valid. As a result, the DAGScheduler and BlockManagerMaster are > never notified about the loss of executor A. > When GC eventually happens, the ContextCleaner will try to clean up > un-referenced objects. Because the executor A was not removed from the > blockManagerIdByExecutor map, BlockManagerMasterEndpoint will send out > requests to clean the references to the non-existent executor, producing a > lot of error message like this in the driver log: > ERROR [2017-08-08 00:00:23,596] > org.apache.spark.network.client.TransportClient: Failed to send RPC xxx to > xxx/xxx:x: java.nio.channels.ClosedChannelException -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org