We use spark on windows 2008 R2 servers. We use one spark context which create 
one spark executor. We run spark master, slave, driver, executor on one single 
machine.

>From time to time, we found that the executor JAVA process was terminated. I 
>cannot fig out why it was terminated. Can anybody help me on how to find out 
>why the executor was terminated?

The spark slave log. It shows that it kill the executor process
2015-10-13 09:58:06,087 INFO  [sparkWorker-akka.actor.default-dispatcher-16] 
worker.Worker (Logging.scala:logInfo(59)) - Asked to kill executor 
app-20151009201453-0000/0

But why does it do that?

Here is the detailed logs from spark slave

2015-10-13 09:58:04,915 WARN  [sparkWorker-akka.actor.default-dispatcher-16] 
remote.ReliableDeliverySupervisor (Slf4jLogger.scala:apply$mcV$sp(71)) - 
Association with remote system 
[akka.tcp://sparkexecu...@qa1-cas01.pcc.lexisnexis.com:61234] has failed, 
address is now gated for [5000] ms. Reason is: [Disassociated].
2015-10-13 09:58:05,134 INFO  [sparkWorker-akka.actor.default-dispatcher-16] 
actor.LocalActorRef (Slf4jLogger.scala:apply$mcV$sp(74)) - Message 
[akka.remote.EndpointWriter$AckIdleCheckTimer$] from 
Actor[akka://sparkWorker/system/endpointManager/reliableEndpointWriter-akka.tcp%3A%2F%2FsparkExecutor%40QA1-CAS01.pcc.lexisnexis.com%3A61234-2/endpointWriter#-175670388]
 to 
Actor[akka://sparkWorker/system/endpointManager/reliableEndpointWriter-akka.tcp%3A%2F%2FsparkExecutor%40QA1-CAS01.pcc.lexisnexis.com%3A61234-2/endpointWriter#-175670388]
 was not delivered. [2] dead letters encountered. This logging can be turned 
off or adjusted with configuration settings 'akka.log-dead-letters' and 
'akka.log-dead-letters-during-shutdown'.
2015-10-13 09:58:05,134 INFO  [sparkWorker-akka.actor.default-dispatcher-16] 
actor.LocalActorRef (Slf4jLogger.scala:apply$mcV$sp(74)) - Message 
[akka.remote.transport.AssociationHandle$Disassociated] from 
Actor[akka://sparkWorker/deadLetters] to 
Actor[akka://sparkWorker/system/transports/akkaprotocolmanager.tcp0/akkaProtocol-tcp%3A%2F%2FsparkWorker%4010.196.116.184%3A61236-3#-1210125680]
 was not delivered. [3] dead letters encountered. This logging can be turned 
off or adjusted with configuration settings 'akka.log-dead-letters' and 
'akka.log-dead-letters-during-shutdown'.
2015-10-13 09:58:05,134 INFO  [sparkWorker-akka.actor.default-dispatcher-16] 
actor.LocalActorRef (Slf4jLogger.scala:apply$mcV$sp(74)) - Message 
[akka.remote.transport.ActorTransportAdapter$DisassociateUnderlying] from 
Actor[akka://sparkWorker/deadLetters] to 
Actor[akka://sparkWorker/system/transports/akkaprotocolmanager.tcp0/akkaProtocol-tcp%3A%2F%2FsparkWorker%4010.196.116.184%3A61236-3#-1210125680]
 was not delivered. [4] dead letters encountered. This logging can be turned 
off or adjusted with configuration settings 'akka.log-dead-letters' and 
'akka.log-dead-letters-during-shutdown'.
2015-10-13 09:58:06,087 INFO  [sparkWorker-akka.actor.default-dispatcher-16] 
worker.Worker (Logging.scala:logInfo(59)) - Asked to kill executor 
app-20151009201453-0000/0
2015-10-13 09:58:06,103 INFO  [ExecutorRunner for app-20151009201453-0000/0] 
worker.ExecutorRunner (Logging.scala:logInfo(59)) - Runner thread for executor 
app-20151009201453-0000/0 interrupted
2015-10-13 09:58:06,118 INFO  [ExecutorRunner for app-20151009201453-0000/0] 
worker.ExecutorRunner (Logging.scala:logInfo(59)) - Killing process!
2015-10-13 09:58:06,509 INFO  [sparkWorker-akka.actor.default-dispatcher-16] 
worker.Worker (Logging.scala:logInfo(59)) - Executor app-20151009201453-0000/0 
finished with state KILLED exitStatus 1
2015-10-13 09:58:06,509 INFO  [sparkWorker-akka.actor.default-dispatcher-16] 
worker.Worker (Logging.scala:logInfo(59)) - Cleaning up local directories for 
application app-20151009201453-0000

Thanks
Ningjun Wang

Reply via email to