[ 
https://issues.apache.org/jira/browse/SPARK-2892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14262078#comment-14262078
 ] 

Ilayaperumal Gopinathan edited comment on SPARK-2892 at 12/31/14 10:29 AM:
---------------------------------------------------------------------------

[~joshrosen] yeah, serialization could be the real issue here.
 But after trying the suggested change (in my environment where I call 
streamingContext stop). But I see the active streaming job is only cancelled 
upon spark context shutdown), I see the below:

ERROR sparkDriver-akka.actor.default-dispatcher-19 scheduler.TaskSchedulerImpl 
- Lost executor 0 on 192.168.2.6: remote Akka client disassociated
WARN sparkDriver-akka.actor.default-dispatcher-20 
remote.ReliableDeliverySupervisor - Association with remote system 
[akka.tcp://sparkExecutor@192.168.2.6:54259] has failed, address is now gated 
for [5000] ms. Reason is: [Disassociated].
WARN sparkDriver-akka.actor.default-dispatcher-19 scheduler.TaskSetManager - 
Lost task 0.0 in stage 8.0 (TID 76, 192.168.2.6): ExecutorLostFailure (executor 
0 lost)
 ERROR sparkDriver-akka.actor.default-dispatcher-19 
cluster.SparkDeploySchedulerBackend - Asked to remove non-existent executor 0
ERROR sparkDriver-akka.actor.default-dispatcher-19 
cluster.SparkDeploySchedulerBackend - Asked to remove non-existent executor 0
WARN main-EventThread scheduler.ReceiverTracker - All of the receivers have not 
deregistered, Map(0 -> 
ReceiverInfo(0,MessageBusReceiver-0,Actor[akka.tcp://sparkExecutor@192.168.2.6:54279/user/Receiver-0-1420019455242#1553271505],true,192.168.2.6,,))
Exception in thread "Thread-42" org.apache.spark.SparkException: Job cancelled 
because SparkContext was shut down
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:702)
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:701)
        at scala.collection.mutable.HashSet.foreach(HashSet.scala:79)
        at 
org.apache.spark.scheduler.DAGScheduler.cleanUpAfterSchedulerStop(DAGScheduler.scala:701)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessActor.postStop(DAGScheduler.scala:1428)
        at akka.actor.Actor$class.aroundPostStop(Actor.scala:475)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessActor.aroundPostStop(DAGScheduler.scala:1375)
        at 
akka.actor.dungeon.FaultHandling$class.akka$actor$dungeon$FaultHandling$$finishTerminate(FaultHandling.scala:210)
        at 
akka.actor.dungeon.FaultHandling$class.terminate(FaultHandling.scala:172)
        at akka.actor.ActorCell.terminate(ActorCell.scala:369)
        at akka.actor.ActorCell.invokeAll$1(ActorCell.scala:462)
        at akka.actor.ActorCell.systemInvoke(ActorCell.scala:478)
        at akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:263)
        at akka.dispatch.Mailbox.run(Mailbox.scala:219)
        at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:393)
        at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
        at 
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
        at 
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
        at 
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)

Let me know what do you think?


was (Author: ilayaperumalg):
[~joshrosen] yeah, serialization could be the real issue here.
 But after trying the suggested change (in my environment where I call 
streamingContext stop). But I see the active streaming job is only cancelled 
upon spark context shutdown), I see the below:

ERROR akka.actor.default-dispatcher-19 scheduler.TaskSchedulerImpl - Lost 
executor 0 on 192.168.2.6: remote Akka client disassociated
WARN akka.actor.default-dispatcher-20 remote.ReliableDeliverySupervisor - 
Association with remote system [akka.tcp://sparkExecutor@192.168.2.6:54259] has 
failed, address is now gated for [5000] ms. Reason is: [Disassociated].
WARN akka.actor.default-dispatcher-19 scheduler.TaskSetManager - Lost task 0.0 
in stage 8.0 (TID 76, 192.168.2.6): ExecutorLostFailure (executor 0 lost)
 ERROR akka.actor.default-dispatcher-19 cluster.SparkDeploySchedulerBackend - 
Asked to remove non-existent executor 0
ERROR akka.actor.default-dispatcher-19 cluster.SparkDeploySchedulerBackend - 
Asked to remove non-existent executor 0
WARN main-EventThread scheduler.ReceiverTracker - All of the receivers have not 
deregistered, Map(0 -> 
ReceiverInfo(0,MessageBusReceiver-0,Actor[akka.tcp://sparkExecutor@192.168.2.6:54279/user/Receiver-0-1420019455242#1553271505],true,192.168.2.6,,))
Exception in thread "Thread-42" org.apache.spark.SparkException: Job cancelled 
because SparkContext was shut down
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:702)
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:701)
        at scala.collection.mutable.HashSet.foreach(HashSet.scala:79)
        at 
org.apache.spark.scheduler.DAGScheduler.cleanUpAfterSchedulerStop(DAGScheduler.scala:701)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessActor.postStop(DAGScheduler.scala:1428)
        at akka.actor.Actor$class.aroundPostStop(Actor.scala:475)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessActor.aroundPostStop(DAGScheduler.scala:1375)
        at 
akka.actor.dungeon.FaultHandling$class.akka$actor$dungeon$FaultHandling$$finishTerminate(FaultHandling.scala:210)
        at 
akka.actor.dungeon.FaultHandling$class.terminate(FaultHandling.scala:172)
        at akka.actor.ActorCell.terminate(ActorCell.scala:369)
        at akka.actor.ActorCell.invokeAll$1(ActorCell.scala:462)
        at akka.actor.ActorCell.systemInvoke(ActorCell.scala:478)
        at akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:263)
        at akka.dispatch.Mailbox.run(Mailbox.scala:219)
        at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:393)
        at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
        at 
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
        at 
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
        at 
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)

Let me know what do you think?

> Socket Receiver does not stop when streaming context is stopped
> ---------------------------------------------------------------
>
>                 Key: SPARK-2892
>                 URL: https://issues.apache.org/jira/browse/SPARK-2892
>             Project: Spark
>          Issue Type: Bug
>          Components: Streaming
>    Affects Versions: 1.0.2
>            Reporter: Tathagata Das
>            Assignee: Tathagata Das
>            Priority: Critical
>
> Running NetworkWordCount with
> {quote}      
> ssc.start(); Thread.sleep(10000); ssc.stop(stopSparkContext = false); 
> Thread.sleep(60000)
> {quote}
> gives the following error
> {quote}
> 14/08/06 18:37:13 INFO TaskSetManager: Finished task 0.0 in stage 0.0 (TID 0) 
> in 10047 ms on localhost (1/1)
> 14/08/06 18:37:13 INFO DAGScheduler: Stage 0 (runJob at 
> ReceiverTracker.scala:275) finished in 10.056 s
> 14/08/06 18:37:13 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose tasks 
> have all completed, from pool
> 14/08/06 18:37:13 INFO SparkContext: Job finished: runJob at 
> ReceiverTracker.scala:275, took 10.179263 s
> 14/08/06 18:37:13 INFO ReceiverTracker: All of the receivers have been 
> terminated
> 14/08/06 18:37:13 WARN ReceiverTracker: All of the receivers have not 
> deregistered, Map(0 -> 
> ReceiverInfo(0,SocketReceiver-0,null,false,localhost,Stopped by driver,))
> 14/08/06 18:37:13 INFO ReceiverTracker: ReceiverTracker stopped
> 14/08/06 18:37:13 INFO JobGenerator: Stopping JobGenerator immediately
> 14/08/06 18:37:13 INFO RecurringTimer: Stopped timer for JobGenerator after 
> time 1407375433000
> 14/08/06 18:37:13 INFO JobGenerator: Stopped JobGenerator
> 14/08/06 18:37:13 INFO JobScheduler: Stopped JobScheduler
> 14/08/06 18:37:13 INFO StreamingContext: StreamingContext stopped successfully
> 14/08/06 18:37:43 INFO SocketReceiver: Stopped receiving
> 14/08/06 18:37:43 INFO SocketReceiver: Closed socket to localhost:9999
> {quote}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to