This is happening because spark context shuts down without shutting down the ssc first. This was behavior till spark 1.4 ans was addressed in later releases. https://github.com/apache/spark/pull/6307
Which version of spark are you on? Thanks Deepak On Thu, May 12, 2016 at 12:14 PM, Rakesh H (Marketing Platform-BLR) < rakes...@flipkart.com> wrote: > Yes, it seems to be the case. > In this case executors should have continued logging values till 300, but > they are shutdown as soon as i do "yarn kill ......" > > On Thu, May 12, 2016 at 12:11 PM Deepak Sharma <deepakmc...@gmail.com> > wrote: > >> So in your case , the driver is shutting down gracefully , but the >> executors are not. >> IS this the problem? >> >> Thanks >> Deepak >> >> On Thu, May 12, 2016 at 11:49 AM, Rakesh H (Marketing Platform-BLR) < >> rakes...@flipkart.com> wrote: >> >>> Yes, it is set to true. >>> Log of driver : >>> >>> 16/05/12 10:18:29 ERROR yarn.ApplicationMaster: RECEIVED SIGNAL 15: SIGTERM >>> 16/05/12 10:18:29 INFO streaming.StreamingContext: Invoking >>> stop(stopGracefully=true) from shutdown hook >>> 16/05/12 10:18:29 INFO scheduler.JobGenerator: Stopping JobGenerator >>> gracefully >>> 16/05/12 10:18:29 INFO scheduler.JobGenerator: Waiting for all received >>> blocks to be consumed for job generation >>> 16/05/12 10:18:29 INFO scheduler.JobGenerator: Waited for all received >>> blocks to be consumed for job generation >>> >>> Log of executor: >>> 16/05/12 10:18:29 ERROR executor.CoarseGrainedExecutorBackend: Driver >>> xx.xx.xx.xx:xxxxx disassociated! Shutting down. >>> 16/05/12 10:18:29 WARN remote.ReliableDeliverySupervisor: Association with >>> remote system [xx.xx.xx.xx:xxxxx] has failed, address is now gated for >>> [5000] ms. Reason: [Disassociated] >>> 16/05/12 10:18:29 INFO storage.DiskBlockManager: Shutdown hook called >>> 16/05/12 10:18:29 INFO processors.StreamJobRunner$: VALUE -------------> >>> 204 //This is value i am logging >>> 16/05/12 10:18:29 INFO util.ShutdownHookManager: Shutdown hook called >>> 16/05/12 10:18:29 INFO processors.StreamJobRunner$: VALUE -------------> 205 >>> 16/05/12 10:18:29 INFO processors.StreamJobRunner$: VALUE -------------> 206 >>> >>> >>> >>> >>> >>> >>> On Thu, May 12, 2016 at 11:45 AM Deepak Sharma <deepakmc...@gmail.com> >>> wrote: >>> >>>> Hi Rakesh >>>> Did you tried setting *spark.streaming.stopGracefullyOnShutdown to >>>> true *for your spark configuration instance? >>>> If not try this , and let us know if this helps. >>>> >>>> Thanks >>>> Deepak >>>> >>>> On Thu, May 12, 2016 at 11:42 AM, Rakesh H (Marketing Platform-BLR) < >>>> rakes...@flipkart.com> wrote: >>>> >>>>> Issue i am having is similar to the one mentioned here : >>>>> >>>>> http://stackoverflow.com/questions/36911442/how-to-stop-gracefully-a-spark-streaming-application-on-yarn >>>>> >>>>> I am creating a rdd from sequence of 1 to 300 and creating streaming >>>>> RDD out of it. >>>>> >>>>> val rdd = ssc.sparkContext.parallelize(1 to 300) >>>>> val dstream = new ConstantInputDStream(ssc, rdd) >>>>> dstream.foreachRDD{ rdd => >>>>> rdd.foreach{ x => >>>>> log(x) >>>>> Thread.sleep(50) >>>>> } >>>>> } >>>>> >>>>> >>>>> When i kill this job, i expect elements 1 to 300 to be logged before >>>>> shutting down. It is indeed the case when i run it locally. It wait for >>>>> the >>>>> job to finish before shutting down. >>>>> >>>>> But when i launch the job in custer with "yarn-cluster" mode, it >>>>> abruptly shuts down. >>>>> Executor prints following log >>>>> >>>>> ERROR executor.CoarseGrainedExecutorBackend: >>>>> Driver xx.xx.xx.xxx:yyyyy disassociated! Shutting down. >>>>> >>>>> and then it shuts down. It is not a graceful shutdown. >>>>> >>>>> Anybody knows how to do it in yarn ? >>>>> >>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> Thanks >>>> Deepak >>>> www.bigdatabig.com >>>> www.keosha.net >>>> >>> >> >> >> -- >> Thanks >> Deepak >> www.bigdatabig.com >> www.keosha.net >> > -- Thanks Deepak www.bigdatabig.com www.keosha.net