Have you used awaitTermination() on your ssc ? --> Yes, i have used that.
Also try setting the deployment mode to yarn-client. --> Is this not
supported on yarn-cluster mode?  I am trying to find root cause for
yarn-cluster mode.
Have you tested graceful shutdown on yarn-cluster mode?

On Fri, May 13, 2016 at 11:54 AM Deepak Sharma <deepakmc...@gmail.com>
wrote:

> Rakesh
> Have you used awaitTermination() on your ssc ?
> If not , dd this and see if it changes the behavior.
> I am guessing this issue may be related to yarn deployment mode.
> Also try setting the deployment mode to yarn-client.
>
> Thanks
> Deepak
>
>
> On Fri, May 13, 2016 at 10:17 AM, Rakesh H (Marketing Platform-BLR) <
> rakes...@flipkart.com> wrote:
>
>> Ping!!
>> Has anybody tested graceful shutdown of a spark streaming in yarn-cluster
>> mode?It looks like a defect to me.
>>
>>
>> On Thu, May 12, 2016 at 12:53 PM Rakesh H (Marketing Platform-BLR) <
>> rakes...@flipkart.com> wrote:
>>
>>> We are on spark 1.5.1
>>> Above change was to add a shutdown hook.
>>> I am not adding shutdown hook in code, so inbuilt shutdown hook is being
>>> called.
>>> Driver signals that it is going to to graceful shutdown, but executor
>>> sees that Driver is dead and it shuts down abruptly.
>>> Could this issue be related to yarn? I see correct behavior locally. I
>>> did "yarn kill ...." to kill the job.
>>>
>>>
>>> On Thu, May 12, 2016 at 12:28 PM Deepak Sharma <deepakmc...@gmail.com>
>>> wrote:
>>>
>>>> This is happening because spark context shuts down without shutting
>>>> down the ssc first.
>>>> This was behavior till spark 1.4 ans was addressed in later releases.
>>>> https://github.com/apache/spark/pull/6307
>>>>
>>>> Which version of spark are you on?
>>>>
>>>> Thanks
>>>> Deepak
>>>>
>>>> On Thu, May 12, 2016 at 12:14 PM, Rakesh H (Marketing Platform-BLR) <
>>>> rakes...@flipkart.com> wrote:
>>>>
>>>>> Yes, it seems to be the case.
>>>>> In this case executors should have continued logging values till 300,
>>>>> but they are shutdown as soon as i do "yarn kill ......"
>>>>>
>>>>> On Thu, May 12, 2016 at 12:11 PM Deepak Sharma <deepakmc...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> So in your case , the driver is shutting down gracefully , but the
>>>>>> executors are not.
>>>>>> IS this the problem?
>>>>>>
>>>>>> Thanks
>>>>>> Deepak
>>>>>>
>>>>>> On Thu, May 12, 2016 at 11:49 AM, Rakesh H (Marketing Platform-BLR) <
>>>>>> rakes...@flipkart.com> wrote:
>>>>>>
>>>>>>> Yes, it is set to true.
>>>>>>> Log of driver :
>>>>>>>
>>>>>>> 16/05/12 10:18:29 ERROR yarn.ApplicationMaster: RECEIVED SIGNAL 15: 
>>>>>>> SIGTERM
>>>>>>> 16/05/12 10:18:29 INFO streaming.StreamingContext: Invoking 
>>>>>>> stop(stopGracefully=true) from shutdown hook
>>>>>>> 16/05/12 10:18:29 INFO scheduler.JobGenerator: Stopping JobGenerator 
>>>>>>> gracefully
>>>>>>> 16/05/12 10:18:29 INFO scheduler.JobGenerator: Waiting for all received 
>>>>>>> blocks to be consumed for job generation
>>>>>>> 16/05/12 10:18:29 INFO scheduler.JobGenerator: Waited for all received 
>>>>>>> blocks to be consumed for job generation
>>>>>>>
>>>>>>> Log of executor:
>>>>>>> 16/05/12 10:18:29 ERROR executor.CoarseGrainedExecutorBackend: Driver 
>>>>>>> xx.xx.xx.xx:xxxxx disassociated! Shutting down.
>>>>>>> 16/05/12 10:18:29 WARN remote.ReliableDeliverySupervisor: Association 
>>>>>>> with remote system [xx.xx.xx.xx:xxxxx] has failed, address is now gated 
>>>>>>> for [5000] ms. Reason: [Disassociated]
>>>>>>> 16/05/12 10:18:29 INFO storage.DiskBlockManager: Shutdown hook called
>>>>>>> 16/05/12 10:18:29 INFO processors.StreamJobRunner$: VALUE 
>>>>>>> -------------> 204 //This is value i am logging
>>>>>>> 16/05/12 10:18:29 INFO util.ShutdownHookManager: Shutdown hook called
>>>>>>> 16/05/12 10:18:29 INFO processors.StreamJobRunner$: VALUE 
>>>>>>> -------------> 205
>>>>>>> 16/05/12 10:18:29 INFO processors.StreamJobRunner$: VALUE 
>>>>>>> -------------> 206
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Thu, May 12, 2016 at 11:45 AM Deepak Sharma <
>>>>>>> deepakmc...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Hi Rakesh
>>>>>>>> Did you tried setting *spark.streaming.stopGracefullyOnShutdown to
>>>>>>>> true *for your spark configuration instance?
>>>>>>>> If not try this , and let us know if this helps.
>>>>>>>>
>>>>>>>> Thanks
>>>>>>>> Deepak
>>>>>>>>
>>>>>>>> On Thu, May 12, 2016 at 11:42 AM, Rakesh H (Marketing Platform-BLR)
>>>>>>>> <rakes...@flipkart.com> wrote:
>>>>>>>>
>>>>>>>>> Issue i am having is similar to the one mentioned here :
>>>>>>>>>
>>>>>>>>> http://stackoverflow.com/questions/36911442/how-to-stop-gracefully-a-spark-streaming-application-on-yarn
>>>>>>>>>
>>>>>>>>> I am creating a rdd from sequence of 1 to 300 and creating
>>>>>>>>> streaming RDD out of it.
>>>>>>>>>
>>>>>>>>> val rdd = ssc.sparkContext.parallelize(1 to 300)
>>>>>>>>> val dstream = new ConstantInputDStream(ssc, rdd)
>>>>>>>>> dstream.foreachRDD{ rdd =>
>>>>>>>>>   rdd.foreach{ x =>
>>>>>>>>>     log(x)
>>>>>>>>>     Thread.sleep(50)
>>>>>>>>>   }
>>>>>>>>> }
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> When i kill this job, i expect elements 1 to 300 to be logged
>>>>>>>>> before shutting down. It is indeed the case when i run it locally. It 
>>>>>>>>> wait
>>>>>>>>> for the job to finish before shutting down.
>>>>>>>>>
>>>>>>>>> But when i launch the job in custer with "yarn-cluster" mode, it
>>>>>>>>> abruptly shuts down.
>>>>>>>>> Executor prints following log
>>>>>>>>>
>>>>>>>>> ERROR executor.CoarseGrainedExecutorBackend:
>>>>>>>>> Driver xx.xx.xx.xxx:yyyyy disassociated! Shutting down.
>>>>>>>>>
>>>>>>>>>  and then it shuts down. It is not a graceful shutdown.
>>>>>>>>>
>>>>>>>>> Anybody knows how to do it in yarn ?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Thanks
>>>>>>>> Deepak
>>>>>>>> www.bigdatabig.com
>>>>>>>> www.keosha.net
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Thanks
>>>>>> Deepak
>>>>>> www.bigdatabig.com
>>>>>> www.keosha.net
>>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Thanks
>>>> Deepak
>>>> www.bigdatabig.com
>>>> www.keosha.net
>>>>
>>>
>
>
> --
> Thanks
> Deepak
> www.bigdatabig.com
> www.keosha.net
>

Reply via email to