Yes, it seems to be the case.
In this case executors should have continued logging values till 300, but
they are shutdown as soon as i do "yarn kill ......"

On Thu, May 12, 2016 at 12:11 PM Deepak Sharma <deepakmc...@gmail.com>
wrote:

> So in your case , the driver is shutting down gracefully , but the
> executors are not.
> IS this the problem?
>
> Thanks
> Deepak
>
> On Thu, May 12, 2016 at 11:49 AM, Rakesh H (Marketing Platform-BLR) <
> rakes...@flipkart.com> wrote:
>
>> Yes, it is set to true.
>> Log of driver :
>>
>> 16/05/12 10:18:29 ERROR yarn.ApplicationMaster: RECEIVED SIGNAL 15: SIGTERM
>> 16/05/12 10:18:29 INFO streaming.StreamingContext: Invoking 
>> stop(stopGracefully=true) from shutdown hook
>> 16/05/12 10:18:29 INFO scheduler.JobGenerator: Stopping JobGenerator 
>> gracefully
>> 16/05/12 10:18:29 INFO scheduler.JobGenerator: Waiting for all received 
>> blocks to be consumed for job generation
>> 16/05/12 10:18:29 INFO scheduler.JobGenerator: Waited for all received 
>> blocks to be consumed for job generation
>>
>> Log of executor:
>> 16/05/12 10:18:29 ERROR executor.CoarseGrainedExecutorBackend: Driver 
>> xx.xx.xx.xx:xxxxx disassociated! Shutting down.
>> 16/05/12 10:18:29 WARN remote.ReliableDeliverySupervisor: Association with 
>> remote system [xx.xx.xx.xx:xxxxx] has failed, address is now gated for 
>> [5000] ms. Reason: [Disassociated]
>> 16/05/12 10:18:29 INFO storage.DiskBlockManager: Shutdown hook called
>> 16/05/12 10:18:29 INFO processors.StreamJobRunner$: VALUE -------------> 204 
>> //This is value i am logging
>> 16/05/12 10:18:29 INFO util.ShutdownHookManager: Shutdown hook called
>> 16/05/12 10:18:29 INFO processors.StreamJobRunner$: VALUE -------------> 205
>> 16/05/12 10:18:29 INFO processors.StreamJobRunner$: VALUE -------------> 206
>>
>>
>>
>>
>>
>>
>> On Thu, May 12, 2016 at 11:45 AM Deepak Sharma <deepakmc...@gmail.com>
>> wrote:
>>
>>> Hi Rakesh
>>> Did you tried setting *spark.streaming.stopGracefullyOnShutdown to true
>>> *for your spark configuration instance?
>>> If not try this , and let us know if this helps.
>>>
>>> Thanks
>>> Deepak
>>>
>>> On Thu, May 12, 2016 at 11:42 AM, Rakesh H (Marketing Platform-BLR) <
>>> rakes...@flipkart.com> wrote:
>>>
>>>> Issue i am having is similar to the one mentioned here :
>>>>
>>>> http://stackoverflow.com/questions/36911442/how-to-stop-gracefully-a-spark-streaming-application-on-yarn
>>>>
>>>> I am creating a rdd from sequence of 1 to 300 and creating streaming
>>>> RDD out of it.
>>>>
>>>> val rdd = ssc.sparkContext.parallelize(1 to 300)
>>>> val dstream = new ConstantInputDStream(ssc, rdd)
>>>> dstream.foreachRDD{ rdd =>
>>>>   rdd.foreach{ x =>
>>>>     log(x)
>>>>     Thread.sleep(50)
>>>>   }
>>>> }
>>>>
>>>>
>>>> When i kill this job, i expect elements 1 to 300 to be logged before
>>>> shutting down. It is indeed the case when i run it locally. It wait for the
>>>> job to finish before shutting down.
>>>>
>>>> But when i launch the job in custer with "yarn-cluster" mode, it
>>>> abruptly shuts down.
>>>> Executor prints following log
>>>>
>>>> ERROR executor.CoarseGrainedExecutorBackend:
>>>> Driver xx.xx.xx.xxx:yyyyy disassociated! Shutting down.
>>>>
>>>>  and then it shuts down. It is not a graceful shutdown.
>>>>
>>>> Anybody knows how to do it in yarn ?
>>>>
>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> Thanks
>>> Deepak
>>> www.bigdatabig.com
>>> www.keosha.net
>>>
>>
>
>
> --
> Thanks
> Deepak
> www.bigdatabig.com
> www.keosha.net
>

Reply via email to