If we want to stop the  application after fix-time period , how it will
work . (How to give the duration in logic , in my case  sleep(t.s.)  is not
working .)  So i used to kill coarseGrained job at each slave by script
.Please suggest something .

On Thu, Jul 30, 2015 at 5:14 AM, Tathagata Das <t...@databricks.com> wrote:

> StreamingContext.stop(stopGracefully = true) stops the streaming context
> gracefully.
> Then you can safely terminate the Spark cluster. They are two different
> steps and needs to be done separately ensuring that the driver process has
> been completely terminated before the Spark cluster is the terminated.
>
> On Wed, Jul 29, 2015 at 6:43 AM, Michal Čizmazia <mici...@gmail.com>
> wrote:
>
>> How to initiate graceful shutdown from outside of the Spark Streaming
>> driver process? Both for the local and cluster mode of Spark Standalone as
>> well as EMR.
>>
>> Does sbin/stop-all.sh stop the context gracefully? How is it done? Is
>> there a signal sent to the driver process?
>>
>> For EMR, is there a way how to terminate an EMR cluster with Spark
>> Streaming graceful shutdown?
>>
>> Thanks!
>>
>>
>>
>


-- 
Thanks & Regards,
Anshu Shukla

Reply via email to