Hi Rakesh Did you tried setting *spark.streaming.stopGracefullyOnShutdown to true *for your spark configuration instance? If not try this , and let us know if this helps.
Thanks Deepak On Thu, May 12, 2016 at 11:42 AM, Rakesh H (Marketing Platform-BLR) < rakes...@flipkart.com> wrote: > Issue i am having is similar to the one mentioned here : > > http://stackoverflow.com/questions/36911442/how-to-stop-gracefully-a-spark-streaming-application-on-yarn > > I am creating a rdd from sequence of 1 to 300 and creating streaming RDD > out of it. > > val rdd = ssc.sparkContext.parallelize(1 to 300) > val dstream = new ConstantInputDStream(ssc, rdd) > dstream.foreachRDD{ rdd => > rdd.foreach{ x => > log(x) > Thread.sleep(50) > } > } > > > When i kill this job, i expect elements 1 to 300 to be logged before > shutting down. It is indeed the case when i run it locally. It wait for the > job to finish before shutting down. > > But when i launch the job in custer with "yarn-cluster" mode, it abruptly > shuts down. > Executor prints following log > > ERROR executor.CoarseGrainedExecutorBackend: > Driver xx.xx.xx.xxx:yyyyy disassociated! Shutting down. > > and then it shuts down. It is not a graceful shutdown. > > Anybody knows how to do it in yarn ? > > > > -- Thanks Deepak www.bigdatabig.com www.keosha.net