StreamingContext.stop(...) if using scala JavaStreamingContext.stop(...) if using Java
ᐧ On Wed, Aug 3, 2016 at 9:14 PM, Tony Lane <tonylane....@gmail.com> wrote: > SparkSession exposes stop() method > > On Wed, Aug 3, 2016 at 8:53 AM, Pradeep <pradeep.mi...@mail.com> wrote: > >> Thanks Park. I am doing the same. Was trying to understand if there are >> other ways. >> >> Thanks, >> Pradeep >> >> > On Aug 2, 2016, at 10:25 PM, Park Kyeong Hee <kh1979.p...@samsung.com> >> wrote: >> > >> > So sorry. Your name was Pradeep !! >> > >> > -----Original Message----- >> > From: Park Kyeong Hee [mailto:kh1979.p...@samsung.com] >> > Sent: Wednesday, August 03, 2016 11:24 AM >> > To: 'Pradeep'; 'user@spark.apache.org' >> > Subject: RE: Stop Spark Streaming Jobs >> > >> > Hi. Paradeep >> > >> > >> > Did you mean, how to kill the job? >> > If yes, you should kill the driver and follow next. >> > >> > on yarn-client >> > 1. find pid - "ps -es | grep <your_jobs_main_class>" >> > 2. kill it - "kill -9 <pid>" >> > 3. check executors were down - "yarn application -list" >> > >> > on yarn-cluster >> > 1. find driver's application ID - "yarn application -list" >> > 2. stop it - "yarn application -kill <app_ID>" >> > 3. check driver and executors were down - "yarn application -list" >> > >> > >> > Thanks. >> > >> > -----Original Message----- >> > From: Pradeep [mailto:pradeep.mi...@mail.com] >> > Sent: Wednesday, August 03, 2016 10:48 AM >> > To: user@spark.apache.org >> > Subject: Stop Spark Streaming Jobs >> > >> > Hi All, >> > >> > My streaming job reads data from Kafka. The job is triggered and pushed >> to >> > background with nohup. >> > >> > What are the recommended ways to stop job either on yarn-client or >> cluster >> > mode. >> > >> > Thanks, >> > Pradeep >> > >> > --------------------------------------------------------------------- >> > To unsubscribe e-mail: user-unsubscr...@spark.apache.org >> > >> > >> > >> > >> > --------------------------------------------------------------------- >> > To unsubscribe e-mail: user-unsubscr...@spark.apache.org >> > >> >> >> --------------------------------------------------------------------- >> To unsubscribe e-mail: user-unsubscr...@spark.apache.org >> >> > -- * Regards* * Sandeep Nemuri*