+1 Regards, Dhaval Modi dhavalmod...@gmail.com
On 8 November 2017 at 00:06, Bryan Jeffrey <bryan.jeff...@gmail.com> wrote: > Hello. > > I am running Spark 2.1, Scala 2.11. We're running several Spark streaming > jobs. In some cases we restart these jobs on an occasional basis. We have > code that looks like the following: > > logger.info("Starting the streaming context!") > ssc.start() > logger.info("Waiting for termination!") > Option(config.getInt(Parameters.RestartMinutes)).getOrElse(0) match { > case restartMinutes: Int if restartMinutes > 0 => > logger.info(s"Waiting for ${restartMinutes} before terminating job") > ssc.awaitTerminationOrTimeout(restartMinutes * > DateUtils.millisecondsPerMinute) > case _ => ssc.awaitTermination() > } > logger.info("Calling 'stop'") > ssc.stop(stopSparkContext = true, stopGracefully = true) > > > In several cases we've observed jobs where we've called 'stop' not > stopping. I went and wrote a simple job that reads from Kafka and does > nothing (prints a count of data). After several minutes it simply calls > 'ssc.stop(true, true)'. In some cases this will stop the context. In > others it will not stop the context. If we call 'stop' several times over > an interval one of them eventually succeeds. > > It looks like this is a bug. I looked in Jira and did not see an open > issue. Is this a known problem? If not I'll open a bug. > > Regards, > > Bryan Jeffrey > > >