Hello. I am running Spark 2.1, Scala 2.11. We're running several Spark streaming jobs. In some cases we restart these jobs on an occasional basis. We have code that looks like the following:
logger.info("Starting the streaming context!") ssc.start() logger.info("Waiting for termination!") Option(config.getInt(Parameters.RestartMinutes)).getOrElse(0) match { case restartMinutes: Int if restartMinutes > 0 => logger.info(s"Waiting for ${restartMinutes} before terminating job") ssc.awaitTerminationOrTimeout(restartMinutes * DateUtils.millisecondsPerMinute) case _ => ssc.awaitTermination() } logger.info("Calling 'stop'") ssc.stop(stopSparkContext = true, stopGracefully = true) In several cases we've observed jobs where we've called 'stop' not stopping. I went and wrote a simple job that reads from Kafka and does nothing (prints a count of data). After several minutes it simply calls 'ssc.stop(true, true)'. In some cases this will stop the context. In others it will not stop the context. If we call 'stop' several times over an interval one of them eventually succeeds. It looks like this is a bug. I looked in Jira and did not see an open issue. Is this a known problem? If not I'll open a bug. Regards, Bryan Jeffrey