[ https://issues.apache.org/jira/browse/SPARK-31969?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
ThimmeGowda updated SPARK-31969: -------------------------------- Component/s: Web UI Spark Core Scheduler > StreamingJobProgressListener threw an exception > java.util.NoSuchElementException for Long Running Streaming Job > --------------------------------------------------------------------------------------------------------------- > > Key: SPARK-31969 > URL: https://issues.apache.org/jira/browse/SPARK-31969 > Project: Spark > Issue Type: Bug > Components: DStreams, Scheduler, Spark Core, Web UI > Affects Versions: 2.4.0 > Environment: Kubernetes > Spark 2.4.0 > Reporter: ThimmeGowda > Priority: Major > Attachments: driver_log, executor_log > > > We are running a long running streaming job and Below exception is seen > continuosly after sometime. After the jobs starts all of a sudden our Spark > streaming application's batch durations start to increase. At around the same > time there starts to appear an error log that does not refer to the > application code at all. We couldn't find any other significant errors in the > driver logs. > Refrerred ticket : https://issues.apache.org/jira/browse/SPARK-21065 for > similar issue, in our case we are not setting anything for > spark.streaming.concurrentJobs and default value is taken. > \{"type":"log", "level":"ERROR", "name":"STREAMING_OTHERS", > "time":"2020-06-09T04:31:43.918Z", "timezone":"UTC", > "class":"spark-listener-group-appStatus", > "method":"streaming.scheduler.StreamingListenerBus.logError(91)", > "log":"Listener StreamingJobProgressListener threw an > exception\u000Ajava.util.NoSuchElementException: key not found: 1591677100000 > ms\u000A\u0009at > scala.collection.MapLike$class.default(MapLike.scala:228)\u000A\u0009at > scala.collection.AbstractMap.default(Map.scala:59)\u000A\u0009at > scala.collection.mutable.HashMap.apply(HashMap.scala:65)\u000A\u0009at > org.apache.spark.streaming.ui.StreamingJobProgressListener.onOutputOperationCompleted(StreamingJobProgressListener.scala:134)\u000A\u0009at > > org.apache.spark.streaming.scheduler.StreamingListenerBus.doPostEvent(StreamingListenerBus.scala:67)\u000A\u0009at > > org.apache.spark.streaming.scheduler.StreamingListenerBus.doPostEvent(StreamingListenerBus.scala:29)\u000A\u0009at > > org.apache.spark.util.ListenerBus$class.postToAll(ListenerBus.scala:91)\u000A\u0009at > > org.apache.spark.streaming.scheduler.StreamingListenerBus.postToAll(StreamingListenerBus.scala:29)\u000A\u0009at > > org.apache.spark.streaming.scheduler.StreamingListenerBus.onOtherEvent(StreamingListenerBus.scala:43)\u000A\u0009at > > org.apache.spark.scheduler.SparkListenerBus$class.doPostEvent(SparkListenerBus.scala:80)\u000A\u0009at > > org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)\u000A\u0009at > > org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)\u000A\u0009at > > org.apache.spark.util.ListenerBus$class.postToAll(ListenerBus.scala:91)\u000A\u0009at > > org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$super$postToAll(AsyncEventQueue.scala:92)\u000A\u0009at > > org.apache.spark.scheduler.AsyncEventQueue$$anonfun$org$apache$spark$scheduler$AsyncEventQueue$$dispatch$1.apply$mcJ$sp(AsyncEventQueue.scala:92)\u000A\u0009at > > org.apache.spark.scheduler.AsyncEventQueue$$anonfun$org$apache$spark$scheduler$AsyncEventQueue$$dispatch$1.apply(AsyncEventQueue.scala:87)\u000A\u0009at > > org.apache.spark.scheduler.AsyncEventQueue$$anonfun$org$apache$spark$scheduler$AsyncEventQueue$$dispatch$1.apply(AsyncEventQueue.scala:87)\u000A\u0009at > scala.util.DynamicVariable.withValue(DynamicVariable.scala:58)\u000A\u0009at > org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$dispatch(AsyncEventQueue.scala:87)\u000A\u0009at > > org.apache.spark.scheduler.AsyncEventQueue$$anon$1$$anonfun$run$1.apply$mcV$sp(AsyncEventQueue.scala:83)\u000A\u0009at > > org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1302)\u000A\u0009at > > org.apache.spark.scheduler.AsyncEventQueue$$anon$1.run(AsyncEventQueue.scala:82)\u000A"} > java.util.NoSuchElementException: key not found: 1591677100000 ms > at scala.collection.MapLike$class.default(MapLike.scala:228) > ~[scala-library-2.11.12.jar:?] > at scala.collection.AbstractMap.default(Map.scala:59) > ~[scala-library-2.11.12.jar:?] > at scala.collection.mutable.HashMap.apply(HashMap.scala:65) > ~[scala-library-2.11.12.jar:?] > at > org.apache.spark.streaming.ui.StreamingJobProgressListener.onOutputOperationCompleted(StreamingJobProgressListener.scala:134) > ~[spark-streaming_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.streaming.scheduler.StreamingListenerBus.doPostEvent(StreamingListenerBus.scala:67) > ~[spark-streaming_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.streaming.scheduler.StreamingListenerBus.doPostEvent(StreamingListenerBus.scala:29) > [spark-streaming_2.11-2.4.0.jar:2.4.0] > at org.apache.spark.util.ListenerBus$class.postToAll(ListenerBus.scala:91) > [spark-core_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.streaming.scheduler.StreamingListenerBus.postToAll(StreamingListenerBus.scala:29) > [spark-streaming_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.streaming.scheduler.StreamingListenerBus.onOtherEvent(StreamingListenerBus.scala:43) > [spark-streaming_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.scheduler.SparkListenerBus$class.doPostEvent(SparkListenerBus.scala:80) > [spark-core_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37) > [spark-core_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37) > [spark-core_2.11-2.4.0.jar:2.4.0] > at org.apache.spark.util.ListenerBus$class.postToAll(ListenerBus.scala:91) > [spark-core_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$super$postToAll(AsyncEventQueue.scala:92) > [spark-core_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.scheduler.AsyncEventQueue$$anonfun$org$apache$spark$scheduler$AsyncEventQueue$$dispatch$1.apply$mcJ$sp(AsyncEventQueue.scala:92) > [spark-core_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.scheduler.AsyncEventQueue$$anonfun$org$apache$spark$scheduler$AsyncEventQueue$$dispatch$1.apply(AsyncEventQueue.scala:87) > [spark-core_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.scheduler.AsyncEventQueue$$anonfun$org$apache$spark$scheduler$AsyncEventQueue$$dispatch$1.apply(AsyncEventQueue.scala:87) > [spark-core_2.11-2.4.0.jar:2.4.0] > at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58) > [scala-library-2.11.12.jar:?] > at > org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$dispatch(AsyncEventQueue.scala:87) > [spark-core_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.scheduler.AsyncEventQueue$$anon$1$$anonfun$run$1.apply$mcV$sp(AsyncEventQueue.scala:83) > [spark-core_2.11-2.4.0.jar:2.4.0] > at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1302) > [spark-core_2.11-2.4.0.jar:2.4.0] > at > org.apache.spark.scheduler.AsyncEventQueue$$anon$1.run(AsyncEventQueue.scala:82) > [spark-core_2.11-2.4.0.jar:2.4.0] -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org