[ https://issues.apache.org/jira/browse/SPARK-22012?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Sean Owen resolved SPARK-22012. ------------------------------- Resolution: Invalid > CLONE - Spark Streaming, Kafka receiver, "Failed to get records for ... after > polling for 512" > ---------------------------------------------------------------------------------------------- > > Key: SPARK-22012 > URL: https://issues.apache.org/jira/browse/SPARK-22012 > Project: Spark > Issue Type: Bug > Components: DStreams > Affects Versions: 2.0.0 > Environment: Apache Spark 2.1.0.2.6.0.3-8, Kafka 0.10 for Java 8 > Reporter: Karan Singh > Priority: Blocker > > My Spark Streaming duration is 5 seconds (5000) and kafka is all at its > default properties , i am still facing this issue , Can anyone tell me how to > resolve it what i am doing wrong ? > JavaStreamingContext ssc = new JavaStreamingContext(sc, > new Duration(5000)); > Exception in Spark Streamings > Job aborted due to stage failure: Task 6 in stage 289.0 failed 4 times, most > recent failure: Lost task 6.3 in stage 289.0 (xxxxxx, executor 2): > java.lang.AssertionError: assertion failed: Failed to get records for > spark-executor-xxx pulse 1 163684030 after polling for 512 > at scala.Predef$.assert(Predef.scala:170) > at > org.apache.spark.streaming.kafka010.CachedKafkaConsumer.get(CachedKafkaConsumer.scala:74) > at > org.apache.spark.streaming.kafka010.KafkaRDD$KafkaRDDIterator.next(KafkaRDD.scala:227) > at > org.apache.spark.streaming.kafka010.KafkaRDD$KafkaRDDIterator.next(KafkaRDD.scala:193) > at > scala.collection.convert.Wrappers$IteratorWrapper.next(Wrappers.scala:31) > at > com.olacabs.analytics.engine.EngineManager$1$1.call(EngineManager.java:165) > at > com.olacabs.analytics.engine.EngineManager$1$1.call(EngineManager.java:132) > at > org.apache.spark.api.java.JavaRDDLike$$anonfun$foreachPartition$1.apply(JavaRDDLike.scala:219) > at > org.apache.spark.api.java.JavaRDDLike$$anonfun$foreachPartition$1.apply(JavaRDDLike.scala:219) > at > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$29.apply(RDD.scala:926) > at > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$29.apply(RDD.scala:926) > at > org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1951) > at > org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1951) > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87) > at org.apache.spark.scheduler.Task.run(Task.scala:99) > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:322) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org