Hi, We use the following Spark Streaming code to collect and process Kafka event :
kafkaStream.foreachRDD(rdd => { rdd.collect().foreach(event => { process(event._1, event._2) }) }) This work fine. But without /collect()/ function, the following exception is raised for call to function process: *Loss was due to java.lang.ExceptionInInitializerError* We attempt to rewrite like this but the same exception is raised : kafkaStream.foreachRDD(rdd => { rdd.foreachPartition(iter => iter.foreach (event => { process(event._1, event._2) }) ) }) Does anybody can explain to us why and how to solve this issue ? Thank's Regards -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Spark-steaming-work-with-collect-but-not-without-collect-tp20622.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org