Hi, I have a standalone cluster. One Master + One Slave. I'm getting below "NULL POINTER" exception.
Could you please help me on this issue. *Code Block :-* val accum = sc.accumulator(0) sc.parallelize(Array(1, 2, 3, 4)).foreach(x => accum += x) *==> This line giving exception.* Exception :- 15/12/22 09:18:26 WARN scheduler.TaskSetManager: Lost task 1.0 in stage 0.0 (TID 1, 172.25.111.123): *java.lang.NullPointerException* at com.cc.ss.etl.Main$$anonfun$1.apply$mcVI$sp(Main.scala:25) at com.cc.ss.etl.Main$$anonfun$1.apply(Main.scala:25) at com.cc.ss.etl.Main$$anonfun$1.apply(Main.scala:25) at scala.collection.Iterator$class.foreach(Iterator.scala:727) at org.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28) at org.apache.spark.rdd.RDD$$anonfun$foreach$1$$anonfun$apply$28.apply(RDD.scala:890) at org.apache.spark.rdd.RDD$$anonfun$foreach$1$$anonfun$apply$28.apply(RDD.scala:890) at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848) at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:88) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) 15/12/22 09:18:26 INFO scheduler.TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0) on executor 172.25.111.123: java.lang.NullPointerException (null) [duplicate 1] 15/12/22 09:18:26 INFO scheduler.TaskSetManager: Starting task 0.1 in stage 0.0 (TID 2, 172.25.111.123, PROCESS_LOCAL, 2155 bytes) 15/12/22 09:18:26 INFO scheduler.TaskSetManager: Starting task 1.1 in stage 0.0 (TID 3, 172.25.111.123, PROCESS_LOCAL, 2155 bytes) 15/12/22 09:18:26 WARN scheduler.TaskSetManager: Lost task 1.1 in stage 0.0 (TID 3, 172.25.111.123): Regards, Rajesh