Hi Sai, I recently also ran into this problem on 0.9.1. The problem is that spark tries to read yarn's class path but when it finds it be empty does not fallback to it's default value. To resolve this, either set yarn.application.classpath in yarn-site.xml to its default value or put in a bug fix. FYI, this issue seems to be resolved in master branch (1.0).
Reference: https://groups.google.com/forum/#!topic/shark-users/8qFsy9JSt4E http://hadoop.apache.org/docs/r2.3.0/hadoop-yarn/hadoop-yarn-common/yarn-default.xml Thanks, Rahul Singhal From: Sai Prasanna <ansaiprasa...@gmail.com<mailto:ansaiprasa...@gmail.com>> Reply-To: "user@spark.apache.org<mailto:user@spark.apache.org>" <user@spark.apache.org<mailto:user@spark.apache.org>> Date: Monday 7 April 2014 6:56 PM To: "user@spark.apache.org<mailto:user@spark.apache.org>" <user@spark.apache.org<mailto:user@spark.apache.org>> Subject: Null Pointer Exception in Spark Application with Yarn Client Mode Hi All, I wanted Spark on Yarn to up and running. I did "SPARK_HADOOP_VERSION=2.3.0 SPARK_YARN=true ./sbt/sbt assembly" Then i ran "SPARK_JAR=./assembly/target/scala-2.9.3/spark-assembly-0.8.1-incubating-hadoop2.3.0.jar SPARK_YARN_APP_JAR=examples/target/scala-2.9.3/spark-examples_2.9.3-0.8.1-incubating.jar MASTER=yarn-client ./spark-shell" I have SPARK_HOME, YARN_CONF_DIR/HADOOP_CONF_DIR set. Still i get the following error. Any clues ?? ERROR: ........................................... Using Scala version 2.9.3 (Java HotSpot(TM) 64-Bit Server VM, Java 1.7.0_51) Initializing interpreter... Creating SparkContext... java.lang.NullPointerException at scala.collection.mutable.ArrayOps$ofRef.length(ArrayOps.scala:115) at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) at scala.collection.mutable.ArrayOps.foreach(ArrayOps.scala:38) at org.apache.spark.deploy.yarn.Client$.populateHadoopClasspath(Client.scala:489) at org.apache.spark.deploy.yarn.Client$.populateClasspath(Client.scala:510) at org.apache.spark.deploy.yarn.Client.setupLaunchEnv(Client.scala:327) at org.apache.spark.deploy.yarn.Client.runApp(Client.scala:90) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:71) at org.apache.spark.scheduler.cluster.ClusterScheduler.start(ClusterScheduler.scala:119) at org.apache.spark.SparkContext.<init>(SparkContext.scala:273) at org.apache.spark.repl.SparkILoop.createSparkContext(SparkILoop.scala:862) at <init>(<console>:10) at <init>(<console>:22) at <init>(<console>:24) at .<init>(<console>:28) at .<clinit>(<console>) at .<init>(<console>:7) at .<clinit>(<console>) at $export(<console>) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:629) at org.apache.spark.repl.SparkIMain$Request$$anonfun$10.apply(SparkIMain.scala:897) at scala.tools.nsc.interpreter.Line$$anonfun$1.apply$mcV$sp(Line.scala:43) at scala.tools.nsc.io.package$$anon$2.run(package.scala:25) at java.lang.Thread.run(Thread.java:744) .....................................................................