Since its a windows machine, you are very likely to be hitting this one https://issues.apache.org/jira/browse/SPARK-2356
Thanks Best Regards On Wed, Jul 1, 2015 at 12:36 AM, Sourav Mazumder < sourav.mazumde...@gmail.com> wrote: > Hi, > > I'm running Spark 1.4.0 without Hadoop. I'm using the binary > spark-1.4.0-bin-hadoop2.6. > > I start the spark-shell as : > > spark-shell --master local[2] --packages > com.databricks:spark-csv_2.11:1.1.0 --executor-memory 2G --conf > spark.local.dir="C:/Users/Sourav". > > Then I run : > > val df = > sqlContext.read.format("com.databricks.spark.csv").load("file:///C:/Users/Sourav/Work/SparkDataScience/test.csv"). > > It gives a null pointer exception - > > 15/06/30 12:03:44 INFO Executor: Running task 0.0 in stage 0.0 (TID 0) > 15/06/30 12:03:44 INFO Executor: Fetching > http://9.49.140.239:64868/jars/com.dat > abricks_spark-csv_2.11-1.1.0.jar with timestamp 1435690997767 > 15/06/30 12:03:44 INFO Utils: Fetching > http://9.49.140.239:64868/jars/com.databr > icks_spark-csv_2.11-1.1.0.jar to > C:\Users\Sourav\spark-18eb9880-4a19-46be-8c23-c > > 7f7e000c454\userFiles-d4df579c-4672-46ee-836c-d4dd9ea9be23\fetchFileTemp40728667 > 75534302313.tmp > 15/06/30 12:03:44 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID > 0) > java.lang.NullPointerException > at java.lang.ProcessBuilder.start(Unknown Source) > at org.apache.hadoop.util.Shell.runCommand(Shell.java:482) > at org.apache.hadoop.util.Shell.run(Shell.java:455) > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java: > 715) > at org.apache.hadoop.fs.FileUtil.chmod(FileUtil.java:873) > at org.apache.hadoop.fs.FileUtil.chmod(FileUtil.java:853) > at org.apache.spark.util.Utils$.fetchFile(Utils.scala:465) > at > org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor > $Executor$$updateDependencies$5.apply(Executor.scala:398) > at > org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor > $Executor$$updateDependencies$5.apply(Executor.scala:390) > at > scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply( > TraversableLike.scala:772) > at > scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.sca > la:98) > at > scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.sca > la:98) > at > scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala > :226) > at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) > at scala.collection.mutable.HashMap.foreach(HashMap.scala:98) > at > scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.s > cala:771) > at org.apache.spark.executor.Executor.org > $apache$spark$executor$Executor > $$updateDependencies(Executor.scala:390) > at > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:193) > > at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown > Source) > at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown > Source) > at java.lang.Thread.run(Unknown Source) > 15/06/30 12:03:44 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, > localh > ost): java.lang.NullPointerException > at java.lang.ProcessBuilder.start(Unknown Source) > at org.apache.hadoop.util.Shell.runCommand(Shell.java:482) > > > Any idea what is going wrong. > > Regards, > Sourav > >