Hi I could run spark trunk code on top of yarn 2.0.5-alpha by
SPARK_JAR=./core/target/spark-core-assembly-0.8.0-SNAPSHOT.jar ./run spark.deploy.yarn.Client \ --jar examples/target/scala-2.9.3/spark-examples_2.9.3-0.8.0-SNAPSHOT.jar \ --class spark.examples.SparkPi \ --args yarn-standalone \ --num-workers 3 \ --worker-memory 2g \ --worker-cores 2 While, if I use make-distribution.sh to build a release package and use this package on the cluster. Then it fails to run up. I do copy examples jar to jars/ dir. The other mode say standalone/mesos/local runs well with the release package. The error encounter is : Exception in thread "main" java.io.IOException: No FileSystem for scheme: hdfs at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2265) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2272) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:86) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2311) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2293) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:317) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:163) at spark.deploy.yarn.Client.prepareLocalResources(Client.scala:117) at spark.deploy.yarn.Client.run(Client.scala:59) at spark.deploy.yarn.Client$.main(Client.scala:318) at spark.deploy.yarn.Client.main(Client.scala) google result seems leading to hdfs core-default.xml not included in the fat jar. While I checked that it did. Any idea on this issue? Thanks! Best Regards, Raymond Liu