Since you are running in yarn-cluster mode, and you are supply the spark assembly jar file. There is no need to install spark on each node. Is it possible two spark jars have different version ?
Chester Sent from my iPad On Jul 16, 2014, at 22:49, cmti95035 <cmti95...@gmail.com> wrote: > Hi, > > I need some help for running Spark over Yarn: > > I set up a cluster running HDP 2.0.6 with 6 nodes, and then installed the > spark-1.0.1-bin-hadoop2 on each node. When running the SparkPi example with > the following command: > > ./bin/spark-submit --class org.apache.spark.examples.SparkPi --master > yarn-cluster --num-executors 5 --driver-memory 4g --executor-memory > 2g --executor-cores 1 --jars lib/spark-assembly-1.0.1-hadoop2.2.0.jar > lib/spark-examples*.jar 10 > > The job failed with the following error: > > INFO yarn.Client: Application report from ASM: > application identifier: application_1405545630872_0023 > appId: 23 > clientToAMToken: null > appDiagnostics: Application application_1405545630872_0023 failed 2 > times > due to AM Container for appattempt_1405545630872_0023_000002 exited with > exitCode: -1000 due to: Resource > hdfs://ip-172-31-9-187.us-west-1.compute.internal:8020/user/hdfs/.sparkStaging/application_1405545630872_0023/spark-assembly-1.0.1-hadoop2.2.0.jar > changed on src filesystem (expected 1405574940411, was 1405574941940 > .Failing this attempt.. Failing the application. > > I searched online for solutions and tried to sync up ntp but it doesn't seem > to work. > > Can someone help? Your help is highly appreciated! > > Thanks, > > Jian > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/jar-changed-on-src-filesystem-tp10011.html > Sent from the Apache Spark User List mailing list archive at Nabble.com.