Make sure you built spark with -Pyarn, and check whether you have class ExecutorLauncher in your spark assembly jar.
On Wed, Jun 22, 2016 at 2:04 PM, Yash Sharma <yash...@gmail.com> wrote: > How about supplying the jar directly in spark submit - > > ./bin/spark-submit \ >> --class org.apache.spark.examples.SparkPi \ >> --master yarn-client \ >> --driver-memory 512m \ >> --num-executors 2 \ >> --executor-memory 512m \ >> --executor-cores 2 \ >> /user/shihj/spark_lib/spark-examples-1.6.1-hadoop2.6.0.jar > > > On Wed, Jun 22, 2016 at 3:59 PM, 另一片天 <958943...@qq.com> wrote: > >> i config this para at spark-defaults.conf >> spark.yarn.jar >> hdfs://master:9000/user/shihj/spark_lib/spark-examples-1.6.1-hadoop2.6.0.jar >> >> then ./bin/spark-submit --class org.apache.spark.examples.SparkPi >> --master yarn-client --driver-memory 512m --num-executors 2 >> --executor-memory 512m --executor-cores 2 10: >> >> >> >> - Error: Could not find or load main class >> org.apache.spark.deploy.yarn.ExecutorLauncher >> >> but i don't config that para ,there no error why???that para is only >> avoid Uploading resource file(jar package)?? >> > > -- Best Regards Jeff Zhang