Awesome! It's documented here: https://spark.apache.org/docs/latest/submitting-applications.html
-Sandy On Mon, May 18, 2015 at 8:03 PM, xiaohe lan <zombiexco...@gmail.com> wrote: > Hi Sandy, > > Thanks for your information. Yes, spark-submit --master yarn > --num-executors 5 --executor-cores 4 > target/scala-2.10/simple-project_2.10-1.0.jar --class scala.SimpleApp is > working awesomely. Is there any documentations pointing to this ? > > Thanks, > Xiaohe > > On Tue, May 19, 2015 at 12:07 AM, Sandy Ryza <sandy.r...@cloudera.com> > wrote: > >> Hi Xiaohe, >> >> The all Spark options must go before the jar or they won't take effect. >> >> -Sandy >> >> On Sun, May 17, 2015 at 8:59 AM, xiaohe lan <zombiexco...@gmail.com> >> wrote: >> >>> Sorry, them both are assigned task actually. >>> >>> Aggregated Metrics by Executor >>> Executor IDAddressTask TimeTotal TasksFailed TasksSucceeded TasksInput >>> Size / RecordsShuffle Write Size / RecordsShuffle Spill (Memory)Shuffle >>> Spill (Disk)1host1:61841.7 min505640.0 MB / 12318400382.3 MB / >>> 121007701630.4 >>> MB295.4 MB2host2:620721.7 min505640.0 MB / 12014510386.0 MB / 109269121646.6 >>> MB304.8 MB >>> >>> On Sun, May 17, 2015 at 11:50 PM, xiaohe lan <zombiexco...@gmail.com> >>> wrote: >>> >>>> bash-4.1$ ps aux | grep SparkSubmit >>>> xilan 1704 13.2 1.2 5275520 380244 pts/0 Sl+ 08:39 0:13 >>>> /scratch/xilan/jdk1.8.0_45/bin/java -cp >>>> /scratch/xilan/spark/conf:/scratch/xilan/spark/lib/spark-assembly-1.3.1-hadoop2.4.0.jar:/scratch/xilan/spark/lib/datanucleus-core-3.2.10.jar:/scratch/xilan/spark/lib/datanucleus-api-jdo-3.2.6.jar:/scratch/xilan/spark/lib/datanucleus-rdbms-3.2.9.jar:/scratch/xilan/hadoop/etc/hadoop >>>> -Xms512m -Xmx512m org.apache.spark.deploy.SparkSubmit --master yarn >>>> target/scala-2.10/simple-project_2.10-1.0.jar --class scala.SimpleApp >>>> --num-executors 5 --executor-cores 4 >>>> xilan 1949 0.0 0.0 103292 800 pts/1 S+ 08:40 0:00 grep >>>> --color SparkSubmit >>>> >>>> >>>> When look at the sparkui, I see the following: >>>> Aggregated Metrics by ExecutorExecutor IDAddressTask TimeTotal TasksFailed >>>> TasksSucceeded TasksShuffle Read Size / Records1host1:304836 s101127.1 >>>> MB / 28089782host2:49970 ms00063.4 MB / 1810945 >>>> >>>> So executor 2 is not even assigned a task ? Maybe I have some problems >>>> in my setting, but I don't know what could be the possible settings I set >>>> wrong or have not set. >>>> >>>> >>>> Thanks, >>>> Xiaohe >>>> >>>> On Sun, May 17, 2015 at 11:16 PM, Akhil Das <ak...@sigmoidanalytics.com >>>> > wrote: >>>> >>>>> Did you try --executor-cores param? While you submit the job, do a ps >>>>> aux | grep spark-submit and see the exact command parameters. >>>>> >>>>> Thanks >>>>> Best Regards >>>>> >>>>> On Sat, May 16, 2015 at 12:31 PM, xiaohe lan <zombiexco...@gmail.com> >>>>> wrote: >>>>> >>>>>> Hi, >>>>>> >>>>>> I have a 5 nodes yarn cluster, I used spark-submit to submit a simple >>>>>> app. >>>>>> >>>>>> spark-submit --master yarn >>>>>> target/scala-2.10/simple-project_2.10-1.0.jar --class scala.SimpleApp >>>>>> --num-executors 5 >>>>>> >>>>>> I have set the number of executor to 5, but from sparkui I could see >>>>>> only two executors and it ran very slow. What did I miss ? >>>>>> >>>>>> Thanks, >>>>>> Xiaohe >>>>>> >>>>> >>>>> >>>> >>> >> >