>> "The proper way to specify this is through "spark.master" in your config
or the "--master" parameter to spark-submit."

By "this" I mean configuring which master the driver connects to (not which
port and address the standalone Master binds to).


2014-07-08 16:43 GMT-07:00 Andrew Or <and...@databricks.com>:

> Hi Mikhail,
>
> It looks like the documentation is a little out-dated. Neither is true
> anymore. In general, we try to shift away from short options ("-em", "-dm"
> etc.) in favor of more explicit ones ("--executor-memory",
> "--driver-memory"). These options, and "--cores", refer to the arguments
> passed in to org.apache.spark.deploy.Client, the now deprecated way of
> launching an application in standalone clusters.
>
> SPARK_MASTER_IP/PORT are only used for binding the Master, not to
> configure which master the driver connects to. The proper way to specify
> this is through "spark.master" in your config or the "--master" parameter
> to spark-submit.
>
> We will update the documentation shortly. Thanks for letting us know.
> Andrew
>
>
>
> 2014-07-08 16:29 GMT-07:00 Mikhail Strebkov <streb...@gmail.com>:
>
> Hi! I've been using Spark compiled from 1.0 branch at some point (~2 month
>> ago). The setup is a standalone cluster with 4 worker machines and 1
>> master
>> machine. I used to run spark shell like this:
>>
>>   ./bin/spark-shell -c 30 -em 20g -dm 10g
>>
>> Today I've finally updated to Spark 1.0 release. Now I can only run spark
>> shell like this:
>>
>>   ./bin/spark-shell --master spark://10.2.1.5:7077
>> --total-executor-cores 30
>> --executor-memory 20g --driver-memory 10g
>>
>> The documentation at
>> http://spark.apache.org/docs/latest/spark-standalone.html says:
>>
>> "You can also pass an option --cores <numCores> to control the number of
>> cores that spark-shell uses on the cluster."
>> This doesn't work, you need to pass "--total-executor-cores <numCores>"
>> instead.
>>
>> "Note that if you are running spark-shell from one of the spark cluster
>> machines, the bin/spark-shell script will automatically set MASTER from
>> the
>> SPARK_MASTER_IP and SPARK_MASTER_PORT variables in conf/spark-env.sh."
>> This is not working for me too. I run the shell from the master machine,
>> and
>> I do have SPARK_MASTER_IP set up in conf/spark-env.sh like this:
>> export SPARK_MASTER_IP='10.2.1.5'
>> But if I omit "--master spark://10.2.1.5:7077" then the console starts
>> but I
>> can't see it in the UI at http://10.2.1.5:8080. But when I go to
>> http://10.2.1.5:4040 (the application UI) I see that the app is using
>> only
>> master as a worker.
>>
>> My question is: are those just the bugs in the documentation? That there
>> is
>> no --cores option and that SPARK_MASTER_IP is not used anymore when I run
>> the Spark shell from the master?
>>
>>
>>
>> --
>> View this message in context:
>> http://apache-spark-user-list.1001560.n3.nabble.com/issues-with-bin-spark-shell-for-standalone-mode-tp9107.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>
>
>

Reply via email to