They should be - in the sense that the docs now recommend using
spark-submit and thus include entirely different invocations.
On Fri, May 30, 2014 at 12:46 AM, Reynold Xin r...@databricks.com wrote:
Can you take a look at the latest Spark 1.0 docs and see if they are fixed?
https://github.com/apache/spark/tree/master/docs
Thanks.
On Thu, May 29, 2014 at 5:29 AM, Lizhengbing (bing, BIPA)
zhengbing...@huawei.com wrote:
The instruction address is in
http://spark.apache.org/docs/0.9.0/spark-standalone.html#launching-applications-inside-the-cluster
or
http://spark.apache.org/docs/0.9.1/spark-standalone.html#launching-applications-inside-the-cluster
Origin instruction is:
./bin/spark-class org.apache.spark.deploy.Client launch
[client-options] \
cluster-url application-jar-url main-class \
[application-options]
If I follow this instruction, I will not run my program deployed in a
spark standalone cluster properly.
Based on source code, This instruction should be changed to
./bin/spark-class org.apache.spark.deploy.Client [client-options]
launch \
cluster-url application-jar-url main-class \
[application-options]
That is to say: [client-options] must be put ahead of launch