Can you take a look at the latest Spark 1.0 docs and see if they are fixed?
https://github.com/apache/spark/tree/master/docs Thanks. On Thu, May 29, 2014 at 5:29 AM, Lizhengbing (bing, BIPA) < zhengbing...@huawei.com> wrote: > The instruction address is in > http://spark.apache.org/docs/0.9.0/spark-standalone.html#launching-applications-inside-the-cluster > or > http://spark.apache.org/docs/0.9.1/spark-standalone.html#launching-applications-inside-the-cluster > > Origin instruction is: > "./bin/spark-class org.apache.spark.deploy.Client launch > [client-options] \ > <cluster-url> <application-jar-url> <main-class> \ > [application-options] " > > If I follow this instruction, I will not run my program deployed in a > spark standalone cluster properly. > > Based on source code, This instruction should be changed to > "./bin/spark-class org.apache.spark.deploy.Client [client-options] launch \ > <cluster-url> <application-jar-url> <main-class> \ > [application-options] " > > That is to say: [client-options] must be put ahead of launch >