After start the dfs, yarn and spark, I run these code under the root
directory of spark on my master host: 
`MASTER=yarn ./bin/run-example ml.LogisticRegressionExample
data/mllib/sample_libsvm_data.txt`

Actually I get these code from spark's README. And here is the source code
about LogisticRegressionExample on GitHub: 
https://github.com/apache/spark/blob/master/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionExample.scala
<https://github.com/apache/spark/blob/master/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionExample.scala>
  

Then, error occurs: 
`Exception in thread "main" org.apache.spark.sql.AnalysisException: Path
does notexist:
hdfs://master:9000/user/root/data/mllib/sample_libsvm_data.txt;`

Firstly, I don't know why it's `hdfs://master:9000/user/root`, I do set
namenode's IP address to `hdfs://master:9000`, but why spark chose the
directory `/user/root`?

Then, I make a directory `/user/root/data/mllib/sample_libsvm_data.txt` on
every host of the cluster, so I hope spark can find this file. But the same
error occurs again. 



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Run-spark-machine-learning-example-on-Yarn-failed-tp28435.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to