xjlin0 wrote > I cannot enter REPL shell in 1.4.0/1.4.1/1.5.0/1.5.1(with pre-built with > or without Hadoop or home compiled with ant or maven). There was no error > message in v1.4.x, system prompt nothing. On v1.5.x, once I enter > $SPARK_HOME/bin/pyspark or spark-shell, I got > > Error: Could not find or load main class org.apache.spark.launcher.Main
I have the same problem (on MacOS X Yosemite, all spark versions since 1.4, installed both with homebrew and downloaded manually). I've been trying to start the pyspark shell, but it also fails in the same way for spark-shell and spark-sql and spark-submit. I've narrowed it down to the following line in the spark-class script: done < <("$RUNNER" -cp "$LAUNCH_CLASSPATH" org.apache.spark.launcher.Main "$@") (where $RUNNER is "java" and $LAUNCH_CLASSPATH is "/usr/local/Cellar/apache-spark/1.5.1/libexec/lib/spark-assembly-1.5.1-hadoop2.6.0.jar", which does exist and does contain the org.apache.spark.launcher.Main class, despite the message that it can't be found) If I run it manually, using: SPARK_HOME=/usr/local/Cellar/apache-spark/1.5.1/libexec java -cp /usr/local/Cellar/apache-spark/1.5.1/libexec/lib/spark-assembly-1.5.1-hadoop2.6.0.jar org.apache.spark.launcher.Main org.apache.spark.deploy.SparkSubmit pyspark-shell-main --name PySparkShell It runs without that error, and instead prints out (where "\0" is a nul character): env\0PYSPARK_SUBMIT_ARGS="--name" "PySparkShell" "pyspark-shell"\0python\0 I'm not really sure what to try next, maybe with this extra information someone has an idea what's going wrong, and how to fix it. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Cannot-start-REPL-shell-since-1-4-0-tp24921p25176.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org