May be you should update your spark version to the latest one. Thanks Best Regards
On Wed, Jun 10, 2015 at 11:04 AM, Chandrashekhar Kotekar < shekhar.kote...@gmail.com> wrote: > Hi, > > I have configured Spark to run on YARN. Whenever I start spark shell using > 'spark-shell' command, it automatically gets killed. Output looks like > below: > > ubuntu@dev-cluster-gateway:~$ ls shekhar/ > edx-spark > ubuntu@dev-cluster-gateway:~$ spark-shell > Welcome to > ____ __ > / __/__ ___ _____/ /__ > _\ \/ _ \/ _ `/ __/ '_/ > /___/ .__/\_,_/_/ /_/\_\ version 1.2.0-SNAPSHOT > /_/ > > Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_75) > Type in expressions to have them evaluated. > Type :help for more information. > 15/06/10 05:20:45 WARN Utils: Your hostname, dev-cluster-gateway resolves > to a loopback address: 127.0.0.1; using 10.182.149.171 instead (on > interface eth0) > 15/06/10 05:20:45 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to > another address > 15/06/10 05:21:20 WARN NativeCodeLoader: Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > /usr/lib/spark/bin/spark-shell: line 48: 15573 Killed > "$FWDIR"/bin/spark-submit --class org.apache.spark.repl.Main > "${SUBMISSION_OPTS[@]}" spark-shell "${APPLICATION_OPTS[@]}" > > > Any clue why spark shell gets killed? Please let me know if other > configuration/information is required. > > Thanks, > Chandrash3khar Kotekar >