Hi Sam,

Did you specify the MASTER in your spark-env.sh? I ask because I didn't see
a --master in your launch command. Also, your app seems to take in a master
("yarn-standalone"). This is not exactly correct because by the time the
SparkContext is launched locally, which is the default, it is too late to
use "yarn-cluster" mode by definition, since the driver should launched
within one of the containers on the worker machines.

I would suggest the following:
- change your application to not take in the Spark master as a command line
argument
- use "yarn-cluster" instead of "yarn-standalone" (which is deprecated)
- add "--master yarn-cluster" in your spark-submit command

Another worrying thing is the warning from your logs:
14/07/21 22:38:42 WARN spark.SparkConf: null jar passed to SparkContext
constructor

How are you creating your SparkContext?

Andrew



2014-07-21 7:47 GMT-07:00 Sam Liu <liuqiyun_sp...@sina.com>:

> Hi Experts,
>
>
> I setup Yarn and Spark env: all services runs on a single node. And then
> submited a WordCount job using spark-submit script with command:
>
> ./bin/spark-submit tests/wordcount-spark-scala.jar --class
> scala.spark.WordCount --num-executors 1 --driver-memory 300M
> --executor-memory 300M --executor-cores 1 "yarn-standalone"
> "hdfs://hostname/tmp/input" "hdfs://hostname/tmp/output"
>
>
> However, the command hangs and no job is submited to Yarn. Any comments?
>
>
> output:
>
> Spark assembly has been built with Hive, including Datanucleus jars on
> classpath
> 14/07/21 22:38:42 WARN spark.SparkConf: null jar passed to SparkContext
> constructor
> 14/07/21 22:38:43 INFO spark.SecurityManager: Changing view acls to:
> biadmin
> 14/07/21 22:38:43 INFO spark.SecurityManager: SecurityManager:
> authentication disabled; ui acls disabled; users with view permissions:
> Set(biadmin)
> 14/07/21 22:38:43 INFO slf4j.Slf4jLogger: Slf4jLogger started
> 14/07/21 22:38:43 INFO Remoting: Starting remoting
> 14/07/21 22:38:43 INFO Remoting: Remoting started; listening on addresses
> :[akka.tcp://spark@hostname:56903]
> 14/07/21 22:38:43 INFO Remoting: Remoting now listens on addresses:
> [akka.tcp://spark@hostname:56903]
> 14/07/21 22:38:43 INFO spark.SparkEnv: Registering MapOutputTracker
> 14/07/21 22:38:43 INFO spark.SparkEnv: Registering BlockManagerMaster
> 14/07/21 22:38:43 INFO storage.DiskBlockManager: Created local directory
> at /tmp/spark-local-20140721223843-75cd
> 14/07/21 22:38:43 INFO storage.MemoryStore: MemoryStore started with
> capacity 180.0 MB.
> 14/07/21 22:38:43 INFO network.ConnectionManager: Bound socket to port
> 57453 with id = ConnectionManagerId(hostname,57453)
> 14/07/21 22:38:43 INFO storage.BlockManagerMaster: Trying to register
> BlockManager
> 14/07/21 22:38:43 INFO storage.BlockManagerInfo: Registering block manager
> hostname:57453 with 180.0 MB RAM
> 14/07/21 22:38:43 INFO storage.BlockManagerMaster: Registered BlockManager
> 14/07/21 22:38:43 INFO spark.HttpServer: Starting HTTP Server
> 14/07/21 22:38:43 INFO server.Server: jetty-8.y.z-SNAPSHOT
> 14/07/21 22:38:43 INFO server.AbstractConnector: Started
> SocketConnector@0.0.0.0:19323
> 14/07/21 22:38:43 INFO broadcast.HttpBroadcast: Broadcast server started
> at http://9.123.99.10:19323
> 14/07/21 22:38:43 INFO spark.HttpFileServer: HTTP File server directory is
> /tmp/spark-e224a31b-4517-43d8-9778-4b6af07dcad2
> 14/07/21 22:38:43 INFO spark.HttpServer: Starting HTTP Server
> 14/07/21 22:38:43 INFO server.Server: jetty-8.y.z-SNAPSHOT
> 14/07/21 22:38:43 INFO server.AbstractConnector: Started
> SocketConnector@0.0.0.0:35420
> 14/07/21 22:38:43 INFO server.Server: jetty-8.y.z-SNAPSHOT
> 14/07/21 22:38:43 INFO server.AbstractConnector: Started
> SelectChannelConnector@0.0.0.0:4040
> 14/07/21 22:38:43 INFO ui.SparkUI: Started SparkUI at http://hostname:4040
> 14/07/21 22:38:44 WARN util.NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> 14/07/21 22:38:44 WARN spark.SparkContext: "yarn-standalone" is deprecated
> as of Spark 1.0. Use "yarn-cluster" instead.
> 14/07/21 22:38:44 INFO cluster.YarnClusterScheduler: Created
> YarnClusterScheduler
> 14/07/21 22:38:44 INFO yarn.ApplicationMaster$$anon$1: Adding shutdown
> hook for context org.apache.spark.SparkContext@610c610c
>
>
>
>
> Thanks!
> --------------------------------
> Sam Liu
>
>

Reply via email to