You're specifying the queue in the spark-submit command line:

  --queue thequeue

Are you sure that queue exists?


On Thu, Jan 15, 2015 at 11:23 AM, Manoj Samel <manojsamelt...@gmail.com> wrote:
> Hi,
>
> Setup is as follows
>
> Hadoop Cluster 2.3.0 (CDH5.0)
> - Namenode HA
> - Resource manager HA
> - Secured with Kerberos
>
> Spark 1.2
>
> Run SparkPi as follows
> - conf/spark-defaults.conf has following entries
> spark.yarn.queue myqueue
> spark.yarn.access.namenodes hdfs://namespace (remember this is namenode HA)
> - Do kinit with some user keytab
> - submit SparkPi as follows
> spark-submit --class org.apache.spark.examples.SparkPi --master yarn-client
> --num-executors 3 --driver-memory 4g --executor-memory 2g --executor-cores 1
> --queue thequeue $MY_SPARK_DIR/lib/spark-examples*.jar 10
>
> Gives following trace (not sure why it shows unknown queue when queue name
> is specified in the spark-defaults.conf above.
>
> 15/01/15 19:18:27 INFO impl.YarnClientImpl: Submitted application
> application_1415648563285_31469
> 15/01/15 19:18:28 INFO yarn.Client: Application report for
> application_1415648563285_31469 (state: FAILED)
> 15/01/15 19:18:28 INFO yarn.Client:
> client token: N/A
> diagnostics: Application application_1415648563285_31469 submitted by user
> XYZ to unknown queue: thequeue <<--- WHY UNKNOWN QUEUE ???
> ApplicationMaster host: N/A
> ApplicationMaster RPC port: -1
> queue: thequeue   <<--- WHY UNKNOWN QUEUE ???
> start time: 1421349507652
> final status: FAILED
> tracking URL: N/A
> user: XYZ
> Exception in thread "main" org.apache.spark.SparkException: Yarn application
> has already ended! It might have been killed or unable to launch application
> master.
> at
> org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication(YarnClientSchedulerBackend.scala:102)
> at
> org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:58)



-- 
Marcelo

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to