Re: Error when running SparkPi on Secure HA Hadoop cluster

2015-01-15 Thread Marcelo Vanzin
You're specifying the queue in the spark-submit command line:

  --queue thequeue

Are you sure that queue exists?


On Thu, Jan 15, 2015 at 11:23 AM, Manoj Samel manojsamelt...@gmail.com wrote:
 Hi,

 Setup is as follows

 Hadoop Cluster 2.3.0 (CDH5.0)
 - Namenode HA
 - Resource manager HA
 - Secured with Kerberos

 Spark 1.2

 Run SparkPi as follows
 - conf/spark-defaults.conf has following entries
 spark.yarn.queue myqueue
 spark.yarn.access.namenodes hdfs://namespace (remember this is namenode HA)
 - Do kinit with some user keytab
 - submit SparkPi as follows
 spark-submit --class org.apache.spark.examples.SparkPi --master yarn-client
 --num-executors 3 --driver-memory 4g --executor-memory 2g --executor-cores 1
 --queue thequeue $MY_SPARK_DIR/lib/spark-examples*.jar 10

 Gives following trace (not sure why it shows unknown queue when queue name
 is specified in the spark-defaults.conf above.

 15/01/15 19:18:27 INFO impl.YarnClientImpl: Submitted application
 application_1415648563285_31469
 15/01/15 19:18:28 INFO yarn.Client: Application report for
 application_1415648563285_31469 (state: FAILED)
 15/01/15 19:18:28 INFO yarn.Client:
 client token: N/A
 diagnostics: Application application_1415648563285_31469 submitted by user
 XYZ to unknown queue: thequeue --- WHY UNKNOWN QUEUE ???
 ApplicationMaster host: N/A
 ApplicationMaster RPC port: -1
 queue: thequeue   --- WHY UNKNOWN QUEUE ???
 start time: 1421349507652
 final status: FAILED
 tracking URL: N/A
 user: XYZ
 Exception in thread main org.apache.spark.SparkException: Yarn application
 has already ended! It might have been killed or unable to launch application
 master.
 at
 org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication(YarnClientSchedulerBackend.scala:102)
 at
 org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:58)



-- 
Marcelo

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



Error when running SparkPi on Secure HA Hadoop cluster

2015-01-15 Thread Manoj Samel
Hi,

Setup is as follows

Hadoop Cluster 2.3.0 (CDH5.0)
- Namenode HA
- Resource manager HA
- Secured with Kerberos

Spark 1.2

Run SparkPi as follows
- conf/spark-defaults.conf has following entries
spark.yarn.queue myqueue
spark.yarn.access.namenodes hdfs://namespace (remember this is namenode HA)
- Do kinit with some user keytab
- submit SparkPi as follows
spark-submit --class org.apache.spark.examples.SparkPi --master yarn-client
--num-executors 3 --driver-memory 4g --executor-memory 2g --executor-cores
1 --queue thequeue $MY_SPARK_DIR/lib/spark-examples*.jar 10

Gives following trace (not sure why it shows unknown queue when queue name
is specified in the spark-defaults.conf above.

15/01/15 19:18:27 INFO impl.YarnClientImpl: Submitted application
application_1415648563285_31469
15/01/15 19:18:28 INFO yarn.Client: Application report for
application_1415648563285_31469 (state: FAILED)
15/01/15 19:18:28 INFO yarn.Client:
 client token: N/A
 diagnostics: Application application_1415648563285_31469 submitted by user
XYZ to unknown queue: thequeue --- WHY UNKNOWN QUEUE ???
 ApplicationMaster host: N/A
 ApplicationMaster RPC port: -1
 queue: thequeue   --- WHY UNKNOWN QUEUE ???
 start time: 1421349507652
 final status: FAILED
 tracking URL: N/A
 user: XYZ
Exception in thread main org.apache.spark.SparkException: Yarn
application has already ended! It might have been killed or unable to
launch application master.
at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication(YarnClientSchedulerBackend.scala:102)
at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:58)