Hi, When i trying to execute the program from my laptop by connecting to HDP environment (on which Spark also configured), i'm getting the warning ("Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory") and Job is being terminated. My console has following log statements.
Note: I could able to run the same client program by using spark-submit command. Whatever parameters i passed to spark-submit command, i passed the same to to SparkConf object. But still getting the same error. Any clue on this? 14/11/24 16:07:09 INFO scheduler.DAGScheduler: Submitting 2 missing tasks from Stage 0 (MappedRDD[4] at map at JavaSchemaRDD.scala:42) 14/11/24 16:07:09 INFO scheduler.TaskSchedulerImpl: Adding task set 0.0 with 2 tasks 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor updated: app-20141124023636-0004/0 is now EXITED (Command exited with code 1) 14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Executor app-20141124023636-0004/0 removed: Command exited with code 1 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor added: app-20141124023636-0004/2 on worker-20141124021958-STI-SM-DEV-SYS4-51561 (STI-SM-DEV-SYS4:51561) with 4 cores 14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Granted executor ID app-20141124023636-0004/2 on hostPort STI-SM-DEV-SYS4:51561 with 4 cores, 8.0 GB RAM 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor updated: app-20141124023636-0004/1 is now EXITED (Command exited with code 1) 14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Executor app-20141124023636-0004/1 removed: Command exited with code 1 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor added: app-20141124023636-0004/3 on worker-20141124022001-STI-SM-DEV-SYS5-50404 (STI-SM-DEV-SYS5:50404) with 4 cores 14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Granted executor ID app-20141124023636-0004/3 on hostPort STI-SM-DEV-SYS5:50404 with 4 cores, 8.0 GB RAM 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor updated: app-20141124023636-0004/2 is now RUNNING 14/11/24 16:07:10 INFO client.AppClient$ClientActor: Executor updated: app-20141124023636-0004/3 is now RUNNING 14/11/24 16:07:24 WARN scheduler.TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory 14/11/24 16:07:39 WARN scheduler.TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory 14/11/24 16:07:43 INFO client.AppClient$ClientActor: Executor updated: app-20141124023636-0004/3 is now EXITED (Command exited with code 1) 14/11/24 16:07:43 INFO cluster.SparkDeploySchedulerBackend: Executor app-20141124023636-0004/3 removed: Command exited with code 1 Thanks in advance. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/issue-while-running-the-code-in-standalone-mode-Initial-job-has-not-accepted-any-resources-check-you-tp19628.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org