This can happen mainly because of the following: - Wrong master url (Make sure you give the master url which is listed on top left corner of the webui - running on 8080)
- Allocated more memory/cores while creating the sparkContext. Thanks Best Regards On Mon, Nov 24, 2014 at 4:13 PM, vdiwakar.malladi < vdiwakar.mall...@gmail.com> wrote: > Hi, > > When i trying to execute the program from my laptop by connecting to HDP > environment (on which Spark also configured), i'm getting the warning > ("Initial job has not accepted any resources; check your cluster UI to > ensure that workers are registered and have sufficient memory") and Job is > being terminated. My console has following log statements. > > Note: I could able to run the same client program by using spark-submit > command. Whatever parameters i passed to spark-submit command, i passed the > same to to SparkConf object. But still getting the same error. Any clue on > this? > > 14/11/24 16:07:09 INFO scheduler.DAGScheduler: Submitting 2 missing tasks > from Stage 0 (MappedRDD[4] at map at JavaSchemaRDD.scala:42) > 14/11/24 16:07:09 INFO scheduler.TaskSchedulerImpl: Adding task set 0.0 > with > 2 tasks > 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor updated: > app-20141124023636-0004/0 is now EXITED (Command exited with code 1) > 14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Executor > app-20141124023636-0004/0 removed: Command exited with code 1 > 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor added: > app-20141124023636-0004/2 on worker-20141124021958-STI-SM-DEV-SYS4-51561 > (STI-SM-DEV-SYS4:51561) with 4 cores > 14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Granted > executor > ID app-20141124023636-0004/2 on hostPort STI-SM-DEV-SYS4:51561 with 4 > cores, > 8.0 GB RAM > 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor updated: > app-20141124023636-0004/1 is now EXITED (Command exited with code 1) > 14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Executor > app-20141124023636-0004/1 removed: Command exited with code 1 > 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor added: > app-20141124023636-0004/3 on worker-20141124022001-STI-SM-DEV-SYS5-50404 > (STI-SM-DEV-SYS5:50404) with 4 cores > 14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Granted > executor > ID app-20141124023636-0004/3 on hostPort STI-SM-DEV-SYS5:50404 with 4 > cores, > 8.0 GB RAM > 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor updated: > app-20141124023636-0004/2 is now RUNNING > 14/11/24 16:07:10 INFO client.AppClient$ClientActor: Executor updated: > app-20141124023636-0004/3 is now RUNNING > 14/11/24 16:07:24 WARN scheduler.TaskSchedulerImpl: Initial job has not > accepted any resources; check your cluster UI to ensure that workers are > registered and have sufficient memory > 14/11/24 16:07:39 WARN scheduler.TaskSchedulerImpl: Initial job has not > accepted any resources; check your cluster UI to ensure that workers are > registered and have sufficient memory > 14/11/24 16:07:43 INFO client.AppClient$ClientActor: Executor updated: > app-20141124023636-0004/3 is now EXITED (Command exited with code 1) > 14/11/24 16:07:43 INFO cluster.SparkDeploySchedulerBackend: Executor > app-20141124023636-0004/3 removed: Command exited with code 1 > > Thanks in advance. > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/issue-while-running-the-code-in-standalone-mode-Initial-job-has-not-accepted-any-resources-check-you-tp19628.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >