bq. Started SparkUI at http://192.168.2.103:4040 bq. Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources
Can you check UI ? Thanks On Thu, Mar 10, 2016 at 6:57 AM, Shams ul Haque <sham...@cashcare.in> wrote: > Hi, > > *Release of Spark:* 1.6.0, i downloaded it and made a built using > 'sbt/sbt assembly' > > *command for submitting your app: *bin/spark-submit --master > spark://shams-machine:7077 --executor-cores 2 --class > in.myapp.email.combiner.CombinerRealtime > /opt/dev/workspace-luna/combiner_spark/target/combiner-0.0.1-SNAPSHOT.jar > 2>&1 & > > *code snippet of your app: *i developed a lot chained transormations and > connected with Kafka, MongoDB, Cassandra. But tested all of them using > *local[2] > *setting in *conf.setMaster *method. Everything is working there. > > *pastebin of log:* http://pastebin.com/0LjTWLfm > > > Thanks > Shams > > On Thu, Mar 10, 2016 at 8:11 PM, Ted Yu <yuzhih...@gmail.com> wrote: > >> Can you provide a bit more information ? >> >> Release of Spark >> command for submitting your app >> code snippet of your app >> pastebin of log >> >> Thanks >> >> On Thu, Mar 10, 2016 at 6:32 AM, Shams ul Haque <sham...@cashcare.in> >> wrote: >> >>> Hi, >>> >>> I have developed a spark realtime app and started spark-standalone on my >>> laptop. But when i tried to submit that app in Spark it is always >>> in WAITING state & Cores is always Zero. >>> >>> I have set: >>> export SPARK_WORKER_CORES="2" >>> export SPARK_EXECUTOR_CORES="1" >>> >>> in spark-env.sh, but still nothing happend. And same log entry in: >>> *TaskSchedulerImpl:70 - Initial job has not accepted any resources* >>> >>> So, does i need a seperate machine for all this? >>> >>> Please help me to sort that out. >>> >>> Thanks >>> Shams >>> >> >> >