bq. Started SparkUI at http://192.168.2.103:4040
bq. Initial job has not accepted any resources; check your cluster UI to
ensure that workers are registered and have sufficient resources

Can you check UI ?

Thanks

On Thu, Mar 10, 2016 at 6:57 AM, Shams ul Haque <sham...@cashcare.in> wrote:

> Hi,
>
> *Release of Spark:* 1.6.0, i downloaded it and made a built using
> 'sbt/sbt assembly'
>
> *command for submitting your app: *bin/spark-submit --master
> spark://shams-machine:7077 --executor-cores 2 --class
> in.myapp.email.combiner.CombinerRealtime
> /opt/dev/workspace-luna/combiner_spark/target/combiner-0.0.1-SNAPSHOT.jar
> 2>&1 &
>
> *code snippet of your app: *i developed a lot chained transormations and
> connected with Kafka, MongoDB, Cassandra. But tested all of them using 
> *local[2]
> *setting in *conf.setMaster *method. Everything is working there.
>
> *pastebin of log:* http://pastebin.com/0LjTWLfm
>
>
> Thanks
> Shams
>
> On Thu, Mar 10, 2016 at 8:11 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>
>> Can you provide a bit more information ?
>>
>> Release of Spark
>> command for submitting your app
>> code snippet of your app
>> pastebin of log
>>
>> Thanks
>>
>> On Thu, Mar 10, 2016 at 6:32 AM, Shams ul Haque <sham...@cashcare.in>
>> wrote:
>>
>>> Hi,
>>>
>>> I have developed a spark realtime app and started spark-standalone on my
>>> laptop. But when i tried to submit that app in Spark it is always
>>> in WAITING state & Cores is always Zero.
>>>
>>> I have set:
>>> export SPARK_WORKER_CORES="2"
>>> export SPARK_EXECUTOR_CORES="1"
>>>
>>> in spark-env.sh, but still nothing happend. And same log entry in:
>>> *TaskSchedulerImpl:70 - Initial job has not accepted any resources*
>>>
>>> So, does i need a seperate machine for all this?
>>>
>>> Please help me to sort that out.
>>>
>>> Thanks
>>> Shams
>>>
>>
>>
>

Reply via email to