In your eclipse, while you create your SparkContext, set the master uri as
shown in the web UI's top left corner like: spark://someIPorHost:7077 and
it should be fine.

Thanks
Best Regards

On Mon, Apr 20, 2015 at 12:22 PM, madhvi <madhvi.gu...@orkash.com> wrote:

> Hi All,
>
> I am new to spark and have installed spark cluster over my system having
> hadoop cluster.I want to process data stored in HDFS through spark.
>
> When I am running code in eclipse it is giving the following warning
> repeatedly:
> scheduler.TaskSchedulerImpl: Initial job has not accepted any resources;
> check your cluster UI to ensure that workers are registered and have
> sufficient resources.
>
> I have made changes to spark-env.sh file as below
> export SPARK_WORKER_INSTANCES=1
> export HADOOP_CONF_DIR=/root/Documents/hadoop/etc/hadoop
> export SPARK_WORKER_MEMORY=1g
> export SPARK_WORKER_CORES=2
> export SPARK_EXECUTOR_MEMORY=1g
>
> I am running the spark standalone cluster.In cluster UI it is showing all
> workers with allocated resources but still its not working.
> what other configurations are needed to be changed?
>
> Thanks
> Madhvi Gupta
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to