just as Marcelo Vanzin said there are two possible reasons for this problem.
I solved reason2 several days ago.

my process is, ssh to one of the worker node, read its log output , find a
line that says
"Remoting started"
after that line their should be some line of "connecting to xxxxx"
MAKE SURE worker node can really connect to the designated host

my problem is caused by hostname misconfiguration, after fix that, problem
is solved.

the error message complaining resource allocation is really misleading in
this case.


On Tue, Aug 12, 2014 at 9:51 AM, ldmtwo <ldm...@gmail.com> wrote:

> I see this error too. I have never found a fix and I've been working on
> this
> for a few months.
>
> For me, I have 4 nodes with 46GB and 8 cores each. If I change the executor
> to use 8GB, if fails. If I use 6GB, it works. I request 2 cores only. On
> another cluster, I have different limits.  My workload is extremely memory
> intensive and I can't even get the smaller loads to run.
>
> Every "solution" says that we have too few cores or RAM, but they are
> wrong.
> Something is either misleading or not working. I'm using 1.0.1 and 1.0.2.
>
> I have checked all nodes and see plenty of free RAM. The driver/master node
> will run and do it's data loading and processing, but the executors never
> start up, attach, connect or w/e to do the real work.
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/Initial-job-has-not-accepted-any-resources-tp11668p11938.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to