I apologize in advance if this is too much of a newbie question but I'm
having a hard time running any Mahout example code in a distributed Spark
cluster. The code runs as advertised when Spark is running locally on one
machine but the minute I point Spark to a cluster and master url, I can't
get it to work, drawing the error: "WARN scheduler.TaskSchedulerImpl:
Initial job has not accepted any resources; check your cluster UI to ensure
that workers are registered and have sufficient memory"

I know my Spark cluster is configured and working correctly because I ran
non-Mahout code and it runs on a distributed cluster fine. What am I doing
wrong? The only thing I can think of is that my Spark version is too recent
-- 2.1.1 -- for the Mahout version I'm using -- 0.13.0. Is that it or am I
doing something else wrong?

Thanks for any advice,
Hoa

Reply via email to