I am a dork please disregard this issue. I did not have the slaves
correctly configured. This error is very misleading


On Tue, Apr 15, 2014 at 11:21 AM, Paul Schooss <paulmscho...@gmail.com>wrote:

> Hello,
>
> Currently I deployed 0.9.1 spark using a new way of starting up spark
>
>     exec start-stop-daemon --start --pidfile /var/run/spark.pid
> --make-pidfile --chuid ${SPARK_USER}:${SPARK_GROUP} --chdir ${SPARK_HOME}
> --exec /usr/bin/java -- -cp ${CLASSPATH}
> -Dcom.sun.management.jmxremote.authenticate=false
> -Dcom.sun.management.jmxremote.ssl=false
> -Dcom.sun.management.jmxremote.port=10111
> -Dspark.akka.logLifecycleEvents=true -Djava.library.path=
> -XX:ReservedCodeCacheSize=512M -XX:+UseCodeCacheFlushing
> -XX:+CMSClassUnloadingEnabled -XX:+UseConcMarkSweepGC
> -Dspark.executor.memory=10G -Xmx10g ${MAIN_CLASS} ${MAIN_CLASS_ARGS}
>
>
> where class path points to the spark jar that we compile with sbt. When I
> try to run a job I receive the following warning
>
> WARN TaskSchedulerImpl: Initial job has not accepted any resources; check
> your cluster UI to ensure that workers are registered and have sufficient
> memory
>
>
> My first question is do I need the entire spark project on disk in order
> to run jobs? Or what else am I doing wrong?
>

Reply via email to