Thanks for your reply. Setting it as
--conf spark.executor.cores=1 when I start spark-shell (as an example application) indeed sets the number of cores per executor as 1 (which is 4 before), but I still have 1 executor per worker. What I am really looking for is having 1 worker with 4 executor (each with one core) per machine when I run my application. Based one the documentation it seems it is feasible, but it is not clear as how. Thanks. On Mon, Sep 28, 2015 at 8:46 PM, Jeff Zhang <zjf...@gmail.com> wrote: > use "--executor-cores 1" you will get 4 executors per worker since you > have 4 cores per worker > > > > On Tue, Sep 29, 2015 at 8:24 AM, James Pirz <james.p...@gmail.com> wrote: > >> Hi, >> >> I am using speak 1.5 (standalone mode) on a cluster with 10 nodes while >> each machine has 12GB of RAM and 4 cores. On each machine I have one worker >> which is running one executor that grabs all 4 cores. I am interested to >> check the performance with "one worker but 4 executors per machine - each >> with one core". >> >> I can see that "running multiple executors per worker in Standalone mode" >> is possible based on the closed issue: >> >> https://issues.apache.org/jira/browse/SPARK-1706 >> >> But I can not find a way to do that. "SPARK_EXECUTOR_INSTANCES" is only >> available for the Yarn mode, and in the standalone mode I can just set >> "SPARK_WORKER_INSTANCES" and "SPARK_WORKER_CORES" and "SPARK_WORKER_MEMORY". >> >> Any hint or suggestion would be great. >> >> > > > -- > Best Regards > > Jeff Zhang >