Thanks for the reply. Yeah I found the same doc and am able to use multiple
cores in spark-shell, however, when I use pyspark, it appears only to use
one core, I am wondering if this is something I did't configure correctly
or something supported in pyspark.

On Fri, Mar 24, 2017 at 3:52 PM, Kadam, Gangadhar (GE Aviation, Non-GE) <
gangadhar.ka...@ge.com> wrote:

> In Local Mode  all processes are executed inside a single JVM.
> Application is started in a local mode by setting master to local,
> local[*] or local[n].
> spark.executor.cores and spark.executor.cores are not applicable in the
> local mode because there is only one embedded executor.
>
>
> In Standalone mode, you need  standalone Spark cluster<
> https://spark.apache.org/docs/latest/spark-standalone.html>.
>
> It requires a master node (can be started using
> SPARK_HOME/sbin/start-master.sh script) and at least one worker node (can
> be started using SPARK_HOME/sbin/start-slave.sh script).SparkConf should
> use master node address to create (spark://host:port)
>
> Thanks!
>
> Gangadhar
> From: Li Jin <ice.xell...@gmail.com<mailto:ice.xell...@gmail.com>>
> Date: Friday, March 24, 2017 at 3:43 PM
> To: "user@spark.apache.org<mailto:user@spark.apache.org>" <
> user@spark.apache.org<mailto:user@spark.apache.org>>
> Subject: EXT: Multiple cores/executors in Pyspark standalone mode
>
> Hi,
>
> I am wondering does pyspark standalone (local) mode support multi
> cores/executors?
>
> Thanks,
> Li
>

Reply via email to