Thanks a lot Andrew! Yeah I actually realized that later. I made a silly
mistake here.
On Mon, Oct 20, 2014 at 6:03 PM, Andrew Ash and...@andrewash.com wrote:
Hi Anny, SPARK_WORKER_INSTANCES is the number of copies of spark workers
running on a single box. If you change the number you change how the
hardware you have is split up (useful for breaking large servers into 32GB
heaps each which perform better) but doesn't change the amount of hardware
you have. Because the hardware's the same, you're not going to see huge
performance improvements unless you were in the huge heap scenario.
Typically you should configure the parameters so that SPARK_WORKER_CORES *
SPARK_WORKER_INSTANCES = the number of cores on your machine. If you have
an 8 core box, then you should lower SPARK_WORKER_CORES as you raise
SPARK_WORKER_INSTANCES.
Cheers!
Andrew
On Mon, Oct 20, 2014 at 3:21 PM, anny9699 anny9...@gmail.com wrote:
Hi,
I have a question about the worker_instances setting and worker_cores
setting in aws ec2 cluster. I understand it is a cluster and the default
setting in the cluster is
*SPARK_WORKER_CORES = 8
SPARK_WORKER_INSTANCES = 1*
However after I changed it to
*SPARK_WORKER_CORES = 8
SPARK_WORKER_INSTANCES = 8*
Seems the speed doesn't change very much. Could anyone give an explanation
about this? Maybe more details about work_cores vs worker_instances?
Thanks a lot!
Anny
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/worker-instances-vs-worker-cores-tp16855.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org