Barak,
The SPARK_WORKER_MEMORYsetting is used for allocating memory to executors.
You can use SPARK_DAEMON_MEMORY to set memory for the worker JVM.
Mohammed
From: Barak Yaish [mailto:barak.ya...@gmail.com]
Sent: Wednesday, January 13, 2016 12:59 AM
To: user@spark.apache.org
Subject: Spark ignores SPARK_WORKER_MEMORY?
Hello,
Although I'm setting SPARK_WORKER_MEMORY in spark-env.sh, looks like this
setting is ignored. I can't find any indication at the scripts under bin/sbin
that -Xms/-Xmx are set.
If I ps the worker pid, it looks like memory set to 1G:
[hadoop@sl-env1-hadoop1 spark-1.5.2-bin-hadoop2.6]$ ps -ef | grep 20232
hadoop 20232 1 0 02:01 ?00:00:22 /usr/java/latest//bin/java -cp
/workspace/3rd-party/spark/spark-1.5.2-bin-hadoop2.6/sbin/../conf/:/workspace/3rd-party/spark/spark-1.5.2-bin-hadoop2.6/lib/spark-assembly-1.5.2-hadoop2.6.0.jar:/workspace/3rd-party/spark/spark-1.5.2-bin-hadoop2.6/lib/datanucleus-api-jdo-3.2.6.jar:/workspace/3rd-party/spark/spark-1.5.2-bin-hadoop2.6/lib/datanucleus-rdbms-3.2.9.jar:/workspace/3rd-party/spark/spark-1.5.2-bin-hadoop2.6/lib/datanucleus-core-3.2.10.jar:/workspace/3rd-party/hadoop/2.6.3//etc/hadoop/
-Xms1g -Xmx1g org.apache.spark.deploy.worker.Worker --webui-port 8081
spark://10.52.39.92:7077<http://10.52.39.92:7077>
Am I missing something?
Thanks.