Can anyone help me set memory for standalone cluster?

2014-06-01 Thread Yunmeng Ban
Hi,

I'm running the example of JavaKafkaWordCount in a standalone cluster. I
want to set 1600MB memory for each slave node. I wrote in the
spark/conf/spark-env.sh

SPARK_WORKER_MEMORY=1600m

But the logs on slave nodes looks this:
Spark Executor Command: /usr/java/latest/bin/java -cp
:/~path/spark/conf:/~path/spark/assembly/target/scala-2.10/spark-assembly_2.10-0.9.1-hadoop2.2.0.jar
-Xms512M -Xmx512M
org.apache.spark.executor.CoarseGrainedExecutorBackend

The memory seems to be the default number, not 1600M.
I don't how to make SPARK_WORKER_MEMORY work.
Can anyone help me?
Many thanks in advance.

Yunmeng


Re: Can anyone help me set memory for standalone cluster?

2014-06-01 Thread Aaron Davidson
In addition to setting the Standalone memory, you'll also need to tell your
SparkContext to claim the extra resources. Set spark.executor.memory to
1600m as well. This should be a system property set in SPARK_JAVA_OPTS in
conf/spark-env.sh (in 0.9.1, which you appear to be using) -- e.g.,
export SPARK_JAVA_OPTS=-Dspark.executor.memory=1600mb


On Sun, Jun 1, 2014 at 7:36 PM, Yunmeng Ban banyunm...@gmail.com wrote:

 Hi,

 I'm running the example of JavaKafkaWordCount in a standalone cluster. I
 want to set 1600MB memory for each slave node. I wrote in the
 spark/conf/spark-env.sh

 SPARK_WORKER_MEMORY=1600m

 But the logs on slave nodes looks this:
 Spark Executor Command: /usr/java/latest/bin/java -cp
 :/~path/spark/conf:/~path/spark/assembly/target/scala-2.10/spark-assembly_2.10-0.9.1-hadoop2.2.0.jar
 -Xms512M -Xmx512M
 org.apache.spark.executor.CoarseGrainedExecutorBackend

 The memory seems to be the default number, not 1600M.
 I don't how to make SPARK_WORKER_MEMORY work.
 Can anyone help me?
 Many thanks in advance.

 Yunmeng