You are correct. If you are just using spark-shell in local mode (i.e., without cluster), you can set the SPARK_MEM environment variable to give the driver more memory. E.g.: SPARK_MEM=24g ./spark-shell
Otherwise, if you're using a real cluster, the driver shouldn't require a significant amount of memory, so SPARK_MEM should not have to be used. On Tue, Oct 29, 2013 at 12:40 PM, Soumya Simanta <soumya.sima...@gmail.com>wrote: > I'm new to Spark. I want to try out a few simple example from the Spark > shell. However, I'm not sure how to configure it so that I can make the > max. use of memory on my workers. > > On average I've around 48 GB of RAM on each node on my cluster. I've > around 10 nodes. > > Based on the documentation I could find memory based configuration in two > places. > > *1. $SPARK_INSTALL_DIR/dist/conf/spark-env.sh * > > *SPARK_WORKER_MEMORY* Total amount of memory to allow Spark applications > to use on the machine, e.g. 1000m, 2g (default: total memory minus 1 GB); > note that each application's *individual* memory is configured using its > spark.executor.memory property. > *2. spark.executor.memory JVM flag. * > spark.executor.memory512m Amount of memory to use per executor process, > in the same format as JVM memory strings (e.g. 512m, 2g). > > http://spark.incubator.apache.org/docs/latest/configuration.html#system-properties > > In my case I want to use the max. memory possible on each node. My > understanding is that I don't have to change *SPARK_WORKER_MEMORY *and I > will have to increase spark.executor.memory to something big (e.g., 24g or > 32g). Is this correct? If yes, what is the correct way of setting this > property if I just want to use the spark-shell. > > > Thanks. > -Soumya > >