thanks but my use case requires I specify different start and max heap sizes. Looks like spark sets start and max sizes same value.
On Thu, Jul 2, 2015 at 1:08 PM, Todd Nist <tsind...@gmail.com> wrote: > You should use: > > spark.executor.memory > > from the docs <https://spark.apache.org/docs/latest/configuration.html>: > spark.executor.memory512mAmount of memory to use per executor process, in > the same format as JVM memory strings (e.g.512m, 2g). > > -Todd > > > > On Thu, Jul 2, 2015 at 3:36 PM, Mulugeta Mammo <mulugeta.abe...@gmail.com> > wrote: > >> tried that one and it throws error - extraJavaOptions is not allowed to >> alter memory settings, use spakr.executor.memory instead. >> >> On Thu, Jul 2, 2015 at 12:21 PM, Benjamin Fradet < >> benjamin.fra...@gmail.com> wrote: >> >>> Hi, >>> >>> You can set those parameters through the >>> >>> spark.executor.extraJavaOptions >>> >>> Which is documented in the configuration guide: >>> spark.apache.org/docs/latest/configuration.htnl >>> On 2 Jul 2015 9:06 pm, "Mulugeta Mammo" <mulugeta.abe...@gmail.com> >>> wrote: >>> >>>> Hi, >>>> >>>> I'm running Spark 1.4.0, I want to specify the start and max size (-Xms >>>> and Xmx) of the jvm heap size for my executors, I tried: >>>> >>>> executor.cores.memory="-Xms1g -Xms8g" >>>> >>>> but doesn't work. How do I specify? >>>> >>>> Appreciate your help. >>>> >>>> Thanks, >>>> >>>> >> >