spark-env.sh works for me in Spark 1.4 but not spark.executor.extraJavaOptions.
On Sun, Aug 23, 2015 at 11:27 AM Raghavendra Pandey < raghavendra.pan...@gmail.com> wrote: > I think the only way to pass on environment variables to worker node is to > write it in spark-env.sh file on each worker node. > > On Sun, Aug 23, 2015 at 8:16 PM, Hemant Bhanawat <hemant9...@gmail.com> > wrote: > >> Check for spark.driver.extraJavaOptions and >> spark.executor.extraJavaOptions in the following article. I think you can >> use -D to pass system vars: >> >> spark.apache.org/docs/latest/configuration.html#runtime-environment >> Hi, >> >> I am starting a spark streaming job in standalone mode with spark-submit. >> >> Is there a way to make the UNIX environment variables with which >> spark-submit is started available to the processes started on the worker >> nodes? >> >> Jan >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> For additional commands, e-mail: user-h...@spark.apache.org >> >> >