Stavros Kontopoulos created SPARK-24429:
-------------------------------------------

             Summary: Add support for spark.driver.extraJavaOptions in cluster 
mode for Spark on K8s
                 Key: SPARK-24429
                 URL: https://issues.apache.org/jira/browse/SPARK-24429
             Project: Spark
          Issue Type: Bug
          Components: Kubernetes
    Affects Versions: 2.4.0
            Reporter: Stavros Kontopoulos


 

Right now in cluster mode only extraJavaOptions targeting the executor are set. 

According to the implementation and the docs:

"In client mode, this config must not be set through the {{SparkConf}} directly 
in your application, because the driver JVM has already started at that point. 
Instead, please set this through the {{--driver-java-options}} command line 
option or in your default properties file."

A typical driver call now is of the form:

"/usr/lib/jvm/java-1.8-openjdk/bin/java -cp /opt/spark/conf/:/opt/spark/jars/* 
-Xmx1g org.apache.spark.deploy.SparkSubmit --deploy-mode client --conf 
spark.driver.bindAddress=9.0.7.116 --properties-file 
/opt/spark/conf/spark.properties --class org.apache.spark.examples.SparkPi 
spark-internal 10000"

Also at the entrypoint.sh file there is no management of the driver's java 
Opts. 

We propose to set an env var to pass the extra java opts to the driver (like in 
the case of the executor), rename the env vars in the container as the one for 
the executor is a bit misleading, and use --driver-java-options to pass the 
required options.

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to