[ https://issues.apache.org/jira/browse/SPARK-34870?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Cory Maklin updated SPARK-34870: -------------------------------- Description: When Spark is run in local mode, it works as expected. However, when Spark is run in client mode, it copies the jars to the executor ($SPARK_HOME/work/<app id>/<executor id>), but never adds them to the classpath. It might be worth noting that `spark.jars` does add the jars to the classpath, but unlike `spark.jars.packages` it doesn't automatically download the jar's compiled dependencies. ``` spark = SparkSession.builder\ .master(SPARK_MASTER)\ .appName(APP_NAME)\ ... .config("spark.jars.packages", DEPENDENCY_PACKAGES) \ ... .getOrCreate() ``` was: When Spark is run in local mode, it works as expected. However, when Spark is run in client mode, it copies the jars to the executor ($SPARK_HOME/work/<app id>/<executor id>), but never adds them to the classpath. It might be worth noting that `spark.jars` does add the jars to the classpath, but unlike `spark.jars.packages` it doesn't automatically download the jar's compiled dependencies. ``` spark = SparkSession.builder\ .master(SPARK_MASTER)\ .appName(APP_NAME)\ ... .config("spark.jars.packages", DEPENDENCY_PACKAGES) \ ... .getOrCreate() ``` > Jars downloaded with the --packages argument are not added to the classpath > for executors. > ------------------------------------------------------------------------------------------ > > Key: SPARK-34870 > URL: https://issues.apache.org/jira/browse/SPARK-34870 > Project: Spark > Issue Type: Bug > Components: Spark Submit > Affects Versions: 3.0.0 > Environment: Spark worker running inside a Kubernetes pod with a > Bitnami Spark image, and the driver running inside of a Jupyter Spark > Kubernetes pod. > Reporter: Cory Maklin > Priority: Major > > When Spark is run in local mode, it works as expected. However, when Spark is > run in client mode, it copies the jars to the executor ($SPARK_HOME/work/<app > id>/<executor id>), but never adds them to the classpath. > It might be worth noting that `spark.jars` does add the jars to the > classpath, but unlike `spark.jars.packages` it doesn't automatically download > the jar's compiled dependencies. > > ``` > spark = SparkSession.builder\ > .master(SPARK_MASTER)\ > .appName(APP_NAME)\ > ... > .config("spark.jars.packages", DEPENDENCY_PACKAGES) \ > ... > .getOrCreate() > ``` > -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org