[ https://issues.apache.org/jira/browse/SPARK-25934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16674191#comment-16674191 ]
Apache Spark commented on SPARK-25934: -------------------------------------- User 'mpmolek' has created a pull request for this issue: https://github.com/apache/spark/pull/22937 > Mesos: SPARK_CONF_DIR should not be propogated by spark submit > -------------------------------------------------------------- > > Key: SPARK-25934 > URL: https://issues.apache.org/jira/browse/SPARK-25934 > Project: Spark > Issue Type: Bug > Components: Mesos > Affects Versions: 2.3.2 > Reporter: Matt Molek > Priority: Major > > This is very similar to how SPARK_HOME caused problems for spark on Mesos in > SPARK-12345 > The `spark submit` command is setting spark.mesos.driverEnv.SPARK_CONF_DIR to > whatever the SPARK_CONF_DIR was for the command that submitted the job. > This is doesn't make sense for most mesos situations, and it broke spark for > my team when we upgraded from 2.2.0 to 2.3.2. I haven't tested it but I think > 2.4.0 will have the same issue. > It's preventing spark-env.sh from running because now SPARK_CONF_DIR points > to some non-existent directory, instead of the unpacked spark binary in the > Mesos sandbox like it should. > I'm not that familiar with the spark code base, but I think this could be > fixed by simply adding a `&& k != "SPARK_CONF_DIR"` clause to this filter > statement: > https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/deploy/rest/RestSubmissionClient.scala#L421 -- This message was sent by Atlassian JIRA (v7.6.3#76005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org