[ https://issues.apache.org/jira/browse/SPARK-23857?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16424144#comment-16424144 ]
Apache Spark commented on SPARK-23857: -------------------------------------- User 'skonto' has created a pull request for this issue: https://github.com/apache/spark/pull/20967 > In mesos cluster mode keytab Spark submit requires the keytab to be available > on the local file system. > ------------------------------------------------------------------------------------------------------- > > Key: SPARK-23857 > URL: https://issues.apache.org/jira/browse/SPARK-23857 > Project: Spark > Issue Type: Bug > Components: Mesos > Affects Versions: 2.3.0 > Reporter: Stavros Kontopoulos > Priority: Minor > > Users could submit their jobs from an external to the cluster host which may > not have the required keytab locally (also discussed here). > Moreover, in cluster mode it does not make much sense to reference a local > resource unless this is uploaded/stored somewhere in the cluster. For yarn > HDFS is used, on mesos and certainly on DC/OS right now the secret store is > used for storing secrets and consequently keytabs. There is a check > [here|https://github.com/apache/spark/blob/7cf9fab33457ccc9b2d548f15dd5700d5e8d08ef/core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala#L387] > that makes spark submit difficult to use in such deployment scenarios. > On DC/OS the workaround is to directly submit to the mesos dispatcher rest > api by passing the spark.yarn.tab property pointing to a path within the > driver's container where the keytab will be mounted after its fetched from > the secret store, at container's launch time. Target is to allow spark submit > be flexible enough for mesos in cluster mode, as DC/OS users often want to > deploy using that. -- This message was sent by Atlassian JIRA (v7.6.3#76005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org