Thanks Li, Inread through the ticket, be able to pass pod YAML file would be amazing.
Do you have any target date for production or incubator? I really want to try out this feature. On Mon, Jul 9, 2018 at 4:48 PM Yinan Li <liyinan...@gmail.com> wrote: > Spark on k8s currently doesn't support specifying a custom SecurityContext > of the driver/executor pods. This will be supported by the solution to > https://issues.apache.org/jira/browse/SPARK-24434. > > On Mon, Jul 9, 2018 at 2:06 PM trung kien <kient...@gmail.com> wrote: > >> Dear all, >> >> Is there any way to includes security context ( >> https://kubernetes.io/docs/tasks/configure-pod-container/security-context/) >> when submitting job through k8s servers? >> >> I'm trying to first spark jobs on Kubernetes through spark-submit: >> >> bin/spark-submit --master k8s://https://API_SERVERS --deploy-mode >> cluster --name spark-pi --class org.apache.spark.examples.SparkPi --conf >> spark.kubernetes.namespace=NAMESPACE --conf spark.executor.instances=3 >> --conf spark.kubernetes.container.image=<SPARK_IMAGE> --conf >> spark.kubernetes.driver.pod.name=spark-pi-driver >> local:///opt/spark/examples/jars/spark-examples_2.11-2.3.1.jar >> >> But the job was rejected because the pod (created by spark-submit) >> doesn't have security context to run as my account (Our policy doesn't >> allow us to runAsUser root) >> >> I check the code under KubernetesClientApplication.scala >> <https://github.com/apache/spark/blob/master/resource-managers/kubernetes/core/src/main/scala/org/apache/spark/deploy/k8s/submit/KubernetesClientApplication.scala>, >> it doesn't seems to support security context by configuration. >> >> Is there any solution to get arround this issue? is there any patch that >> support this? >> >> -- >> Thanks >> Kien >> > -- Thanks Kien