Github user echarles commented on the issue:
https://github.com/apache/zeppelin/pull/2637
@matyix I've given a try on you last commit and can not get the additional
deps (in settings page) working.
I don't see the `spark.jars` property in the generate command (by
interpreter.sh):
```
/opt/spark/bin/spark-submit --class
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer
--driver-class-path
":/opt/zeppelin/interpreter/spark/*:/opt/zeppelin/lib/interpreter/*::/opt/zeppelin/interpreter/spark/zeppelin-spark_2.11-0.8.0-SNAPSHOT.jar:/etc/hdfs-k8s/conf"
--driver-java-options " -Dfile.encoding=UTF-8
-Dlog4j.configuration=file:///opt/zeppelin/conf/log4j.properties
-Dzeppelin.log.file=/opt/zeppelin/logs/zeppelin-interpreter---zeppelin-k8s-hdfs-locality-zeppelin-7cd554b49d-dpq2k.log"
--master k8s://https://kubernetes:443 --conf spark.cores.max='1' --conf
spark.shuffle.service.enabled='false' --conf
spark.yarn.dist.archives=/opt/spark/R/lib/sparkr.zip --conf
spark.executor.instances='3' --conf spark.sql.catalogImplementation='in-memory'
--conf spark.app.name='zeppelin-k8s-spark' --conf spark.executor.memory='1g'
--conf spark.master='k8s://https://kubernetes:443' --conf
spark.kubernetes.namespace='default' --conf
spark.kubernetes.executor.docker.image='datalaye
r/spark-k8s-executor:2.2.0-0.5.0' --conf
spark.kubernetes.driver.docker.image='datalayer/spark-k8s-driver:2.2.0-0.5.0'
--conf
spark.kubernetes.initcontainer.docker.image='datalayer/spark-k8s-init:2.2.0-0.5.0'
--conf spark.kubernetes.resourceStagingServer.uri='http://10.108.197.6:10000'
--conf
```
---