Jerzy J. Gangi created ZEPPELIN-1741:
----------------------------------------
Summary: JAR's specified with spark.jars in spark-defaults.conf
does not affect %pyspark interpreter
Key: ZEPPELIN-1741
URL: https://issues.apache.org/jira/browse/ZEPPELIN-1741
Project: Zeppelin
Issue Type: Bug
Components: Interpreters
Affects Versions: 0.6.2
Environment: Zeppelin 0.6.2
Spark 1.6.2
Reporter: Jerzy J. Gangi
If you specify JAR's with `spark.jars` in your `spark-defaults.conf`, the
`%pyspark` interpreter will not load these JARs.
Currently, a note in the Spark interpreter documentation says, "Note that
adding jar to pyspark is only availabe via `%dep` interpreter at the moment."
This is undesirable for two reasons:
1) `%dep` is deprecated
2) Sysadmins configure `spark-defaults.conf`, and expect these settings to be
honored however Spark is executed.
As a Zeppelin user, I expect that if I configure JAR's in
`spark-defaults.conf`, these JAR's will be available when the %pyspark
interpreter is run.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)