Hi all,

I've been trying the 0.9.0-preview1 build on minikube with the spark
interpreter. It's working, but I'm unable to work with any dependencies
that I've added to the spark interpreter.

(Note: I had to add `SPARK_SUBMIT_OPTIONS=--conf spark.jars.ivy=/tmp/.ivy`
and `SPARK_USER=root` to the default interpreter options.)

I'm trying to connect spark to Cassandra, and I've added the following
dependency to the spark interpreter: `
com.datastax.cassandra:cassandra-driver-core:3.9.0`.

In the `zeppelin-server` pod logs, I see this:

```
 INFO [2020-05-13 02:52:15,840] ({Thread-18}
InterpreterSetting.java[run]:953) - Start to download dependencies for
interpreter: spark
 INFO [2020-05-13 02:52:21,565] ({Thread-18}
InterpreterSetting.java[run]:966) - Finish downloading dependencies for
interpreter: spark
 INFO [2020-05-13 02:52:21,565] ({Thread-18}
InterpreterSetting.java[setStatus]:740) - Set interpreter spark status to
READY
```

However, when a run a cell, I don't see any note of dependencies being
downloaded to the actual spark pod, and I get this error:

```
<console>:23: error: object datastax is not a member of package com import
com.datastax.driver.core.Cluster
```

Are dependency downloads supported with zeppelin and spark over kubernetes?
Or am I required to add the dependency jars directly to my spark docker
image and add them to the classpath?

Thanks,
Hetu

Reply via email to