Yeah, that isn't very clear. I'll improve it.
-- Chris Miller On Wed, Mar 9, 2016 at 5:12 PM, mina lee <mina...@apache.org> wrote: > Glad to hear that it works! > Actually, there is document > https://zeppelin.incubator.apache.org/docs/0.5.5-incubating/interpreter/spark.html > in > dependency management section, but obviously it seems hard to find for new > users. So feel free to improve it. > > > On Wed, Mar 9, 2016 at 6:05 PM Chris Miller <cmiller11...@gmail.com> > wrote: > >> Oh, I see. Yeah, that's not documented... no wonder it's confusing. I'll >> open a PR with some improvements to the documentation for this case when I >> have a moment. >> >> Changing spark-default.conf as you suggested indeed worked. Thanks! >> >> >> -- >> Chris Miller >> >> On Wed, Mar 9, 2016 at 10:04 AM, mina lee <mina...@apache.org> wrote: >> >>> Hi Chris, >>> >>> there are several ways to load dependencies to Zeppelin 0.5.5. >>> Using %dep is one of them. >>> If you want do it by setting spark.jars.packages property, proper way of >>> doing it is editing your SPARK_HOME/conf/spark-default.conf >>> and adding below line.(I assume that you set SPARK_HOME in >>> ZEPPELIN_HOME/conf/zeppelin-env.sh) >>> >>> spark.jars.packages org.apache.avro:avro:1.8.0,org. >>> joda:joda-convert:1.8.1 >>> >>> The reason you can import avro dependency is that spark assembly already >>> includes avro dependencies, not because you added it in Zeppelin >>> interpreter setting. >>> >>> You can add dependencies via GUI with the latest master >>> branch(0.6.0-incubating-SNAPSHOT) which is experimental at the moment. >>> Please let me know it answers your question. >>> >>> Regards, >>> Mina >>> >>> On Wed, Mar 9, 2016 at 1:41 AM Chris Miller <cmiller11...@gmail.com> >>> wrote: >>> >>>> Hi, >>>> >>>> I have a strange situation going on. I'm running Zeppelin 0.5.5 and >>>> Spark 1.6.0 (on Amazon EMR). I added this property to the interpreter >>>> settings (and restarted it): >>>> >>>> >>>> spark.jars.packages: org.apache.avro:avro:1.8.0,org.joda:joda-convert:1.8.1 >>>> >>>> The avro dependency loads fine and I'm able to import and use it. >>>> However, if I try to import something in the joda-convert package (such as, >>>> org.joda.convert.FromString), I get an error that "error: object convert is >>>> not a member of package org.joda". >>>> >>>> If I run the spark-shell from the CLI and include the same string above >>>> in the --package parameter, I'm able to import joda-convert just fine. >>>> Also, if I restart the interpreter and manually import the dependency with >>>> z.load(), it also works fine: >>>> >>>> %dep >>>> z.load("org.joda:joda-convert:1.8.1") >>>> >>>> So, what's going on here? >>>> >>>> -- >>>> Chris Miller >>>> >>> >>