Thanks. I'll file a Jira. Looks like custom repos aren't loading either from either source. I'll try to replicate as well.
Mike On Tue, Aug 16, 2016 at 5:55 PM Mina Lee <[email protected]> wrote: > I also could reproduce it with Spark 2.0.0, but not with Spark 1.6.1. > If you want to use Zeppelin with Spark 2.0, one alternative you can try is > using [1] "dependencies" in GUI interpreter menu. > > [1] http://zeppelin.apache.org/docs/0.6.1/manual/dependencymanagement.html > > On Wed, Aug 17, 2016 at 1:46 AM Jeff Zhang <[email protected]> wrote: > >> I can reproduce it in 0.6.1 & master branch, please file a ticket for >> that. >> >> On Wed, Aug 17, 2016 at 4:09 AM, Michael Sells <[email protected]> wrote: >> >>> Testing out 0.6.1 with Spark 2.0 and discovered the way we load >>> dependencies doesn't seem to be working with the new update. >>> >>> We pass new dependencies in via a SPARK_SUBMIT_OPTIONS environment >>> variable pass the following flags: >>> --packages com.databricks:spark-avro_2.11:3.0.0 >>> >>> Now when I try to import it with: >>> import com.databricks.spark.avro._ >>> >>> I get: >>> <console>:25: error: object databricks is not a member of package com >>> import com.databricks.spark.avro._ >>> >>> I checked the logs are there is no error retrieving the package. So it >>> seems to be something with the classpath. >>> >>> This works in 0.6.0. Any idea if something changed or if we're doing >>> something wrong? I tried this with a few internal packages as well and it >>> doesn't work with those either. >>> >>> Thanks, >>> Mike >>> >>> >>> >>> >>> >>> >> >> >> -- >> Best Regards >> >> Jeff Zhang >> >
