Yes i can see it. Its wird cause i have rebuilt from master yesterday.
Le 23 févr. 2016 11:51 PM, "moon soo Lee" <m...@apache.org> a écrit :

> I tried with current master branch, but i couldn't reproduce.
> After "3. Start zeppelin", before "4. Run a paragraph", if you go to
> interpreter menu, can you see dependency you have added on GUI?
>
> Thanks,
> moon
>
> On Tue, Feb 23, 2016 at 11:41 AM vincent gromakowski <
> vincent.gromakow...@gmail.com> wrote:
>
>> 1. Stop zeppelin
>> 2. Add a dependency in interpreter.json
>> "dependencies": [
>>         {
>>           "groupArtifactVersion":
>> "/........./spark-cassandra-connector-1.5.0_2.10.jar",
>>           "local": false,
>>           "exclusions": []
>>         }
>>       ]
>> 3. Start zeppelin
>> 4. Run a paragraph with
>> import com.datastax.spark.connector._
>> => error: object datastax is not a member of package com
>> 5. Edit Spark interpreter settings, save it and restart it
>> 6. Re run the paragraph
>> => no error
>>
>> 2016-02-23 19:25 GMT+01:00 moon soo Lee <m...@apache.org>:
>>
>>> interpreter.json supposed to be loaded on launch.
>>> Could double check that interpreter.json is not read at zeppelin launch?
>>> Or if it keep happening, could you let me know how to reproduce?
>>>
>>> Thanks,
>>> moon
>>>
>>>
>>> On Tue, Feb 23, 2016 at 8:22 AM vincent gromakowski <
>>> vincent.gromakow...@gmail.com> wrote:
>>>
>>>> What is the best way to configure spark interpreter ?
>>>>
>>>> Should I use zeppelin-env.sh and a very long line of "export
>>>> SPARK_SUBMIT_OPTIONS"
>>>>
>>>> or configure interpreter.json before launching Zeppelin daemon
>>>>
>>>> It seems interpreter.json is not read at zeppelin launch,I need to
>>>> manually go to settings web UI,edit the spark interpreter and restart it...
>>>>
>>>> 2016-02-23 15:15 GMT+01:00 vincent gromakowski <
>>>> vincent.gromakow...@gmail.com>:
>>>>
>>>>> Hi,
>>>>> I am trying to automatcally add jars to spark interpreter with several
>>>>> methods but I cannot achieve it.
>>>>> I am currently generating an interpreter.json file from ansible
>>>>> templates before launching Zeppelin in Marathon.
>>>>> 1.  spark.jars
>>>>> 2.  spark.driver.extraClassPath
>>>>> 3.  groupArtifactVersion  (dependency loading)
>>>>>
>>>>> In all case I get a class not found exception for the spark cassandra
>>>>> connector. The only way to make it works is to go to interpreter settings,
>>>>> edit spark settings, then save and restart the interpreter but it's not
>>>>> automatic at all as we need to do it each time Zeppelin is started.
>>>>>
>>>>> Is the interpreter.json file automatically loaded at the start of
>>>>> Zeppelin ?
>>>>>
>>>>
>>>>
>>

Reply via email to