Re: how to automatically load jars

2016-02-23 Thread vincent gromakowski
Yes i can see it. Its wird cause i have rebuilt from master yesterday.
Le 23 févr. 2016 11:51 PM, "moon soo Lee"  a écrit :

> I tried with current master branch, but i couldn't reproduce.
> After "3. Start zeppelin", before "4. Run a paragraph", if you go to
> interpreter menu, can you see dependency you have added on GUI?
>
> Thanks,
> moon
>
> On Tue, Feb 23, 2016 at 11:41 AM vincent gromakowski <
> vincent.gromakow...@gmail.com> wrote:
>
>> 1. Stop zeppelin
>> 2. Add a dependency in interpreter.json
>> "dependencies": [
>> {
>>   "groupArtifactVersion":
>> "/./spark-cassandra-connector-1.5.0_2.10.jar",
>>   "local": false,
>>   "exclusions": []
>> }
>>   ]
>> 3. Start zeppelin
>> 4. Run a paragraph with
>> import com.datastax.spark.connector._
>> => error: object datastax is not a member of package com
>> 5. Edit Spark interpreter settings, save it and restart it
>> 6. Re run the paragraph
>> => no error
>>
>> 2016-02-23 19:25 GMT+01:00 moon soo Lee :
>>
>>> interpreter.json supposed to be loaded on launch.
>>> Could double check that interpreter.json is not read at zeppelin launch?
>>> Or if it keep happening, could you let me know how to reproduce?
>>>
>>> Thanks,
>>> moon
>>>
>>>
>>> On Tue, Feb 23, 2016 at 8:22 AM vincent gromakowski <
>>> vincent.gromakow...@gmail.com> wrote:
>>>
 What is the best way to configure spark interpreter ?

 Should I use zeppelin-env.sh and a very long line of "export
 SPARK_SUBMIT_OPTIONS"

 or configure interpreter.json before launching Zeppelin daemon

 It seems interpreter.json is not read at zeppelin launch,I need to
 manually go to settings web UI,edit the spark interpreter and restart it...

 2016-02-23 15:15 GMT+01:00 vincent gromakowski <
 vincent.gromakow...@gmail.com>:

> Hi,
> I am trying to automatcally add jars to spark interpreter with several
> methods but I cannot achieve it.
> I am currently generating an interpreter.json file from ansible
> templates before launching Zeppelin in Marathon.
> 1.  spark.jars
> 2.  spark.driver.extraClassPath
> 3.  groupArtifactVersion  (dependency loading)
>
> In all case I get a class not found exception for the spark cassandra
> connector. The only way to make it works is to go to interpreter settings,
> edit spark settings, then save and restart the interpreter but it's not
> automatic at all as we need to do it each time Zeppelin is started.
>
> Is the interpreter.json file automatically loaded at the start of
> Zeppelin ?
>


>>


Re: how to automatically load jars

2016-02-23 Thread moon soo Lee
I tried with current master branch, but i couldn't reproduce.
After "3. Start zeppelin", before "4. Run a paragraph", if you go to
interpreter menu, can you see dependency you have added on GUI?

Thanks,
moon

On Tue, Feb 23, 2016 at 11:41 AM vincent gromakowski <
vincent.gromakow...@gmail.com> wrote:

> 1. Stop zeppelin
> 2. Add a dependency in interpreter.json
> "dependencies": [
> {
>   "groupArtifactVersion":
> "/./spark-cassandra-connector-1.5.0_2.10.jar",
>   "local": false,
>   "exclusions": []
> }
>   ]
> 3. Start zeppelin
> 4. Run a paragraph with
> import com.datastax.spark.connector._
> => error: object datastax is not a member of package com
> 5. Edit Spark interpreter settings, save it and restart it
> 6. Re run the paragraph
> => no error
>
> 2016-02-23 19:25 GMT+01:00 moon soo Lee :
>
>> interpreter.json supposed to be loaded on launch.
>> Could double check that interpreter.json is not read at zeppelin launch?
>> Or if it keep happening, could you let me know how to reproduce?
>>
>> Thanks,
>> moon
>>
>>
>> On Tue, Feb 23, 2016 at 8:22 AM vincent gromakowski <
>> vincent.gromakow...@gmail.com> wrote:
>>
>>> What is the best way to configure spark interpreter ?
>>>
>>> Should I use zeppelin-env.sh and a very long line of "export
>>> SPARK_SUBMIT_OPTIONS"
>>>
>>> or configure interpreter.json before launching Zeppelin daemon
>>>
>>> It seems interpreter.json is not read at zeppelin launch,I need to
>>> manually go to settings web UI,edit the spark interpreter and restart it...
>>>
>>> 2016-02-23 15:15 GMT+01:00 vincent gromakowski <
>>> vincent.gromakow...@gmail.com>:
>>>
 Hi,
 I am trying to automatcally add jars to spark interpreter with several
 methods but I cannot achieve it.
 I am currently generating an interpreter.json file from ansible
 templates before launching Zeppelin in Marathon.
 1.  spark.jars
 2.  spark.driver.extraClassPath
 3.  groupArtifactVersion  (dependency loading)

 In all case I get a class not found exception for the spark cassandra
 connector. The only way to make it works is to go to interpreter settings,
 edit spark settings, then save and restart the interpreter but it's not
 automatic at all as we need to do it each time Zeppelin is started.

 Is the interpreter.json file automatically loaded at the start of
 Zeppelin ?

>>>
>>>
>


Re: how to automatically load jars

2016-02-23 Thread vincent gromakowski
1. Stop zeppelin
2. Add a dependency in interpreter.json
"dependencies": [
{
  "groupArtifactVersion":
"/./spark-cassandra-connector-1.5.0_2.10.jar",
  "local": false,
  "exclusions": []
}
  ]
3. Start zeppelin
4. Run a paragraph with
import com.datastax.spark.connector._
=> error: object datastax is not a member of package com
5. Edit Spark interpreter settings, save it and restart it
6. Re run the paragraph
=> no error

2016-02-23 19:25 GMT+01:00 moon soo Lee :

> interpreter.json supposed to be loaded on launch.
> Could double check that interpreter.json is not read at zeppelin launch?
> Or if it keep happening, could you let me know how to reproduce?
>
> Thanks,
> moon
>
>
> On Tue, Feb 23, 2016 at 8:22 AM vincent gromakowski <
> vincent.gromakow...@gmail.com> wrote:
>
>> What is the best way to configure spark interpreter ?
>>
>> Should I use zeppelin-env.sh and a very long line of "export
>> SPARK_SUBMIT_OPTIONS"
>>
>> or configure interpreter.json before launching Zeppelin daemon
>>
>> It seems interpreter.json is not read at zeppelin launch,I need to
>> manually go to settings web UI,edit the spark interpreter and restart it...
>>
>> 2016-02-23 15:15 GMT+01:00 vincent gromakowski <
>> vincent.gromakow...@gmail.com>:
>>
>>> Hi,
>>> I am trying to automatcally add jars to spark interpreter with several
>>> methods but I cannot achieve it.
>>> I am currently generating an interpreter.json file from ansible
>>> templates before launching Zeppelin in Marathon.
>>> 1.  spark.jars
>>> 2.  spark.driver.extraClassPath
>>> 3.  groupArtifactVersion  (dependency loading)
>>>
>>> In all case I get a class not found exception for the spark cassandra
>>> connector. The only way to make it works is to go to interpreter settings,
>>> edit spark settings, then save and restart the interpreter but it's not
>>> automatic at all as we need to do it each time Zeppelin is started.
>>>
>>> Is the interpreter.json file automatically loaded at the start of
>>> Zeppelin ?
>>>
>>
>>


Re: how to automatically load jars

2016-02-23 Thread moon soo Lee
interpreter.json supposed to be loaded on launch.
Could double check that interpreter.json is not read at zeppelin launch?
Or if it keep happening, could you let me know how to reproduce?

Thanks,
moon


On Tue, Feb 23, 2016 at 8:22 AM vincent gromakowski <
vincent.gromakow...@gmail.com> wrote:

> What is the best way to configure spark interpreter ?
>
> Should I use zeppelin-env.sh and a very long line of "export
> SPARK_SUBMIT_OPTIONS"
>
> or configure interpreter.json before launching Zeppelin daemon
>
> It seems interpreter.json is not read at zeppelin launch,I need to
> manually go to settings web UI,edit the spark interpreter and restart it...
>
> 2016-02-23 15:15 GMT+01:00 vincent gromakowski <
> vincent.gromakow...@gmail.com>:
>
>> Hi,
>> I am trying to automatcally add jars to spark interpreter with several
>> methods but I cannot achieve it.
>> I am currently generating an interpreter.json file from ansible templates
>> before launching Zeppelin in Marathon.
>> 1.  spark.jars
>> 2.  spark.driver.extraClassPath
>> 3.  groupArtifactVersion  (dependency loading)
>>
>> In all case I get a class not found exception for the spark cassandra
>> connector. The only way to make it works is to go to interpreter settings,
>> edit spark settings, then save and restart the interpreter but it's not
>> automatic at all as we need to do it each time Zeppelin is started.
>>
>> Is the interpreter.json file automatically loaded at the start of
>> Zeppelin ?
>>
>
>


Re: how to automatically load jars

2016-02-23 Thread vincent gromakowski
What is the best way to configure spark interpreter ?

Should I use zeppelin-env.sh and a very long line of "export
SPARK_SUBMIT_OPTIONS"

or configure interpreter.json before launching Zeppelin daemon

It seems interpreter.json is not read at zeppelin launch,I need to manually
go to settings web UI,edit the spark interpreter and restart it...

2016-02-23 15:15 GMT+01:00 vincent gromakowski <
vincent.gromakow...@gmail.com>:

> Hi,
> I am trying to automatcally add jars to spark interpreter with several
> methods but I cannot achieve it.
> I am currently generating an interpreter.json file from ansible templates
> before launching Zeppelin in Marathon.
> 1.  spark.jars
> 2.  spark.driver.extraClassPath
> 3.  groupArtifactVersion  (dependency loading)
>
> In all case I get a class not found exception for the spark cassandra
> connector. The only way to make it works is to go to interpreter settings,
> edit spark settings, then save and restart the interpreter but it's not
> automatic at all as we need to do it each time Zeppelin is started.
>
> Is the interpreter.json file automatically loaded at the start of Zeppelin
> ?
>