For 1 marathon on mesos restart zeppelin daemon In case of failure.
For 2 glusterfs fuse mount allows to share notebooks on all mesos nodes.
For 3 not available right now In our  design but a manual restart In
zeppelin config page is acceptable for US.
Le 6 avr. 2016 8:18 AM, "Eran Witkon" <eranwit...@gmail.com> a écrit :

> Yes this is correct.
> For HA disk, if you don't have HA storage and no access to S3 then AFAIK
> you don't have other option at the moment.
> If you like to save notebooks to elastic then I suggest you look at the
> storage interface and implementation for git and s3 and implement that
> yourself. It does sound like an interesting feature
> Best
> Eran
> On Wed, 6 Apr 2016 at 08:57 ashish rawat <dceash...@gmail.com> wrote:
>
>> Thanks Eran. So 3, seems to be something external to Zeppelin, and
>> hopefully 1 only means running "zeppelin-daemon.sh start" on a slave
>> machine, when master become inaccessible. Is that correct?
>>
>> My main concern still remains on the storage front. And I don't really
>> have high availability disks or even hdfs in my setup. I have been using
>> elastic search cluster for data high availability, but was hoping that
>> zeppelin can save notebooks to a Elastic Search (like kibana) or maybe a
>> document store.
>>
>> Any idea if anything is planned in that direction. Don't want to fallback
>> to 'rsync' like options.
>>
>> Regards,
>> Ashish
>>
>>
>>
>> On Tue, Apr 5, 2016 at 11:17 PM, Eran Witkon <eranwit...@gmail.com>
>> wrote:
>>
>>> For 1 you need to have both zeppelin web HA and zeppelin deamon HA
>>> For 2 I guess you can use HDFS if you implement the storage interface
>>> for HDFS. But i am not sure.
>>> For 3 I mean that if you connect to an external cluster for example a
>>> spark cluster you need to make sure your spark cluster is HA. Otherwise you
>>> will have zeppelin running but your notebook will fail as no spark cluster
>>> available.
>>> HTH
>>> Eran
>>>
>>>
>>> On Tue, 5 Apr 2016 at 20:20 ashish rawat <dceash...@gmail.com> wrote:
>>>
>>>> Thanks Eran for your reply.
>>>> For 1) I am assuming that it would similar to HA of any other web
>>>> application, i.e. running multiple instances and switching to the backup
>>>> server when master is down, is it not the case?
>>>> For 2) is it also possible to save it on hdfs?
>>>> Can you please explain 3, are you referring to interpreter config? If I
>>>> am using Spark interpreter and submitting jobs to it, and if zeppelin
>>>> master node goes down, then what could be the problem in slave node
>>>> pointing to the same cluster and submitting jobs?
>>>>
>>>> Regards,
>>>> Ashish
>>>>
>>>> On Tue, Apr 5, 2016 at 10:08 PM, Eran Witkon <eranwit...@gmail.com>
>>>> wrote:
>>>>
>>>>> I would say you need to account for these things
>>>>> 1) availability of the zeppelin deamon
>>>>> 2) availability of the notebookd files
>>>>> 3) availability of the interpreters used.
>>>>>
>>>>> For 1 i don't know of out-of-box solution
>>>>> For 2 any ha storage will do, s3 or any ha external mounted disk
>>>>> For 3 it is up to the interpreter and your big data ha solution
>>>>>
>>>>> On Tue, 5 Apr 2016 at 19:29 ashish rawat <dceash...@gmail.com> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> Is there a suggested architecture to run Zeppelin in high
>>>>>> availability mode. The only option I could find was by saving notebooks 
>>>>>> to
>>>>>> S3. Are there any options if one is not using AWS?
>>>>>>
>>>>>> Regards,
>>>>>> Ashish
>>>>>>
>>>>>
>>>>
>>

Reply via email to