Using it for 3 months without any incident
Le 8 avr. 2016 9:09 AM, "ashish rawat" <dceash...@gmail.com> a écrit :

> Sounds great. How long have you been using glusterfs in prod? and have you
> encountered any challenges. The only difficulty for me to use it, would be
> a lack of expertise to fix broken things, so hope it's stability isn't
> something to be concerned about.
>
> Regards,
> Ashish
>
> On Fri, Apr 8, 2016 at 12:20 PM, vincent gromakowski <
> vincent.gromakow...@gmail.com> wrote:
>
>> use fuse interface. Gluster volume is directly accessible as local
>> storage on all nodes but performance is only 200 Mb/s. More than enough for
>> notebooks. For data prefer tachyon/alluxio on top of gluster...
>> Le 8 avr. 2016 6:35 AM, "ashish rawat" <dceash...@gmail.com> a écrit :
>>
>>> Thanks Eran and Vincent.
>>> Eran, I would definitely like to try it out, since it won't add to the
>>> complexity of my deployment. Would see the S3 implementation, to figure out
>>> how complex it would be.
>>>
>>> Vincent,
>>> I haven't explored glusterfs at all. Would it also require to write an
>>> implementation of storage interface? Or zeppelin can work with it, out of
>>> the box?
>>>
>>> Regards,
>>> Ashish
>>>
>>> On Wed, Apr 6, 2016 at 12:53 PM, vincent gromakowski <
>>> vincent.gromakow...@gmail.com> wrote:
>>>
>>>> For 1 marathon on mesos restart zeppelin daemon In case of failure.
>>>> For 2 glusterfs fuse mount allows to share notebooks on all mesos nodes.
>>>> For 3 not available right now In our  design but a manual restart In
>>>> zeppelin config page is acceptable for US.
>>>> Le 6 avr. 2016 8:18 AM, "Eran Witkon" <eranwit...@gmail.com> a écrit :
>>>>
>>>>> Yes this is correct.
>>>>> For HA disk, if you don't have HA storage and no access to S3 then
>>>>> AFAIK you don't have other option at the moment.
>>>>> If you like to save notebooks to elastic then I suggest you look at
>>>>> the storage interface and implementation for git and s3 and implement that
>>>>> yourself. It does sound like an interesting feature
>>>>> Best
>>>>> Eran
>>>>> On Wed, 6 Apr 2016 at 08:57 ashish rawat <dceash...@gmail.com> wrote:
>>>>>
>>>>>> Thanks Eran. So 3, seems to be something external to Zeppelin, and
>>>>>> hopefully 1 only means running "zeppelin-daemon.sh start" on a slave
>>>>>> machine, when master become inaccessible. Is that correct?
>>>>>>
>>>>>> My main concern still remains on the storage front. And I don't
>>>>>> really have high availability disks or even hdfs in my setup. I have been
>>>>>> using elastic search cluster for data high availability, but was hoping
>>>>>> that zeppelin can save notebooks to a Elastic Search (like kibana) or 
>>>>>> maybe
>>>>>> a document store.
>>>>>>
>>>>>> Any idea if anything is planned in that direction. Don't want to
>>>>>> fallback to 'rsync' like options.
>>>>>>
>>>>>> Regards,
>>>>>> Ashish
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, Apr 5, 2016 at 11:17 PM, Eran Witkon <eranwit...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> For 1 you need to have both zeppelin web HA and zeppelin deamon HA
>>>>>>> For 2 I guess you can use HDFS if you implement the storage
>>>>>>> interface for HDFS. But i am not sure.
>>>>>>> For 3 I mean that if you connect to an external cluster for example
>>>>>>> a spark cluster you need to make sure your spark cluster is HA. 
>>>>>>> Otherwise
>>>>>>> you will have zeppelin running but your notebook will fail as no spark
>>>>>>> cluster available.
>>>>>>> HTH
>>>>>>> Eran
>>>>>>>
>>>>>>>
>>>>>>> On Tue, 5 Apr 2016 at 20:20 ashish rawat <dceash...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Thanks Eran for your reply.
>>>>>>>> For 1) I am assuming that it would similar to HA of any other web
>>>>>>>> application, i.e. running multiple instances and switching to the 
>>>>>>>> backup
>>>>>>>> server when master is down, is it not the case?
>>>>>>>> For 2) is it also possible to save it on hdfs?
>>>>>>>> Can you please explain 3, are you referring to interpreter config?
>>>>>>>> If I am using Spark interpreter and submitting jobs to it, and if 
>>>>>>>> zeppelin
>>>>>>>> master node goes down, then what could be the problem in slave node
>>>>>>>> pointing to the same cluster and submitting jobs?
>>>>>>>>
>>>>>>>> Regards,
>>>>>>>> Ashish
>>>>>>>>
>>>>>>>> On Tue, Apr 5, 2016 at 10:08 PM, Eran Witkon <eranwit...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> I would say you need to account for these things
>>>>>>>>> 1) availability of the zeppelin deamon
>>>>>>>>> 2) availability of the notebookd files
>>>>>>>>> 3) availability of the interpreters used.
>>>>>>>>>
>>>>>>>>> For 1 i don't know of out-of-box solution
>>>>>>>>> For 2 any ha storage will do, s3 or any ha external mounted disk
>>>>>>>>> For 3 it is up to the interpreter and your big data ha solution
>>>>>>>>>
>>>>>>>>> On Tue, 5 Apr 2016 at 19:29 ashish rawat <dceash...@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Hi,
>>>>>>>>>>
>>>>>>>>>> Is there a suggested architecture to run Zeppelin in high
>>>>>>>>>> availability mode. The only option I could find was by saving 
>>>>>>>>>> notebooks to
>>>>>>>>>> S3. Are there any options if one is not using AWS?
>>>>>>>>>>
>>>>>>>>>> Regards,
>>>>>>>>>> Ashish
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>
>>>
>

Reply via email to