Hi all,
I'm setting up a 2 nodes (physical) GlusterFS cluster (3.5 latest nightly) 
where each node acts also as an oVirt host (3.4 latest nightly with self-hosted 
engine; GlusterFS access through NFS/FUSE while waiting for some bug fixing 
etc.).

Since I already properly configured fencing (power management) in oVirt, I'm 
currently configuring each GlusterFS volume with:

    gluster volume set VOLUMENAME cluster.server-quorum-type none
    gluster volume set VOLUMENAME cluster.quorum-type none

since I need the surviving node to remain up and responsive (without excessive 
"delays" at single-node-failure time) in case of a single node failure.

Limited testing has shown that the "surviving" node somewhat "halts" (but I did 
not try waiting for more than few minutes, maybe less) if, for example, I put 
the other node in maintenance through oVirt and reboot it.

Is this the proper way of achieving what I need?

Many thanks in advance for any hint/suggestion/docs-to-read.

Regards,
Giuseppe

PS: I posted the same question some time ago on gluster-devel but then realized 
it's not really devel related...
                                          
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to