On Wed, Oct 16, 2013 at 1:01 PM, Itamar Heim wrote: > Hope to meet many next week in LinuxCon Europe/KVM Forum/oVirt conference in > Edinburgh > Unfortunately not me ;-(
> - LINBIT published "High-Availability oVirt-Cluster with > iSCSI-Storage"[9] > > [9] note: link requires registeration: > http://www.linbit.com/en/company/news/333-high-available-virtualization-at-a-most-reasonable-price > the pdf is: > http://www.linbit.com/en/downloads/tech-guides?download=68:high-availability-ovirt-cluster-with-iscsi-storage About 3 years ago I successfully tested plain Qemu/KVM with two nodes and drbd dual-primary with live migration on Fedora, so I know it works very well. I think the LINBIT guide gives many suggestions, one of them is in particular just another way to have a High Available oVirt Engine for example, that is a hot topic lately Deep attention is to be put in fencing mechanism though, as always it is in general when talking about drbd, pacemaker, oVirt. I don't know it quite well, but the drbd proxy feature (closed source and not free) could also be then an option for DR in small environments I'm going to experiment the solution of the paper (possibly using clusterlabs.org repo for cluster and so using corosync instead of heartbeat). To compare in small environments where two nodes are ok, in my opinion glusterfs is far away from robust at the moment: with 3.4.1 on fedora 19 I cannot put a host in maintenance and get back a workable node after normal reboot. Didn't find any "safe" way for a node in distributed-replicated two nodes config to leave the cluster and get it back without becoming crazy to manual solve the brick differences generated in the mean time. And this was with only one VM running ..... Possibly I have to dig more in this, but I didn't find great resources for it. Any tips/links are welcome I would like to have oVirt more conscious about it and have sort of capability to solve itself the misalignments generated on gluster backend during mainteneance of a node. At the moment it seems to me it only shows volumes are ok in the sense of started, but they could be very different... For example another tab with details about heal info; something like the output of the command gluster volume heal $VOLUME info and/or gluster volume heal $VOLUME info split-brain so that if one finds 0 entries, he/she is calm and at the same doesn't risk to be erroneously calm in the other scenario... just my opinion. Gianluca _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users