Jim - One thing I noticed is that, by accident, I used 'backupvolfile-server=node2:node3' which is apparently a supported setting. It would appear, by reading the man page of mount.glusterfs, the syntax is slightly different. not sure if my setting being different has different impacts
hosted-engine.conf: # cat /etc/ovirt-hosted-engine/hosted-engine.conf | grep -i option mnt_options=backup-volfile-servers=node2:node3 And for my datatest gluster domain I have: backupvolfile-server=node2:node3 I am now curious what happens when I move everything to node1 and drop node2 To that end, will follow up with that test On Fri, Sep 1, 2017 at 7:20 PM, Charles Kozler <ckozler...@gmail.com> wrote: > Jim - > > here is my test: > > - All VM's on node2: hosted engine and 1 test VM > - Test VM on gluster storage domain (with mount options set) > - hosted engine is on gluster as well, with settings persisted to > hosted-engine.conf for backupvol > > All VM's stayed up. Nothing in dmesg of the test vm indicating a pause or > an issue or anything > > However, what I did notice during this, is my /datatest volume doesnt have > quorum set. So I will set that now and report back what happens > > # gluster volume info datatest > > Volume Name: datatest > Type: Replicate > Volume ID: 229c25f9-405e-4fe7-b008-1d3aea065069 > Status: Started > Snapshot Count: 0 > Number of Bricks: 1 x 3 = 3 > Transport-type: tcp > Bricks: > Brick1: node1:/gluster/data/datatest/brick1 > Brick2: node2:/gluster/data/datatest/brick1 > Brick3: node3:/gluster/data/datatest/brick1 > Options Reconfigured: > transport.address-family: inet > nfs.disable: on > > Perhaps quorum may be more trouble than its worth when you have 3 nodes > and/or 2 nodes + arbiter? > > Since I am keeping my 3rd node out of ovirt, I am more content on keeping > it as a warm spare if I **had** to swap it in to ovirt cluster, but keeps > my storage 100% quorum > > On Fri, Sep 1, 2017 at 5:18 PM, Jim Kusznir <j...@palousetech.com> wrote: > >> I can confirm that I did set it up manually, and I did specify backupvol, >> and in the "manage domain" storage settings, I do have under mount >> options, backup-volfile-servers=192.168.8.12:192.168.8.13 (and this was >> done at initial install time). >> >> The "used managed gluster" checkbox is NOT checked, and if I check it and >> save settings, next time I go in it is not checked. >> >> --Jim >> >> On Fri, Sep 1, 2017 at 2:08 PM, Charles Kozler <ckozler...@gmail.com> >> wrote: >> >>> @ Jim - here is my setup which I will test in a few (brand new cluster) >>> and report back what I found in my tests >>> >>> - 3x servers direct connected via 10Gb >>> - 2 of those 3 setup in ovirt as hosts >>> - Hosted engine >>> - Gluster replica 3 (no arbiter) for all volumes >>> - 1x engine volume gluster replica 3 manually configured (not using >>> ovirt managed gluster) >>> - 1x datatest volume (20gb) replica 3 manually configured (not using >>> ovirt managed gluster) >>> - 1x nfstest domain served from some other server in my infrastructure >>> which, at the time of my original testing, was master domain >>> >>> I tested this earlier and all VMs stayed online. However, ovirt cluster >>> reported DC/cluster down, all VM's stayed up >>> >>> As I am now typing this, can you confirm you setup your gluster storage >>> domain with backupvol? Also, confirm you updated hosted-engine.conf with >>> backupvol mount option as well? >>> >>> On Fri, Sep 1, 2017 at 4:22 PM, Jim Kusznir <j...@palousetech.com> wrote: >>> >>>> So, after reading the first document twice and the 2nd link thoroughly >>>> once, I believe that the arbitrator volume should be sufficient and count >>>> for replica / split brain. EG, if any one full replica is down, and the >>>> arbitrator and the other replica is up, then it should have quorum and all >>>> should be good. >>>> >>>> I think my underlying problem has to do more with config than the >>>> replica state. That said, I did size the drive on my 3rd node planning to >>>> have an identical copy of all data on it, so I'm still not opposed to >>>> making it a full replica. >>>> >>>> Did I miss something here? >>>> >>>> Thanks! >>>> >>>> On Fri, Sep 1, 2017 at 11:59 AM, Charles Kozler <ckozler...@gmail.com> >>>> wrote: >>>> >>>>> These can get a little confusing but this explains it best: >>>>> https://gluster.readthedocs.io/en/latest/Administrator >>>>> %20Guide/arbiter-volumes-and-quorum/#replica-2-and-replica-3-volumes >>>>> >>>>> Basically in the first paragraph they are explaining why you cant have >>>>> HA with quorum for 2 nodes. Here is another overview doc that explains >>>>> some >>>>> more >>>>> >>>>> http://openmymind.net/Does-My-Replica-Set-Need-An-Arbiter/ >>>>> >>>>> From my understanding arbiter is good for resolving split brains. >>>>> Quorum and arbiter are two different things though quorum is a mechanism >>>>> to >>>>> help you **avoid** split brain and the arbiter is to help gluster resolve >>>>> split brain by voting and other internal mechanics (as outlined in link >>>>> 1). >>>>> How did you create the volume exactly - what command? It looks to me like >>>>> you created it with 'gluster volume create replica 2 arbiter 1 {....}' per >>>>> your earlier mention of "replica 2 arbiter 1". That being said, if you did >>>>> that and then setup quorum in the volume configuration, this would cause >>>>> your gluster to halt up since quorum was lost (as you saw until you >>>>> recovered node 1) >>>>> >>>>> As you can see from the docs, there is still a corner case for getting >>>>> in to split brain with replica 3, which again, is where arbiter would help >>>>> gluster resolve it >>>>> >>>>> I need to amend my previous statement: I was told that arbiter volume >>>>> does not store data, only metadata. I cannot find anything in the docs >>>>> backing this up however it would make sense for it to be. That being said, >>>>> in my setup, I would not include my arbiter or my third node in my ovirt >>>>> VM >>>>> cluster component. I would keep it completely separate >>>>> >>>>> >>>>> On Fri, Sep 1, 2017 at 2:46 PM, Jim Kusznir <j...@palousetech.com> >>>>> wrote: >>>>> >>>>>> I'm now also confused as to what the point of an arbiter is / what it >>>>>> does / why one would use it. >>>>>> >>>>>> On Fri, Sep 1, 2017 at 11:44 AM, Jim Kusznir <j...@palousetech.com> >>>>>> wrote: >>>>>> >>>>>>> Thanks for the help! >>>>>>> >>>>>>> Here's my gluster volume info for the data export/brick (I have 3: >>>>>>> data, engine, and iso, but they're all configured the same): >>>>>>> >>>>>>> Volume Name: data >>>>>>> Type: Replicate >>>>>>> Volume ID: e670c488-ac16-4dd1-8bd3-e43b2e42cc59 >>>>>>> Status: Started >>>>>>> Snapshot Count: 0 >>>>>>> Number of Bricks: 1 x (2 + 1) = 3 >>>>>>> Transport-type: tcp >>>>>>> Bricks: >>>>>>> Brick1: ovirt1.nwfiber.com:/gluster/brick2/data >>>>>>> Brick2: ovirt2.nwfiber.com:/gluster/brick2/data >>>>>>> Brick3: ovirt3.nwfiber.com:/gluster/brick2/data (arbiter) >>>>>>> Options Reconfigured: >>>>>>> performance.strict-o-direct: on >>>>>>> nfs.disable: on >>>>>>> user.cifs: off >>>>>>> network.ping-timeout: 30 >>>>>>> cluster.shd-max-threads: 8 >>>>>>> cluster.shd-wait-qlength: 10000 >>>>>>> cluster.locking-scheme: granular >>>>>>> cluster.data-self-heal-algorithm: full >>>>>>> performance.low-prio-threads: 32 >>>>>>> features.shard-block-size: 512MB >>>>>>> features.shard: on >>>>>>> storage.owner-gid: 36 >>>>>>> storage.owner-uid: 36 >>>>>>> cluster.server-quorum-type: server >>>>>>> cluster.quorum-type: auto >>>>>>> network.remote-dio: enable >>>>>>> cluster.eager-lock: enable >>>>>>> performance.stat-prefetch: off >>>>>>> performance.io-cache: off >>>>>>> performance.read-ahead: off >>>>>>> performance.quick-read: off >>>>>>> performance.readdir-ahead: on >>>>>>> server.allow-insecure: on >>>>>>> [root@ovirt1 ~]# >>>>>>> >>>>>>> >>>>>>> all 3 of my brick nodes ARE also members of the virtualization >>>>>>> cluster (including ovirt3). How can I convert it into a full replica >>>>>>> instead of just an arbiter? >>>>>>> >>>>>>> Thanks! >>>>>>> --Jim >>>>>>> >>>>>>> On Fri, Sep 1, 2017 at 9:09 AM, Charles Kozler <ckozler...@gmail.com >>>>>>> > wrote: >>>>>>> >>>>>>>> @Kasturi - Looks good now. Cluster showed down for a moment but >>>>>>>> VM's stayed up in their appropriate places. Thanks! >>>>>>>> >>>>>>>> < Anyone on this list please feel free to correct my response to >>>>>>>> Jim if its wrong> >>>>>>>> >>>>>>>> @ Jim - If you can share your gluster volume info / status I can >>>>>>>> confirm (to the best of my knowledge). From my understanding, If you >>>>>>>> setup >>>>>>>> the volume with something like 'gluster volume set <vol> group virt' >>>>>>>> this >>>>>>>> will configure some quorum options as well, Ex: >>>>>>>> http://i.imgur.com/Mya4N5o.png >>>>>>>> >>>>>>>> While, yes, you are configured for arbiter node you're still losing >>>>>>>> quorum by dropping from 2 -> 1. You would need 4 node with 1 being >>>>>>>> arbiter >>>>>>>> to configure quorum which is in effect 3 writable nodes and 1 arbiter. >>>>>>>> If >>>>>>>> one gluster node drops, you still have 2 up. Although in this case, you >>>>>>>> probably wouldnt need arbiter at all >>>>>>>> >>>>>>>> If you are configured, you can drop quorum settings and just let >>>>>>>> arbiter run since you're not using arbiter node in your VM cluster >>>>>>>> part (I >>>>>>>> believe), just storage cluster part. When using quorum, you need > 50% >>>>>>>> of >>>>>>>> the cluster being up at one time. Since you have 3 nodes with 1 >>>>>>>> arbiter, >>>>>>>> you're actually losing 1/2 which == 50 which == degraded / hindered >>>>>>>> gluster >>>>>>>> >>>>>>>> Again, this is to the best of my knowledge based on other quorum >>>>>>>> backed software....and this is what I understand from testing with >>>>>>>> gluster >>>>>>>> and ovirt thus far >>>>>>>> >>>>>>>> On Fri, Sep 1, 2017 at 11:53 AM, Jim Kusznir <j...@palousetech.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Huh...Ok., how do I convert the arbitrar to full replica, then? I >>>>>>>>> was misinformed when I created this setup. I thought the arbitrator >>>>>>>>> held >>>>>>>>> enough metadata that it could validate or refudiate any one replica >>>>>>>>> (kinda >>>>>>>>> like the parity drive for a RAID-4 array). I was also under the >>>>>>>>> impression >>>>>>>>> that one replica + Arbitrator is enough to keep the array online and >>>>>>>>> functional. >>>>>>>>> >>>>>>>>> --Jim >>>>>>>>> >>>>>>>>> On Fri, Sep 1, 2017 at 5:22 AM, Charles Kozler < >>>>>>>>> ckozler...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> @ Jim - you have only two data volumes and lost quorum. >>>>>>>>>> Arbitrator only stores metadata, no actual files. So yes, you were >>>>>>>>>> running >>>>>>>>>> in degraded mode so some operations were hindered. >>>>>>>>>> >>>>>>>>>> @ Sahina - Yes, this actually worked fine for me once I did that. >>>>>>>>>> However, the issue I am still facing, is when I go to create a new >>>>>>>>>> gluster >>>>>>>>>> storage domain (replica 3, hyperconverged) and I tell it "Host to >>>>>>>>>> use" and >>>>>>>>>> I select that host. If I fail that host, all VMs halt. I do not >>>>>>>>>> recall this >>>>>>>>>> in 3.6 or early 4.0. This to me makes it seem like this is "pinning" >>>>>>>>>> a node >>>>>>>>>> to a volume and vice versa like you could, for instance, for a >>>>>>>>>> singular >>>>>>>>>> hyperconverged to ex: export a local disk via NFS and then mount it >>>>>>>>>> via >>>>>>>>>> ovirt domain. But of course, this has its caveats. To that end, I am >>>>>>>>>> using >>>>>>>>>> gluster replica 3, when configuring it I say "host to use: " node 1, >>>>>>>>>> then >>>>>>>>>> in the connection details I give it node1:/data. I fail node1, all >>>>>>>>>> VMs >>>>>>>>>> halt. Did I miss something? >>>>>>>>>> >>>>>>>>>> On Fri, Sep 1, 2017 at 2:13 AM, Sahina Bose <sab...@redhat.com> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> To the OP question, when you set up a gluster storage domain, >>>>>>>>>>> you need to specify backup-volfile-servers=<server2>:<server3> >>>>>>>>>>> where server2 and server3 also have bricks running. When server1 is >>>>>>>>>>> down, >>>>>>>>>>> and the volume is mounted again - server2 or server3 are queried to >>>>>>>>>>> get the >>>>>>>>>>> gluster volfiles. >>>>>>>>>>> >>>>>>>>>>> @Jim, if this does not work, are you using 4.1.5 build with >>>>>>>>>>> libgfapi access? If not, please provide the vdsm and gluster mount >>>>>>>>>>> logs to >>>>>>>>>>> analyse >>>>>>>>>>> >>>>>>>>>>> If VMs go to paused state - this could mean the storage is not >>>>>>>>>>> available. You can check "gluster volume status <volname>" to see if >>>>>>>>>>> atleast 2 bricks are running. >>>>>>>>>>> >>>>>>>>>>> On Fri, Sep 1, 2017 at 11:31 AM, Johan Bernhardsson < >>>>>>>>>>> jo...@kafit.se> wrote: >>>>>>>>>>> >>>>>>>>>>>> If gluster drops in quorum so that it has less votes than it >>>>>>>>>>>> should it will stop file operations until quorum is back to >>>>>>>>>>>> normal.If i >>>>>>>>>>>> rember it right you need two bricks to write for quorum to be met >>>>>>>>>>>> and that >>>>>>>>>>>> the arbiter only is a vote to avoid split brain. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Basically what you have is a raid5 solution without a spare. >>>>>>>>>>>> And when one disk dies it will run in degraded mode. And some raid >>>>>>>>>>>> systems >>>>>>>>>>>> will stop the raid until you have removed the disk or forced it to >>>>>>>>>>>> run >>>>>>>>>>>> anyway. >>>>>>>>>>>> >>>>>>>>>>>> You can read up on it here: https://gluster.readthed >>>>>>>>>>>> ocs.io/en/latest/Administrator%20Guide/arbiter-volumes-and-q >>>>>>>>>>>> uorum/ >>>>>>>>>>>> >>>>>>>>>>>> /Johan >>>>>>>>>>>> >>>>>>>>>>>> On Thu, 2017-08-31 at 22:33 -0700, Jim Kusznir wrote: >>>>>>>>>>>> >>>>>>>>>>>> Hi all: >>>>>>>>>>>> >>>>>>>>>>>> Sorry to hijack the thread, but I was about to start >>>>>>>>>>>> essentially the same thread. >>>>>>>>>>>> >>>>>>>>>>>> I have a 3 node cluster, all three are hosts and gluster nodes >>>>>>>>>>>> (replica 2 + arbitrar). I DO have the >>>>>>>>>>>> mnt_options=backup-volfile-servers= >>>>>>>>>>>> set: >>>>>>>>>>>> >>>>>>>>>>>> storage=192.168.8.11:/engine >>>>>>>>>>>> mnt_options=backup-volfile-servers=192.168.8.12:192.168.8.13 >>>>>>>>>>>> >>>>>>>>>>>> I had an issue today where 192.168.8.11 went down. ALL VMs >>>>>>>>>>>> immediately paused, including the engine (all VMs were running on >>>>>>>>>>>> host2:192.168.8.12). I couldn't get any gluster stuff working >>>>>>>>>>>> until host1 >>>>>>>>>>>> (192.168.8.11) was restored. >>>>>>>>>>>> >>>>>>>>>>>> What's wrong / what did I miss? >>>>>>>>>>>> >>>>>>>>>>>> (this was set up "manually" through the article on setting up >>>>>>>>>>>> self-hosted gluster cluster back when 4.0 was new..I've upgraded >>>>>>>>>>>> it to 4.1 >>>>>>>>>>>> since). >>>>>>>>>>>> >>>>>>>>>>>> Thanks! >>>>>>>>>>>> --Jim >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Aug 31, 2017 at 12:31 PM, Charles Kozler < >>>>>>>>>>>> ckozler...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>> Typo..."Set it up and then failed that **HOST**" >>>>>>>>>>>> >>>>>>>>>>>> And upon that host going down, the storage domain went down. I >>>>>>>>>>>> only have hosted storage domain and this new one - is this why the >>>>>>>>>>>> DC went >>>>>>>>>>>> down and no SPM could be elected? >>>>>>>>>>>> >>>>>>>>>>>> I dont recall this working this way in early 4.0 or 3.6 >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Aug 31, 2017 at 3:30 PM, Charles Kozler < >>>>>>>>>>>> ckozler...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>> So I've tested this today and I failed a node. Specifically, I >>>>>>>>>>>> setup a glusterfs domain and selected "host to use: node1". Set it >>>>>>>>>>>> up and >>>>>>>>>>>> then failed that VM >>>>>>>>>>>> >>>>>>>>>>>> However, this did not work and the datacenter went down. My >>>>>>>>>>>> engine stayed up, however, it seems configuring a domain to pin to >>>>>>>>>>>> a host >>>>>>>>>>>> to use will obviously cause it to fail >>>>>>>>>>>> >>>>>>>>>>>> This seems counter-intuitive to the point of glusterfs or any >>>>>>>>>>>> redundant storage. If a single host has to be tied to its >>>>>>>>>>>> function, this >>>>>>>>>>>> introduces a single point of failure >>>>>>>>>>>> >>>>>>>>>>>> Am I missing something obvious? >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Aug 31, 2017 at 9:43 AM, Kasturi Narra < >>>>>>>>>>>> kna...@redhat.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>> yes, right. What you can do is edit the hosted-engine.conf >>>>>>>>>>>> file and there is a parameter as shown below [1] and replace h2 >>>>>>>>>>>> and h3 with >>>>>>>>>>>> your second and third storage servers. Then you will need to >>>>>>>>>>>> restart >>>>>>>>>>>> ovirt-ha-agent and ovirt-ha-broker services in all the nodes . >>>>>>>>>>>> >>>>>>>>>>>> [1] 'mnt_options=backup-volfile-servers=<h2>:<h3>' >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Aug 31, 2017 at 5:54 PM, Charles Kozler < >>>>>>>>>>>> ckozler...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>> Hi Kasturi - >>>>>>>>>>>> >>>>>>>>>>>> Thanks for feedback >>>>>>>>>>>> >>>>>>>>>>>> > If cockpit+gdeploy plugin would be have been used then that >>>>>>>>>>>> would have automatically detected glusterfs replica 3 volume >>>>>>>>>>>> created during >>>>>>>>>>>> Hosted Engine deployment and this question would not have been >>>>>>>>>>>> asked >>>>>>>>>>>> >>>>>>>>>>>> Actually, doing hosted-engine --deploy it too also auto detects >>>>>>>>>>>> glusterfs. I know glusterfs fuse client has the ability to >>>>>>>>>>>> failover >>>>>>>>>>>> between all nodes in cluster, but I am still curious given the >>>>>>>>>>>> fact that I >>>>>>>>>>>> see in ovirt config node1:/engine (being node1 I set it to in >>>>>>>>>>>> hosted-engine >>>>>>>>>>>> --deploy). So my concern was to ensure and find out exactly how >>>>>>>>>>>> engine >>>>>>>>>>>> works when one node goes away and the fuse client moves over to >>>>>>>>>>>> the other >>>>>>>>>>>> node in the gluster cluster >>>>>>>>>>>> >>>>>>>>>>>> But you did somewhat answer my question, the answer seems to be >>>>>>>>>>>> no (as default) and I will have to use hosted-engine.conf and >>>>>>>>>>>> change the >>>>>>>>>>>> parameter as you list >>>>>>>>>>>> >>>>>>>>>>>> So I need to do something manual to create HA for engine on >>>>>>>>>>>> gluster? Yes? >>>>>>>>>>>> >>>>>>>>>>>> Thanks so much! >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Aug 31, 2017 at 3:03 AM, Kasturi Narra < >>>>>>>>>>>> kna...@redhat.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>> Hi, >>>>>>>>>>>> >>>>>>>>>>>> During Hosted Engine setup question about glusterfs volume >>>>>>>>>>>> is being asked because you have setup the volumes yourself. If >>>>>>>>>>>> cockpit+gdeploy plugin would be have been used then that would have >>>>>>>>>>>> automatically detected glusterfs replica 3 volume created during >>>>>>>>>>>> Hosted >>>>>>>>>>>> Engine deployment and this question would not have been asked. >>>>>>>>>>>> >>>>>>>>>>>> During new storage domain creation when glusterfs is >>>>>>>>>>>> selected there is a feature called 'use managed gluster volumes' >>>>>>>>>>>> and upon >>>>>>>>>>>> checking this all glusterfs volumes managed will be listed and you >>>>>>>>>>>> could >>>>>>>>>>>> choose the volume of your choice from the dropdown list. >>>>>>>>>>>> >>>>>>>>>>>> There is a conf file called >>>>>>>>>>>> /etc/hosted-engine/hosted-engine.conf >>>>>>>>>>>> where there is a parameter called backup-volfile-servers="h1:h2" >>>>>>>>>>>> and if one >>>>>>>>>>>> of the gluster node goes down engine uses this parameter to >>>>>>>>>>>> provide ha / >>>>>>>>>>>> failover. >>>>>>>>>>>> >>>>>>>>>>>> Hope this helps !! >>>>>>>>>>>> >>>>>>>>>>>> Thanks >>>>>>>>>>>> kasturi >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Wed, Aug 30, 2017 at 8:09 PM, Charles Kozler < >>>>>>>>>>>> ckozler...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>> Hello - >>>>>>>>>>>> >>>>>>>>>>>> I have successfully created a hyperconverged hosted engine >>>>>>>>>>>> setup consisting of 3 nodes - 2 for VM's and the third purely for >>>>>>>>>>>> storage. >>>>>>>>>>>> I manually configured it all, did not use ovirt node or anything. >>>>>>>>>>>> Built the >>>>>>>>>>>> gluster volumes myself >>>>>>>>>>>> >>>>>>>>>>>> However, I noticed that when setting up the hosted engine and >>>>>>>>>>>> even when adding a new storage domain with glusterfs type, it >>>>>>>>>>>> still asks >>>>>>>>>>>> for hostname:/volumename >>>>>>>>>>>> >>>>>>>>>>>> This leads me to believe that if that one node goes down (ex: >>>>>>>>>>>> node1:/data), then ovirt engine wont be able to communicate with >>>>>>>>>>>> that >>>>>>>>>>>> volume because its trying to reach it on node 1 and thus, go down >>>>>>>>>>>> >>>>>>>>>>>> I know glusterfs fuse client can connect to all nodes to >>>>>>>>>>>> provide failover/ha but how does the engine handle this? >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Users mailing list >>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Users mailing list >>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Users mailing >>>>>>>>>>>> listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Users mailing list >>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> Users mailing list >>>>>>>>>> Users@ovirt.org >>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users