Proxmox does it quite nicely using LVM ontop of iSCSI targets - the LVM locking only allows an LV to be accessed by one host at a time, but if its deactivated on one host it can be reactivated on another
http://pve.proxmox.com/wiki/Storage_Model#LVM_Groups_with_Network_Backing On 10 March 2011 21:02, <jurgen.depic...@let.be> wrote: > Hi all. > > Presently, 7 VMs (4 windows, 3 ubuntu) are using the same nfs pool. The > machine serving that NFS pool (VLET1) has high load as soon as there is some > continuous disk activity: > > top - 13:25:30 up 23:16, 5 users, load average: 4.51, 4.41, 3.98 > Tasks: 191 total, 1 running, 189 sleeping, 0 stopped, 1 zombie > Cpu0 : 3.0%us, 2.6%sy, 0.0%ni, 44.6%id, 49.8%wa, 0.0%hi, 0.0%si, > 0.0%st > Cpu1 : 4.2%us, 2.9%sy, 0.0%ni, 54.2%id, 38.1%wa, 0.0%hi, 0.6%si, > 0.0%st > Cpu2 : 3.6%us, 2.3%sy, 0.0%ni, 55.4%id, 38.7%wa, 0.0%hi, 0.0%si, > 0.0%st > Cpu3 : 2.9%us, 4.2%sy, 0.0%ni, 56.4%id, 36.5%wa, 0.0%hi, 0.0%si, > 0.0%st > > It's a quad-core Xeon, running SW Raid on sataII disks: > cat /proc/mdstat > Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] > [raid4] [raid10] > md0 : active raid1 sdb1[1] sda1[0] > 966796224 blocks [2/2] [UU] > > When the VM client's disk activity is low, the load on VLET1 is very > minimal: > top - 13:53:30 up 23:44, 5 users, load average: 0.11, 0.81, 1.83 > Tasks: 191 total, 1 running, 189 sleeping, 0 stopped, 1 zombie > Cpu0 : 1.0%us, 2.0%sy, 0.0%ni, 92.1%id, 4.6%wa, 0.0%hi, 0.3%si, > 0.0%st > Cpu1 : 2.2%us, 5.3%sy, 0.0%ni, 90.7%id, 1.9%wa, 0.0%hi, 0.0%si, > 0.0%st > Cpu2 : 4.2%us, 3.6%sy, 0.0%ni, 87.7%id, 4.5%wa, 0.0%hi, 0.0%si, > 0.0%st > Cpu3 : 3.2%us, 3.2%sy, 0.0%ni, 93.3%id, 0.3%wa, 0.0%hi, 0.0%si, > 0.0%st > > What can I do to reduce the load? Now, sometimes a VM refuses to come up > (windows blue screen) except when I launch it on the same host (VLET1) > hosting the nfs share. After booting the VM, I can live migrate it to > another VM host, eg VLET2. > I did spot an error in /var/log/messages at boot failure of the windows VM > when starting it on VLET2: > Mar 10 11:06:46 VLET2 libvirtd: 11:06:46.137: warning : > qemudParsePCIDeviceStrs:1422 : Unexpected exit status '1', qemu probably > failedbut I'm not sure this is related. > > This is a test config; our company's mail server will soon be running in > this cluster of KVM hosts (VLET1 and VLET2 will be joined by VLET3 next > week). Since the Domino mail server is very disk intensive, I'm a bit > worried now. I would rather not run it on local disks, sionce this makes > live migration impossible. I'll need to decide where to put the storage > pool, and using which protocol (NFS or iSCSI). > > Which brings me to my last question: I was wondering whether it would be > better to use iSCSI instead of NFS? I started with this, but couldn't get a > pool defined through virt-manager (it always showed as 100% full, even when > I created a completely new iSCSI target/lun). According to e.g. > http://communities.vmware.com/thread/150775 it doesn't seem to make much > difference, whether I use iSCSI or NFS, performance-wise. Anyhow, I don't > grasp how KVM running on different hosts, connected to the same iSCSI LUN, > can work? It corrupts data, having the same LUN mounted on different > hosts... I obviously understood something wrong about iSCSI... > > Regards, Jürgen > > -- > ubuntu-server mailing list > ubuntu-server@lists.ubuntu.com > https://lists.ubuntu.com/mailman/listinfo/ubuntu-server > More info: https://wiki.ubuntu.com/ServerTeam > -- Adon Metcalfe Labyrinth Data Services Pty Ltd http://www.labyrinthdata.net.au
-- ubuntu-server mailing list ubuntu-server@lists.ubuntu.com https://lists.ubuntu.com/mailman/listinfo/ubuntu-server More info: https://wiki.ubuntu.com/ServerTeam