Re: 5x slower guest disk performance with virtio disk
- Message from "Brian J. Murrell" - Date: Thu, 15 Dec 2011 14:43:23 -0500 From: "Brian J. Murrell" Subject: Re: 5x slower guest disk performance with virtio disk To: kvm@vger.kernel.org Cc: Sasha Levin On 11-12-15 12:27 PM, Daniel P. Berrange wrote: On Thu, Dec 15, 2011 at 07:16:22PM +0200, Sasha Levin wrote: On Thu, 2011-12-15 at 11:55 -0500, Brian J. Murrell wrote: So, about 2/3 of host speed now -- which is much better. Is 2/3 about normal or should I be looking for more? aio=native Thats the qemu setting, I'm not sure where libvirt hides that. ... When I try to "virsh edit" and add that "" it seems to get stripped out of the config (as observed with "virsh dumpxml"). Doing some googling I discovered an alternate syntax (in message https://www.redhat.com/archives/libvirt-users/2011-June/msg4.html): But the "io='native'" seems to get stripped out of that too. FWIW, I have: qemu-kvm-0.12.1.2-2.113.el6.x86_64 libvirt-client-0.8.1-27.el6.x86_64 installed here on CentOS 6.0. Maybe this aio= is not supported in the above package(s)? You need to update CentOS, your version of libvirt does not support io=native I don't believe. You are running into similar issues that I have posted about (mostly unsuccessfully in terms of getting much support LOL) here - https://www.centos.org/modules/newbb/viewtopic.php?topic_id=33708&forum=55 From that thread: With libvirt 0.8.7-6, qemu-kvm 0.12.1.2-2.144, and kernel 2.6.32-115, you can use the "io=native" parameter in the KVM xml files. Bugzilla 591703 has the details, but basically my img file references in the VM xml now reads like this: function='0x0'/> You may also want to search this list for my thread from November with the title "Improving RAID5 write performance in a KVM VM" Cheers Simon. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: 5x slower guest disk performance with virtio disk
On 11-12-15 12:27 PM, Daniel P. Berrange wrote: > On Thu, Dec 15, 2011 at 07:16:22PM +0200, Sasha Levin wrote: >> On Thu, 2011-12-15 at 11:55 -0500, Brian J. Murrell wrote: >>> So, about 2/3 of host speed now -- which is much better. Is 2/3 about >>> normal or should I be looking for more? >> >> aio=native >> >> Thats the qemu setting, I'm not sure where libvirt hides that. > > > > ... > When I try to "virsh edit" and add that "" it seems to get stripped out of the config (as observed with "virsh dumpxml"). Doing some googling I discovered an alternate syntax (in message https://www.redhat.com/archives/libvirt-users/2011-June/msg4.html): But the "io='native'" seems to get stripped out of that too. FWIW, I have: qemu-kvm-0.12.1.2-2.113.el6.x86_64 libvirt-client-0.8.1-27.el6.x86_64 installed here on CentOS 6.0. Maybe this aio= is not supported in the above package(s)? Cheers, b. signature.asc Description: OpenPGP digital signature
Re: 5x slower guest disk performance with virtio disk
On Thu, Dec 15, 2011 at 07:16:22PM +0200, Sasha Levin wrote: > On Thu, 2011-12-15 at 11:55 -0500, Brian J. Murrell wrote: > > So, about 2/3 of host speed now -- which is much better. Is 2/3 about > > normal or should I be looking for more? > > aio=native > > Thats the qemu setting, I'm not sure where libvirt hides that. ... Regards, Daniel -- |: http://berrange.com -o-http://www.flickr.com/photos/dberrange/ :| |: http://libvirt.org -o- http://virt-manager.org :| |: http://autobuild.org -o- http://search.cpan.org/~danberr/ :| |: http://entangle-photo.org -o- http://live.gnome.org/gtk-vnc :| -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: 5x slower guest disk performance with virtio disk
On Thu, 2011-12-15 at 11:55 -0500, Brian J. Murrell wrote: > So, about 2/3 of host speed now -- which is much better. Is 2/3 about > normal or should I be looking for more? aio=native Thats the qemu setting, I'm not sure where libvirt hides that. -- Sasha. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: 5x slower guest disk performance with virtio disk
On 11-12-15 10:47 AM, Stefan Pietsch wrote: > > Did you try to set the cache of the virtio disk to "none"? I didn't. It was set at "default" in virt-manager and I suppose I just assumed that "default" would be reasonable. Changing to "none" has had a good effect indeed: host $ ssh guest dd of=/dev/vdb if=/dev/zero bs=1M oflag=direct count=1 1+0 records in 1+0 records out 1048576 bytes (10 GB) copied, 49.9241 s, 210 MB/s host $ ssh guest dd of=/dev/vdb if=/dev/zero bs=1M oflag=direct count=1 1+0 records in 1+0 records out 1048576 bytes (10 GB) copied, 54.7737 s, 191 MB/s host $ ssh guest dd of=/dev/vdb if=/dev/zero bs=1M oflag=direct count=1 1+0 records in 1+0 records out 1048576 bytes (10 GB) copied, 50.9851 s, 206 MB/s host $ sudo dd of=/dev/datavol/disk1 if=/dev/zero bs=1M oflag=direct count=1 1+0 records in 1+0 records out 1048576 bytes (10 GB) copied, 31.0545 s, 338 MB/s So, about 2/3 of host speed now -- which is much better. Is 2/3 about normal or should I be looking for more? Thanks much! b. signature.asc Description: OpenPGP digital signature
Re: 5x slower guest disk performance with virtio disk
* Brian J. Murrell [2011-12-15 15:28]: > I have a CentOS 6 host system running a CentOS 6 KVM guest and the > guest seems to get about 5x slower disk throughput than the host: > > host $ sudo dd of=/dev/datavol/disk1 if=/dev/zero bs=1M oflag=direct > count=5 > 5+0 records in > 5+0 records out > 5242880 bytes (52 GB) copied, 150.36 s, 349 MB/s > host $ ssh guest dd of=/dev/vdb if=/dev/zero bs=1M oflag=direct count=5 > 5+0 records in > 5+0 records out > 5242880 bytes (52 GB) copied, 731.007 s, 71.7 MB/s > > This is actually a vast improvement over earlier tests where it was a > factor of 40x difference. But even still, I thought virtio disks were > supposed to be much closer to native host speed than 5x. --- snip --- Did you try to set the cache of the virtio disk to "none"? Regards, Stefan -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
5x slower guest disk performance with virtio disk
I have a CentOS 6 host system running a CentOS 6 KVM guest and the guest seems to get about 5x slower disk throughput than the host: host $ sudo dd of=/dev/datavol/disk1 if=/dev/zero bs=1M oflag=direct count=5 5+0 records in 5+0 records out 5242880 bytes (52 GB) copied, 150.36 s, 349 MB/s host $ ssh guest dd of=/dev/vdb if=/dev/zero bs=1M oflag=direct count=5 5+0 records in 5+0 records out 5242880 bytes (52 GB) copied, 731.007 s, 71.7 MB/s This is actually a vast improvement over earlier tests where it was a factor of 40x difference. But even still, I thought virtio disks were supposed to be much closer to native host speed than 5x. /dev/datavol/disk1 is a RAIDed LV in an LVM volume group consisting of 5 disks: $ sudo pvs PV VG Fmt Attr PSize PFree /dev/sda3 lvm2 a- 420.16g 420.16g /dev/sdb datavol lvm2 a- 465.76g 227.66g /dev/sdc datavol lvm2 a- 465.76g 245.76g /dev/sdd datavol lvm2 a- 465.76g 245.76g /dev/sde datavol lvm2 a- 465.76g 245.76g /dev/sdf datavol lvm2 a- 465.76g 245.76g $ sudo lvdisplay --maps /dev/datavol/disk1 --- Logical volume --- LV Name/dev/datavol/disk1 VG Namedatavol LV UUID3gD1N3-ybAU-GzUO-8QBV-b6op-lsK9-GMNm3w LV Write Accessread/write LV Status available # open 1 LV Size500.00 GiB Current LE 128000 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 253:1 --- Segments --- Logical extent 0 to 127999: Typestriped Stripes 5 Stripe size 4.00 KiB Stripe 0: Physical volume /dev/sdb Physical extents 5120 to 30719 Stripe 1: Physical volume /dev/sdc Physical extents 5120 to 30719 Stripe 2: Physical volume /dev/sdd Physical extents 5120 to 30719 Stripe 3: Physical volume /dev/sde Physical extents 5120 to 30719 Stripe 4: Physical volume /dev/sdf Physical extents 5120 to 30719 The KVM command line is as follows: $ tr '\0' '\n' < /proc/9409/cmdline /usr/libexec/qemu-kvm -S -M rhel6.0.0 -enable-kvm -m 1024 -smp 1,sockets=1,cores=1,threads=1 -name guest -uuid e2bf97e0-cfb7-444c-abc3-9efe262d8efe -nodefconfig -nodefaults -chardev socket,id=monitor,path=/var/lib/libvirt/qemu/guest.monitor,server,nowait -mon chardev=monitor,mode=control -rtc base=utc -boot c -drive file=/var/lib/libvirt/images/cdrom.iso,if=none,media=cdrom,id=drive-ide0-1-0,readonly=on,format=raw -device ide-drive,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -drive file=/var/lib/libvirt/images/guest.qcow2,if=none,id=drive-virtio-disk0,boot=on,format=qcow2 -device virtio-blk-pci,bus=pci.0,addr=0x4,drive=drive-virtio-disk0,id=virtio-disk0 -drive file=/dev/datavol/disk1,if=none,id=drive-virtio-disk1,format=raw -device virtio-blk-pci,bus=pci.0,addr=0x6,drive=drive-virtio-disk1,id=virtio-disk1 -netdev tap,fd=23,id=hostnet0 -device virtio-net-pci,netdev=hostnet0,id=net0,mac=52:54:00:ec:8d:47,bus=pci.0,addr=0x3 -chardev pty,id=serial0 -device isa-serial,chardev=serial0 -usb -vnc 127.0.0.1:2 -vga cirrus -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x5 That is started from virt-manager given the following configuration: guest e2bf97e0-cfb7-444c-abc3-9efe262d8efe 1048576 1048576 1 hvm destroy restart restart /usr/libexec/qemu-kvm Any ideas why the huge performance difference? Cheers, b. signature.asc Description: OpenPGP digital signature