Re: [ovirt-users] Ovirt with bad IO performance

2016-09-06 Thread Gabriel Ozaki
Hi Yaniv

This results is averange in sysbench, my machine for example gets
1.3905Mb/sec, i don't know how this test really works and i will search
about it

So i try to make a* bonnie++ test* ( reference
http://support.commgate.net/index.php?/Knowledgebase/Article/View/212 ):

Xenserver speeds:
Write speed: 91076 KB/sec
ReWrite speed: 57885 KB/sec
Read speed: 215457 KB/sec (Strange, too high)
Num of Blocks: 632.4

Ovirt Speeds:
Write speed: 111597 KB/sec (22% more then xenserver)
ReWrite speed: 73402 KB/sec (26% more then xenserver)
Read speed:  121537 KB/sec (44% less then xenserver)
Num of Blocks: 537.2 ( 15% less then xenserver)


result: a draw?


And* DD test *( reference: https://romanrm.net/dd-benchmark )*:*
[root@xenserver teste]# echo 2 > /proc/sys/vm/drop_caches  && sync
[root@xenserver teste]# dd bs=1M count=256 if=/dev/zero of=test
conv=fdatasync
256+0 registros de entrada
256+0 registros de saída
268435456 bytes (268 MB) copiados, 1,40111 s, 192 MB/s (Again, too high)

[root@ovirt teste]# echo 2 > /proc/sys/vm/drop_caches  && sync
[root@ovirt teste]# dd bs=1M count=256 if=/dev/zero of=test conv=fdatasync
256+0 registros de entrada
256+0 registros de saída
268435456 bytes (268 MB) copiados, 2,31288 s, 116 MB/s (Really fair, the
host result is 124 MB/s)


*HDparm *(FAIL on xenserver)
[root@xenserver teste]# hdparm -Tt /dev/xvda1

/dev/xvda1:
 Timing cached reads:   25724 MB in  2.00 seconds = 12882.77 MB/sec
 Timing buffered disk reads: 2984 MB in  3.00 seconds = 994.43 MB/sec ( 8
times the expect value, something is very wrong)

[root@ovirt teste]# hdparm -Tt /dev/vda1

/dev/vda1:
 Timing cached reads:   25042 MB in  2.00 seconds = 12540.21 MB/sec
 Timing buffered disk reads: 306 MB in  3.01 seconds = 101.66 MB/sec(ok
result)


There is something strange in xenserver affecting the results, probably the
best choice is close the thread and  start the studies about benchmarks

Thanks






2016-09-05 12:01 GMT-03:00 Yaniv Kaul <yk...@redhat.com>:

>
>
> On Mon, Sep 5, 2016 at 1:45 PM, Gabriel Ozaki <gabriel.oz...@kemi.com.br>
> wrote:
>
>> Hi Yaniv and Sandro
>>
>> The disk is in the same machine then ovirt-engine
>>
>
> I'm looking back at your results, and something is terribly wrong there:
> For example, sysbench:
>
> Host result:2.9843Mb/sec
> Ovirt result:   1.1561Mb/sec
> Xenserver result:   2.9006Mb/sec
>
> This is slower than a USB1 disk on key performance. I don't know what to
> make of it, but it's completely bogus. Even plain QEMU can get better
> results than this.
> And the 2nd benchmark:
>
>
> **The novabench test:*
> Ovirt result:  79Mb/s
> Xenserver result:  101Mb/s
>
> This is better, but still very slow. If I translate it to MB/s, it's
> ~10-12MBs - still very very slow.
> If, however, this is MB/sec, then this makes sense - and is probably as
> much as you can get from a single spindle.
> The difference between XenServer and oVirt are more likely have to do with
> caching than anything else. I don't know what the caching settings of
> XenServer - can you ensure no caching ('direct IO') is used?
>
>
>
>> Thanks
>>
>>
>>
>>
>>
>> 2016-09-02 15:31 GMT-03:00 Yaniv Kaul <yk...@redhat.com>:
>>
>>>
>>>
>>> On Fri, Sep 2, 2016 at 6:11 PM, Gabriel Ozaki <gabriel.oz...@kemi.com.br
>>> > wrote:
>>>
>>>> Hi Yaniv
>>>>
>>>> Sorry guys, i don't explain well on my first mail, i notice a bad IO
>>>> performance on *disk* benchmarks, the network are working really fine
>>>>
>>>
>>> But where is the disk? If it's across the network, then network is
>>> involved and is certainly a bottleneck.
>>> Y.
>>>
>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> 2016-09-02 12:04 GMT-03:00 Yaniv Kaul <yk...@redhat.com>:
>>>>
>>>>> On Fri, Sep 2, 2016 at 5:33 PM, Gabriel Ozaki <
>>>>> gabriel.oz...@kemi.com.br> wrote:
>>>>>
>>>>>> Hi Nir, thanks for the answer
>>>>>>
>>>>>>
>>>>>> *The nfs server is in the host?*
>>>>>> Yes, i choose NFS to use as storage on ovirt host
>>>>>>
>>>>>> *- Is this 2.9GiB/s or 2.9 MiB/s?*
>>>>>> Is MiB/s, i put the full test on paste bin
>>>>>> centos guest on ovirt:
>>>>>> http://pastebin.com/d48qfvuf
>>>>>>
>>>>>> centos guest on xenserver:
>>>>>> http://pastebin.com/gqN3du29
>>>>

Re: [ovirt-users] Ovirt with bad IO performance

2016-09-05 Thread Gabriel Ozaki
Hi Yaniv and Sandro

The disk is in the same machine then ovirt-engine

Thanks





2016-09-02 15:31 GMT-03:00 Yaniv Kaul <yk...@redhat.com>:

>
>
> On Fri, Sep 2, 2016 at 6:11 PM, Gabriel Ozaki <gabriel.oz...@kemi.com.br>
> wrote:
>
>> Hi Yaniv
>>
>> Sorry guys, i don't explain well on my first mail, i notice a bad IO
>> performance on *disk* benchmarks, the network are working really fine
>>
>
> But where is the disk? If it's across the network, then network is
> involved and is certainly a bottleneck.
> Y.
>
>
>>
>>
>>
>>
>>
>>
>>
>> 2016-09-02 12:04 GMT-03:00 Yaniv Kaul <yk...@redhat.com>:
>>
>>> On Fri, Sep 2, 2016 at 5:33 PM, Gabriel Ozaki <gabriel.oz...@kemi.com.br
>>> > wrote:
>>>
>>>> Hi Nir, thanks for the answer
>>>>
>>>>
>>>> *The nfs server is in the host?*
>>>> Yes, i choose NFS to use as storage on ovirt host
>>>>
>>>> *- Is this 2.9GiB/s or 2.9 MiB/s?*
>>>> Is MiB/s, i put the full test on paste bin
>>>> centos guest on ovirt:
>>>> http://pastebin.com/d48qfvuf
>>>>
>>>> centos guest on xenserver:
>>>> http://pastebin.com/gqN3du29
>>>>
>>>> how the test works:
>>>> https://www.howtoforge.com/how-to-benchmark-your-system-cpu-
>>>> file-io-mysql-with-sysbench
>>>>
>>>> *- Are you testing using NFS in all versions?*
>>>> i am using the v3 version
>>>>
>>>> *- What is the disk format?*
>>>> partion size format
>>>> /20Gb xfs
>>>> swap 2 Gb xfs
>>>> /dados rest of disk xfs   (note, this is the partition where i save the
>>>> ISOs,exports and VM disks)
>>>>
>>>>
>>>> *- How do you test io on the host?*
>>>> I do a clean install of centos and do the test before i install the
>>>> ovirt
>>>> the test:
>>>> http://pastebin.com/7RKU7778
>>>>
>>>> *- What kind of nic is used? (1G, 10G?)*
>>>> Is only a 100mbps :(
>>>>
>>>
>>> 100Mbps will not get you more than several MB/s. 11MB/s on a very bright
>>> day...
>>>
>>>>
>>>> *We need much more details to understand what do you test here.*
>>>> I have problems to upload the benchmark test on orvirt to novabench
>>>> site, so here is the screenshot(i make a mistake on the last email i get
>>>> the wrong value), is 86 Mb/s:
>>>>
>>>
>>> Which is not possible on the wire. Unless it's VM to VM? And the storage
>>> is local, which means it's the bandwidth of the physical disk itself?
>>> Y.
>>>
>>>
>>>
>>>>
>>>> ​
>>>> And the novabench on xenserver:
>>>> https://novabench.com/compare.php?id=ba8dd628e4042dfc1f3d396
>>>> 70b164ab11061671
>>>>
>>>> *- For Xenserver - detailed description of the vm and the storage
>>>> configuration?*
>>>> The host is the same(i install xenserver, do the tests before i install
>>>> centos), the VM i use the same configuration of ovirt, 2 cores, 4 Gb of ram
>>>> and 60 Gb disk(in the default xenserver SR)
>>>>
>>>> *- For ovirt, can you share the vm command line, available in
>>>> /var/log/libvirt/qemu/vmname.**log?*
>>>> 2016-09-01 12:50:28.268+: starting up libvirt version: 1.2.17,
>>>> package: 13.el7_2.5 (CentOS BuildSystem <http://bugs.centos.org>,
>>>> 2016-06-23-14:23:27, worker1.bsys.centos.org), qemu version: 2.3.0
>>>> (qemu-kvm-ev-2.3.0-31.el7.16.1)
>>>> LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
>>>> QEMU_AUDIO_DRV=none /usr/libexec/qemu-kvm -name vmcentos -S -machine
>>>> pc-i440fx-rhel7.2.0,accel=kvm,usb=off -cpu Haswell-noTSX -m
>>>> size=4194304k,slots=16,maxmem=4294967296k -realtime mlock=off -smp
>>>> 2,maxcpus=16,sockets=16,cores=1,threads=1 -numa
>>>> node,nodeid=0,cpus=0-1,mem=4096 -uuid 21872e4b-7699-4502-b1ef-2c058eff1c3c
>>>> -smbios type=1,manufacturer=oVirt,product=oVirt
>>>> Node,version=7-2.1511.el7.centos.2.10,serial=03AA02FC-0414-0
>>>> 5F8-D906-710700080009,uuid=21872e4b-7699-4502-b1ef-2c058eff1c3c
>>>> -no-user-config -nodefaults -chardev socket,id=charmonitor,path=/va
>>>> r/lib/libvirt/qemu/domain-vmcentos/monitor.sock,server,nowait -mon

Re: [ovirt-users] Ovirt with bad IO performance

2016-09-02 Thread Gabriel Ozaki
Hi Yaniv

Sorry guys, i don't explain well on my first mail, i notice a bad IO
performance on *disk* benchmarks, the network are working really fine







2016-09-02 12:04 GMT-03:00 Yaniv Kaul <yk...@redhat.com>:

> On Fri, Sep 2, 2016 at 5:33 PM, Gabriel Ozaki <gabriel.oz...@kemi.com.br>
> wrote:
>
>> Hi Nir, thanks for the answer
>>
>>
>> *The nfs server is in the host?*
>> Yes, i choose NFS to use as storage on ovirt host
>>
>> *- Is this 2.9GiB/s or 2.9 MiB/s?*
>> Is MiB/s, i put the full test on paste bin
>> centos guest on ovirt:
>> http://pastebin.com/d48qfvuf
>>
>> centos guest on xenserver:
>> http://pastebin.com/gqN3du29
>>
>> how the test works:
>> https://www.howtoforge.com/how-to-benchmark-your-system-cpu-
>> file-io-mysql-with-sysbench
>>
>> *- Are you testing using NFS in all versions?*
>> i am using the v3 version
>>
>> *- What is the disk format?*
>> partion size format
>> /20Gb xfs
>> swap 2 Gb xfs
>> /dados rest of disk xfs   (note, this is the partition where i save the
>> ISOs,exports and VM disks)
>>
>>
>> *- How do you test io on the host?*
>> I do a clean install of centos and do the test before i install the ovirt
>> the test:
>> http://pastebin.com/7RKU7778
>>
>> *- What kind of nic is used? (1G, 10G?)*
>> Is only a 100mbps :(
>>
>
> 100Mbps will not get you more than several MB/s. 11MB/s on a very bright
> day...
>
>>
>> *We need much more details to understand what do you test here.*
>> I have problems to upload the benchmark test on orvirt to novabench site,
>> so here is the screenshot(i make a mistake on the last email i get the
>> wrong value), is 86 Mb/s:
>>
>
> Which is not possible on the wire. Unless it's VM to VM? And the storage
> is local, which means it's the bandwidth of the physical disk itself?
> Y.
>
>
>
>>
>> ​
>> And the novabench on xenserver:
>> https://novabench.com/compare.php?id=ba8dd628e4042dfc1f3d396
>> 70b164ab11061671
>>
>> *- For Xenserver - detailed description of the vm and the storage
>> configuration?*
>> The host is the same(i install xenserver, do the tests before i install
>> centos), the VM i use the same configuration of ovirt, 2 cores, 4 Gb of ram
>> and 60 Gb disk(in the default xenserver SR)
>>
>> *- For ovirt, can you share the vm command line, available in
>> /var/log/libvirt/qemu/vmname.**log?*
>> 2016-09-01 12:50:28.268+: starting up libvirt version: 1.2.17,
>> package: 13.el7_2.5 (CentOS BuildSystem <http://bugs.centos.org>,
>> 2016-06-23-14:23:27, worker1.bsys.centos.org), qemu version: 2.3.0
>> (qemu-kvm-ev-2.3.0-31.el7.16.1)
>> LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
>> QEMU_AUDIO_DRV=none /usr/libexec/qemu-kvm -name vmcentos -S -machine
>> pc-i440fx-rhel7.2.0,accel=kvm,usb=off -cpu Haswell-noTSX -m
>> size=4194304k,slots=16,maxmem=4294967296k -realtime mlock=off -smp
>> 2,maxcpus=16,sockets=16,cores=1,threads=1 -numa
>> node,nodeid=0,cpus=0-1,mem=4096 -uuid 21872e4b-7699-4502-b1ef-2c058eff1c3c
>> -smbios type=1,manufacturer=oVirt,product=oVirt
>> Node,version=7-2.1511.el7.centos.2.10,serial=03AA02FC-0414-
>> 05F8-D906-710700080009,uuid=21872e4b-7699-4502-b1ef-2c058eff1c3c
>> -no-user-config -nodefaults -chardev socket,id=charmonitor,path=/va
>> r/lib/libvirt/qemu/domain-vmcentos/monitor.sock,server,nowait -mon
>> chardev=charmonitor,id=monitor,mode=control -rtc
>> base=2016-09-01T09:50:28,driftfix=slew -global
>> kvm-pit.lost_tick_policy=discard -no-hpet -no-shutdown -boot strict=on
>> -device piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
>> virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x3 -device
>> virtio-serial-pci,id=virtio-serial0,max_ports=16,bus=pci.0,addr=0x4
>> -drive file=/rhev/data-center/mnt/ovirt.kemi.intranet:_dados_iso/
>> 52ee9f87-9d38-48ec-8003-193262f81994/images/--
>> --/CentOS-7-x86_64-NetInstall-1511.iso,
>> if=none,id=drive-ide0-1-0,readonly=on,format=raw -device
>> ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0,bootindex=2
>> -drive file=/rhev/data-center/0001-0001-0001-0001-02bb/
>> 4ccdd1f3-ee79-4425-b6ed-5774643003fa/images/2ecfcf18-
>> ae84-4e73-922f-28b9cda9e6e1/800f05bf-23f7-4c9d-8c1d-
>> b2503592875f,if=none,id=drive-virtio-disk0,format=raw,
>> serial=2ecfcf18-ae84-4e73-922f-28b9cda9e6e1,cache=none,
>> werror=stop,rerror=stop,aio=threads -device
>> virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virti
>&

Re: [ovirt-users] Ovirt with bad IO performance

2016-09-02 Thread Gabriel Ozaki
Hi Nir, thanks for the answer


*The nfs server is in the host?*
Yes, i choose NFS to use as storage on ovirt host

*- Is this 2.9GiB/s or 2.9 MiB/s?*
Is MiB/s, i put the full test on paste bin
centos guest on ovirt:
http://pastebin.com/d48qfvuf

centos guest on xenserver:
http://pastebin.com/gqN3du29

how the test works:
https://www.howtoforge.com/how-to-benchmark-your-system-cpu-file-io-mysql-with-sysbench

*- Are you testing using NFS in all versions?*
i am using the v3 version

*- What is the disk format?*
partion size format
/20Gb xfs
swap 2 Gb xfs
/dados rest of disk xfs   (note, this is the partition where i save the
ISOs,exports and VM disks)


*- How do you test io on the host?*
I do a clean install of centos and do the test before i install the ovirt
the test:
http://pastebin.com/7RKU7778

*- What kind of nic is used? (1G, 10G?)*
Is only a 100mbps :(

*We need much more details to understand what do you test here.*
I have problems to upload the benchmark test on orvirt to novabench site,
so here is the screenshot(i make a mistake on the last email i get the
wrong value), is 86 Mb/s:

​
And the novabench on xenserver:
https://novabench.com/compare.php?id=ba8dd628e4042dfc1f3d39670b164ab11061671

*- For Xenserver - detailed description of the vm and the storage
configuration?*
The host is the same(i install xenserver, do the tests before i install
centos), the VM i use the same configuration of ovirt, 2 cores, 4 Gb of ram
and 60 Gb disk(in the default xenserver SR)

*- For ovirt, can you share the vm command line, available in
/var/log/libvirt/qemu/vmname.**log?*
2016-09-01 12:50:28.268+: starting up libvirt version: 1.2.17, package:
13.el7_2.5 (CentOS BuildSystem <http://bugs.centos.org>,
2016-06-23-14:23:27, worker1.bsys.centos.org), qemu version: 2.3.0
(qemu-kvm-ev-2.3.0-31.el7.16.1)
LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
QEMU_AUDIO_DRV=none /usr/libexec/qemu-kvm -name vmcentos -S -machine
pc-i440fx-rhel7.2.0,accel=kvm,usb=off -cpu Haswell-noTSX -m
size=4194304k,slots=16,maxmem=4294967296k -realtime mlock=off -smp
2,maxcpus=16,sockets=16,cores=1,threads=1 -numa
node,nodeid=0,cpus=0-1,mem=4096 -uuid 21872e4b-7699-4502-b1ef-2c058eff1c3c
-smbios type=1,manufacturer=oVirt,product=oVirt
Node,version=7-2.1511.el7.centos.2.10,serial=03AA02FC-0414-05F8-D906-710700080009,uuid=21872e4b-7699-4502-b1ef-2c058eff1c3c
-no-user-config -nodefaults -chardev
socket,id=charmonitor,path=/var/lib/libvirt/qemu/domain-vmcentos/monitor.sock,server,nowait
-mon chardev=charmonitor,id=monitor,mode=control -rtc
base=2016-09-01T09:50:28,driftfix=slew -global
kvm-pit.lost_tick_policy=discard -no-hpet -no-shutdown -boot strict=on
-device piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x3 -device
virtio-serial-pci,id=virtio-serial0,max_ports=16,bus=pci.0,addr=0x4 -drive
file=/rhev/data-center/mnt/ovirt.kemi.intranet:_dados_iso/52ee9f87-9d38-48ec-8003-193262f81994/images/----/CentOS-7-x86_64-NetInstall-1511.iso,if=none,id=drive-ide0-1-0,readonly=on,format=raw
-device
ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0,bootindex=2 -drive
file=/rhev/data-center/0001-0001-0001-0001-02bb/4ccdd1f3-ee79-4425-b6ed-5774643003fa/images/2ecfcf18-ae84-4e73-922f-28b9cda9e6e1/800f05bf-23f7-4c9d-8c1d-b2503592875f,if=none,id=drive-virtio-disk0,format=raw,serial=2ecfcf18-ae84-4e73-922f-28b9cda9e6e1,cache=none,werror=stop,rerror=stop,aio=threads
-device
virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1
-chardev
socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/21872e4b-7699-4502-b1ef-2c058eff1c3c.com.redhat.rhevm.vdsm,server,nowait
-device
virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=com.redhat.rhevm.vdsm
-chardev
socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/21872e4b-7699-4502-b1ef-2c058eff1c3c.org.qemu.guest_agent.0,server,nowait
-device
virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel1,name=org.qemu.guest_agent.0
-device usb-tablet,id=input0 -vnc 192.168.0.189:0,password -k pt-br -device
VGA,id=video0,vgamem_mb=16,bus=pci.0,addr=0x2 -device
virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x5 -msg timestamp=on
2016-09-01T12:50:28.307173Z qemu-kvm: warning: CPU(s) not present in any
NUMA nodes: 2 3 4 5 6 7 8 9 10 11 12 13 14 15
2016-09-01T12:50:28.307371Z qemu-kvm: warning: All CPU(s) up to maxcpus
should be described in NUMA config
qemu: terminating on signal 15 from pid 1
2016-09-01 19:13:47.899+: shutting down


Thanks







2016-09-02 11:05 GMT-03:00 Nir Soffer <nsof...@redhat.com>:

> On Fri, Sep 2, 2016 at 4:44 PM, Gabriel Ozaki <gabriel.oz...@kemi.com.br>
> wrote:
>
>> Hi
>> i am trying Ovirt 4.0 and i am getting some strange results when
>> comparing with Xenserver
>>
>> **The host machine*
>> Intel Core i5-4440 3.1

[ovirt-users] Ovirt with bad IO performance

2016-09-02 Thread Gabriel Ozaki
Hi
i am trying Ovirt 4.0 and i am getting some strange results when comparing
with Xenserver

**The host machine*
Intel Core i5-4440 3.10GHz running at 3093 MHz
8 Gb of RAM (1x8)
500 Gb of Disk (seagate st500dm002 7200rpm)
CentOS 7 (netinstall for the most updated and stable packages)



**How i am testing:*
I choose two benchmark tools, sysbench(epel-repo on centos) and
novabench(for windows guest, https://novabench.com ), then i make a clean
install of xenserver and create two guests(CentOS and Windows 7 SP1)


**The Guest specs*
2 cores
4 Gb of RAM
60 Gb of disk (using virtIO in a NFS storage)
Important note: only the testing guest are up on benchmark and i have
installed the drivers in guest


**The Sysbench disk test(creates 10Gb of data and do the bench):*
# sysbench --test=fileio --file-total-size=10G prepare
# sysbench --test=fileio --file-total-size=10G --file-test-mode=rndrw
--init-rng=on --max-time=300 --max-requests=0 run

Host result:2.9843Mb/sec
Ovirt result:   1.1561Mb/sec
Xenserver result:   2.9006Mb/sec


**The novabench test:*
Ovirt result:  79Mb/s
Xenserver result:  101Mb/s


I do other tests, like cpu, memory and etc, ovirt wins with a small
advantage(less then 1%)


Adicional info:
i use this command to install ovirt repo
yum install http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm

i follow this steps on installation:
http://jensd.be/550/linux/install-ovirt-on-centos-7
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users