On Wed, Mar 20, 2019, 1:16 PM Darrell Budic <bu...@onholyground.com> wrote:

> Inline:
>
> On Mar 20, 2019, at 4:25 AM, Roy Golan <rgo...@redhat.com> wrote:
>
> On Mon, 18 Mar 2019 at 22:14, Darrell Budic <bu...@onholyground.com>
> wrote:
>
>> I agree, been checking some of my more disk intensive VMs this morning,
>> switching them to noop definitely improved responsiveness. All the virtio
>> ones I’ve found were using deadline (with RHEL/Centos guests), but some of
>> the virt-scsi were using deadline and some were noop, so I’m not sure of a
>> definitive answer on that level yet.
>>
>> For the hosts, it depends on what your backend is running. With a
>> separate storage server on my main cluster, it doesn’t matter what the
>> hosts set for me. You mentioned you run hyper converged, so I’d say it
>> depends on what your disks are. If you’re using SSDs, go none/noop as they
>> don’t benefit from the queuing. If they are HDDs, I’d test cfq or deadline
>> and see which gave better latency and throughput to your vms. I’d guess
>> you’ll find deadline to offer better performance, but cfq to share better
>> amongst multiple VMs. Unless you use ZFS underneath, then go noop and let
>> ZFS take care of it.
>>
>> On Mar 18, 2019, at 2:05 PM, Strahil <hunter86...@yahoo.com> wrote:
>>
>> Hi Darrel,
>>
>> Still, based on my experience we shouldn't queue our I/O in the VM, just
>> to do the same in the Host.
>>
>> I'm still considering if I should keep deadline  in my hosts or to switch
>> to 'cfq'.
>> After all, I'm using Hyper-converged oVirt and this needs testing.
>> What I/O scheduler  are  you using on the  host?
>>
>>
> Our internal scale team is testing now 'throughput-performance' tuned
> profile and it gives
> promising results, I suggest you try it as well.
> We will go over the results of a comparison against the virtual-guest
> profile
> , if there will be evidence for improvements we will set it as the default
> (if it won't degrade small,medium scale envs).
>
>
> I don’t think that will make a difference in this case. Both virtual-host
> and virtual-guest include the throughput-performance profile, just with
> “better” virtual memory tunings for guest and hosts. None of those 3 modify
> the disk queue schedulers, by default, at least not on my Centos 7.6
> systems.
>
> Re my testing, I have virtual-host on my hosts and virtual-guest on my
> guests already.
>

Unfortunately, the ideal scheduler really depends on storage configuration.
Gluster, ZFS, iSCSI, FC, and NFS don't align on a single "best"
configuration (to say nothing of direct LUNs on guests), then there's
workload considerations.

The scale team is aiming for a balanced "default" policy rather than one
which is best for a specific environment.

That said, I'm optimistic that the results will let us give better
recommendations if your workload/storage benefits from a different scheduler


>
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/FH5LLYXSEJKXTVVOAZCSMV6AAU33CNCA/
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/74CWGUYUKEKSV3ANEGGEE2L5GJZVCN23/

Reply via email to