Inline:

> On Mar 20, 2019, at 4:25 AM, Roy Golan <rgo...@redhat.com> wrote:
> 
> On Mon, 18 Mar 2019 at 22:14, Darrell Budic <bu...@onholyground.com 
> <mailto:bu...@onholyground.com>> wrote:
> I agree, been checking some of my more disk intensive VMs this morning, 
> switching them to noop definitely improved responsiveness. All the virtio 
> ones I’ve found were using deadline (with RHEL/Centos guests), but some of 
> the virt-scsi were using deadline and some were noop, so I’m not sure of a 
> definitive answer on that level yet. 
> 
> For the hosts, it depends on what your backend is running. With a separate 
> storage server on my main cluster, it doesn’t matter what the hosts set for 
> me. You mentioned you run hyper converged, so I’d say it depends on what your 
> disks are. If you’re using SSDs, go none/noop as they don’t benefit from the 
> queuing. If they are HDDs, I’d test cfq or deadline and see which gave better 
> latency and throughput to your vms. I’d guess you’ll find deadline to offer 
> better performance, but cfq to share better amongst multiple VMs. Unless you 
> use ZFS underneath, then go noop and let ZFS take care of it.
> 
>> On Mar 18, 2019, at 2:05 PM, Strahil <hunter86...@yahoo.com 
>> <mailto:hunter86...@yahoo.com>> wrote:
>> 
>> Hi Darrel,
>> 
>> Still, based on my experience we shouldn't queue our I/O in the VM, just to 
>> do the same in the Host.
>> 
>> I'm still considering if I should keep deadline  in my hosts or to switch to 
>> 'cfq'.
>> After all, I'm using Hyper-converged oVirt and this needs testing.
>> What I/O scheduler  are  you using on the  host?
>> 
> 
> 
> Our internal scale team is testing now 'throughput-performance' tuned profile 
> and it gives
> promising results, I suggest you try it as well.
> We will go over the results of a comparison against the virtual-guest profile
> , if there will be evidence for improvements we will set it as the default 
> (if it won't degrade small,medium scale envs). 

I don’t think that will make a difference in this case. Both virtual-host and 
virtual-guest include the throughput-performance profile, just with “better” 
virtual memory tunings for guest and hosts. None of those 3 modify the disk 
queue schedulers, by default, at least not on my Centos 7.6 systems.

Re my testing, I have virtual-host on my hosts and virtual-guest on my guests 
already.


_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/FH5LLYXSEJKXTVVOAZCSMV6AAU33CNCA/

Reply via email to