My VMs are using virtual-guest tuned profiles and ovirt node hosts are
using virtual-host profile.  Those seem to be good defaults from what I'm
looking at.  I will test I/O schedulers to see if that makes any difference
and also try out high performance VM profile (I was staying away from that
profile due to loss of high-availability).

On Fri, Oct 18, 2019 at 9:18 AM Jayme <jay...@gmail.com> wrote:

> The VMs are basically as stock CentOS 7x as you can get.  There are so
> many layers to deal with in HCI it's difficult to know where to begin with
> tuning.  I was focusing mainly on gluster.  Is it recommended to do tuning
> directly on oVirt host nodes as well such as I/O scheduler and tuned-adm
> profiles etc?
>
> On Fri, Oct 18, 2019 at 6:55 AM Strahil <hunter86...@yahoo.com> wrote:
>
>> What is  your  I/O scheduler and tuned-adm profile in the VM.
>> RedHat based VMs use deadline which prioritizes reads before writes  ->
>> you can use 'noop' or 'none'.
>>
>> For profile, you can use high-performance.
>>
>> Best Regards,
>> Strahil Nikolov
>> On Oct 18, 2019 06:45, Jayme <jay...@gmail.com> wrote:
>>
>> I'm wondering if anyone has any tips to improve file/directory operations
>> in HCI replica 3 (no arbtr) configuration with SSDs and 10Gbe storage
>> network.
>>
>> I am running stock optimize for virt store volume settings currently and
>> am wondering what if any improvements I can make for VM write speed and
>> more specifically anything I can tune to increase performance of small file
>> operations such as copying, untar, npm installs etc.
>>
>> For some context, I'm seeing ~50MB/s write speeds inner VM with: dd
>> if=/dev/zero of=./test bs=512k count=2048 oflag=direct -- I am not sure how
>> this compares to other HCI setups, I feel like it should be higher with SSD
>> backed storage.  Same command from gluster mount is over 400MB/s
>>
>> I've read some things about meta data caching, read ahead and other
>> options.  There are so many and I'm not sure where to start, I'm also not
>> sure which could potentially have a negative impact on VM
>> stability/reliability.
>>
>> Here are options for one of my volumes:
>>
>> Volume Name: prod_b
>> Type: Replicate
>> Volume ID: c3e7447e-8514-4e4a-9ff5-a648fe6aa537
>> Status: Started
>> Snapshot Count: 0
>> Number of Bricks: 1 x 3 = 3
>> Transport-type: tcp
>> Bricks:
>> Brick1: gluster0.example.com:/gluster_bricks/prod_b/prod_b
>> Brick2: gluster1.example.com:/gluster_bricks/prod_b/prod_b
>> Brick3: gluster2.example.com:/gluster_bricks/prod_b/prod_b
>> Options Reconfigured:
>> server.event-threads: 4
>> client.event-threads: 4
>> performance.client-io-threads: on
>> nfs.disable: on
>> transport.address-family: inet
>> performance.quick-read: off
>> performance.read-ahead: off
>> performance.io-cache: off
>> performance.low-prio-threads: 32
>> network.remote-dio: off
>> cluster.eager-lock: enable
>> cluster.quorum-type: auto
>> cluster.server-quorum-type: server
>> cluster.data-self-heal-algorithm: full
>> cluster.locking-scheme: granular
>> cluster.shd-max-threads: 8
>> cluster.shd-wait-qlength: 10000
>> features.shard: on
>> user.cifs: off
>> storage.owner-uid: 36
>> storage.owner-gid: 36
>> network.ping-timeout: 30
>> performance.strict-o-direct: on
>> cluster.granular-entry-heal: enable
>> server.allow-insecure: on
>> cluster.choose-local: off
>>
>>
>>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZIDNBQMAZABKLDPA4KNJCTRNITQABQ37/

Reply via email to