If you seek performance, then set tuned-adm profile in the VM to 'throughput-performance' and scheduler either 'noop' or 'none' (depends if multi queue is enabled).
Usually, if yhou crete the gluster cluster via cockpit and then install the hosted engine via cockpit again - all options on your gluster volumes are the most optimal you need. Best Rregards, Strahil NikolovOn Oct 18, 2019 15:30, Jayme <jay...@gmail.com> wrote: > > My VMs are using virtual-guest tuned profiles and ovirt node hosts are using > virtual-host profile. Those seem to be good defaults from what I'm looking > at. I will test I/O schedulers to see if that makes any difference and also > try out high performance VM profile (I was staying away from that profile due > to loss of high-availability). > > On Fri, Oct 18, 2019 at 9:18 AM Jayme <jay...@gmail.com> wrote: >> >> The VMs are basically as stock CentOS 7x as you can get. There are so many >> layers to deal with in HCI it's difficult to know where to begin with >> tuning. I was focusing mainly on gluster. Is it recommended to do tuning >> directly on oVirt host nodes as well such as I/O scheduler and tuned-adm >> profiles etc? >> >> On Fri, Oct 18, 2019 at 6:55 AM Strahil <hunter86...@yahoo.com> wrote: >>> >>> What is your I/O scheduler and tuned-adm profile in the VM. >>> RedHat based VMs use deadline which prioritizes reads before writes -> >>> you can use 'noop' or 'none'. >>> >>> For profile, you can use high-performance. >>> >>> Best Regards, >>> Strahil Nikolov >>> >>> On Oct 18, 2019 06:45, Jayme <jay...@gmail.com> wrote: >>>> >>>> I'm wondering if anyone has any tips to improve file/directory operations >>>> in HCI replica 3 (no arbtr) configuration with SSDs and 10Gbe storage >>>> network. >>>> >>>> I am running stock optimize for virt store volume settings currently and >>>> am wondering what if any improvements I can make for VM write speed and >>>> more specifically anything I can tune to increase performance of small >>>> file operations such as copying, untar, npm installs etc. >>>> >>>> For some context, I'm seeing ~50MB/s write speeds inner VM with: dd >>>> if=/dev/zero of=./test bs=512k count=2048 oflag=direct -- I am not sure >>>> how this compares to other HCI setups, I feel like it should be higher >>>> with SSD backed storage. Same command from gluster mount is over 400MB/s >>>> >>>> I've read some things about meta data caching, read ahead and other >>>> options. There are so many and I'm not sure where to start, I'm also not >>>> sure which could potentially have a negative impact on VM >>>> stability/reliability. >>>> >>>> Here are options for one of my volumes: >>>> >>>> Volume Name: prod_b >>>> Type: Replicate >>>> Volume ID: c3e7447e-8514-4e4a-9ff5-a648fe6aa537 >>>> Status: Started >>>> Snapshot Count: 0 >>>> Number of Bricks: 1 x 3 = 3 >>>> Transport-type: tcp >>>> Bricks: >>>> Brick1: gluster0.example.com:/gluster_bricks/prod_b/prod_b >>>> Brick2: gluster1.example.com:/gluster_bricks/prod_b/prod_b >>>> Brick3: gluster2.example.com:/gluster_bricks/prod_b/prod_b >>>> Options Reconfigured: >>>> server.event-threads: 4 >>>> client.event-threads: 4 >>>> performance.client-io-threads: on >>>> nfs.disable: on >>>> transport.address-family: inet >>>> performance.quick-read: off >>>> performance.read-ahead: off >>>> performance.io-cache: off >>>> performance.low-prio-threads: 32 >>>> network.remote-dio: off >>>> cluster.eager-lock: enable >>>> cluster.quorum-type: auto >>>> cluster.server-quorum-type: server >>>> cluster.data-self-heal-algorithm: full >>>> cluster.locking-scheme: granular >>>> cluster.shd-max-threads: 8 >>>> cluster.shd-wait-qlength: 10000 >>>> features.shard: on >>>> user.cifs: off >>>> storage.owner-uid: 36 >>>> storage.owner-gid: 36 >>>> network.ping-timeout: 30 >>>> performance.strict-o-direct: on >>>> cluster.granular-entry-heal: enable >>>> server.allow-insecure: on >>>> cluster.choose-local: off >>>> >>>>
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/DIZY4SRG2DYJJSR55WXMYBJYWGEPU64R/