So I've finally completed my first HCI build using the below configuration:

3x
Dell PowerEdge R720
2x 2.9 GHz 8 Core E5-2690
256GB RAM
2x250gb SSD Raid 1 (boot/os)
2x2TB SSD jbod passthrough (used for gluster bricks)
1Gbe Nic for management 10Gbe nic for Gluster

Using Replica 3 with no arbiter.

Installed the latest version of oVirt available at the time 4.2.5.  Created
recommended volumes (with an additional data volume on second SSD). Not
using VDO

First thing I did was setup glusterFS network on 10Gbe and set it to be
used for glusterFS and migration traffic.

I've setup a single test VM using Centos7 minimal on the default "x-large
instance" profile.

Within this VM if I do very basic write test using something like:

dd bs=1M count=256 if=/dev/zero of=test conv=fdatasync

I'm seeing quite slow speeds, only 8mb/sec.

If I do the same from one of the hosts gluster mounts i.e.

host1: /rhev/data-center/mnt/glusterSD/HOST:data

I get about 30mb/sec (which still seems fairly low?)

Am I testing incorrectly here?  Is there anything I should be tuning on the
Gluster volumes to increase performance with SSDs?  Where can I find out
where the bottle neck is here, or is this expected performance of Gluster?
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/NNEAHYFAQONRYM3AORLV7OHGPWLAD5G7/

Reply via email to