On 2019-04-15 12:58, Alex McWhirter wrote:

> On 2019-04-15 12:43, Darrell Budic wrote: Interesting. Who's 10g cards and 
> which offload settings did you disable? Did you do that on the servers or the 
> vm host clients or both? 
> 
> On Apr 15, 2019, at 11:37 AM, Alex McWhirter <a...@triadic.us> wrote: 
> 
> I went in and disabled TCP offload on all the nics, huge performance boost. 
> went from 110MB/s to 240MB/s seq writes, reads lost a bit of performance 
> going down to 680MB/s, but that's a decent trade off. Latency is still really 
> high though, need to work on that. I think some more TCP tuning might help.
> 
> Those changes didn't do a whole lot, but i ended up enabling 
> performance.read-ahead on the gluster volume. my blockdev read ahead values 
> were already 8192, which seemed good enough. Not sure if ovirt set those, or 
> if it's just the defaults of my raid controller. 
> 
> Anyways up to 350MB/s writes, 700MB/s reads. Which so happens to correlate 
> with the saturation of my 10G network. Latency is still a slight issue, but 
> at least now im not blocking :)
> 
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives: 
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/5COPHAIVCVK42KMMGWZQVMNGDH6Q32ZC/

These are dual port qlogic QLGE cards, plugging into dual Cisco Nexus
3064's with VPC to allow me to LACP across two switches. These are
FCoE/10GBE cards, so on the Cisco Switches i had to disable lldp on the
ports to stop FCoE initiator errors from disabling ports (as i don't use
FCoE atm) 

bond options are "mode=4 lacp_rate=1 miimon=100 xmit_hash_policy=1" 

then i have the following /sbin/ifup-local script that triggers on
storage network creation 

#!/bin/bash
case "$1" in
  Storage)
    /sbin/ethtool -K ens2f0 tx off rx off tso off gso off
    /sbin/ethtool -K ens2f1 tx off rx off tso off gso off
    /sbin/ip link set dev ens2f0 txqueuelen 10000
    /sbin/ip link set dev ens2f1 txqueuelen 10000
    /sbin/ip link set dev bond2 txqueuelen 10000
    /sbin/ip link set dev Storage txqueuelen 10000
  ;;
  *)
  ;;
esac
exit 0 

if you have lro, disable it too IMO, these cards do not do lro so it's
not applicable to me. 

This did cut down my read performance by about 50MB/s, but my write went
from 98-110MB/s to about 240MB/s, then enabling read-ahead got me to the
350MB/s it should have been.

Oh and i did it on both, the VM hosts and storage machines. Same cards
in all of them.
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/EZF2G3AWM3GSDR35CLP2SKABALQ6H4KN/

Reply via email to