Interesting. Who’s 10g cards and which offload settings did you disable? Did 
you do that on the servers or the vm host clients or both?

> On Apr 15, 2019, at 11:37 AM, Alex McWhirter <[email protected]> wrote:
>> I went in and disabled TCP offload on all the nics, huge performance boost. 
>> went from 110MB/s to 240MB/s seq writes, reads lost a bit of performance 
>> going down to 680MB/s, but that's a decent trade off. Latency is still 
>> really high though, need to work on that. I think some more TCP tuning might 
>> help.
>> 
>>  
> Those changes didn't do a whole lot, but i ended up enabling 
> performance.read-ahead on the gluster volume. my blockdev read ahead values 
> were already 8192, which seemed good enough. Not sure if ovirt set those, or 
> if it's just the defaults of my raid controller.
> 
> Anyways up to 350MB/s writes, 700MB/s reads. Which so happens to correlate 
> with the saturation of my 10G network. Latency is still a slight issue, but 
> at least now im not blocking :)
> 
>  
> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives: 
> https://lists.ovirt.org/archives/list/[email protected]/message/5COPHAIVCVK42KMMGWZQVMNGDH6Q32ZC/

_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/T3QMRYHIDRZPUTW4QMGGVOCJ3S3VHLRY/

Reply via email to