Hi,

Thanks for info. Glad the issue was resolved.

Best regards,
Dariusz Sosnowski

> From: Tanmay Pandey <[email protected]> 
> Sent: Friday, April 26, 2024 14:22
> To: Dariusz Sosnowski <[email protected]>
> Cc: [email protected]
> Subject: Re: Performance Bottleneck at NIC with Openshift
> 
> External email: Use caution opening links or attachments 
> 
> Hey Darius,
> 
> Thanks for the help. Actually the issue was very silly – our SriovNetwork 
> resource had a maxTxRate and minTxRate specified due to which it was capping 
> > the generation in the first place. 
> 
> Removing that fixed the generation cap. 
> 
> Regards
> Tanmay 
> 
> 
> From: Dariusz Sosnowski <mailto:[email protected]>
> Date: Thursday, 25 April 2024 at 10:08 PM
> To: Tanmay Pandey <mailto:[email protected]>
> Cc: mailto:[email protected] <mailto:[email protected]>
> Subject: RE: Performance Bottleneck at NIC with Openshift
> Hi,
> 
> Since as you mentioned, the similar HW, the same DPDK and PRoX versions were 
> able to achieve much better performance,
> I'd guess that the problem might be related to how processes in pods are 
> being scheduled with OpenShift. Specifically, I would:
> 
> - Check if both pods are not scheduled on the same cores.
> - Verify if cores on which these pods are running, are isolated.
> 
> Anything interrupting threads responsible for generating traffic will hurt 
> the performance.
> 
> Best regards,
> Dariusz Sosnowski
> 
> > From: Tanmay Pandey <mailto:[email protected]> 
> > Sent: Friday, April 5, 2024 14:41
> > To: mailto:[email protected]
> > Subject: Performance Bottleneck at NIC with Openshift
> > 
> > External email: Use caution opening links or attachments 
> > 
> > Hi,
> > 
> > I am using DPDK version 22.11 for performance evaluation running PRoX on an 
> > Openshift Cluster where I have created two pods - I am sending traffic > 
> > from one and receiving on the other and I've found that I'm unable to 
> > utilize more than 6GB of bandwidth in the server at the packet generation 
> > level. > I have tested with a 64-byte frame size and achieved a maximum of 
> > 6.99 MPPS.
> > I've attempted to address this issue by adhering to the recommendations 
> > outlined in the DPDK 22.11 NVIDIA Mellanox NIC performance report available 
> > > at 
> > https://fast.dpdk.org/doc/perf/DPDK_22_11_NVIDIA_Mellanox_NIC_performance_report.pdf
> >  . However, the problem persists.
> > Additionally, I've investigated packet loss at the NIC interface level and 
> > found no anomalies. The bottleneck appears to be related to packet > 
> > generation, but I'm uncertain about the underlying cause.
> > I am very new to DPDK so don't really know how to debug this issue. I 
> > believe there is something happening between the NIC layer and Openshift. 
> > Additionally, I used the same hardware running kubeadm where I was using 
> > the same DPDK and PRoX version with a similar setup and was able to achieve 
> > > much better performance(at least for the packet generation part - where 
> > my current bottleneck occurs).
> > Can someone point me in the right direction? 
> > I would be happy to provide any other required information
> > Below are the SUT details:
> > Nic Model: Ethernet controller: Mellanox Technologies MT2892 Family 
> > [ConnectX-6 Dx]
> > uname -r
> > 5.14.0-284.54.1.rt14.339.el9_2.x86_64
> > 
> > ethtool -i enp216s0f0np0
> > driver: mlx5_core
> > version: 5.14.0-284.54.1.rt14.339.el9_2.
> > firmware-version: 22.35.2000 (MT_0000000359)
> > expansion-rom-version:
> > bus-info: 0000:d8:00.0
> > supports-statistics: yes
> > supports-test: yes
> > supports-eeprom-access: no
> > supports-register-dump: no
> > supports-priv-flags: yes
> > 
> > ## CPU
> > Architecture:            x86_64
> >   CPU op-mode(s):        32-bit, 64-bit
> >   Address sizes:         46 bits physical, 48 bits virtual
> >   Byte Order:            Little Endian
> > CPU(s):                  104
> >   On-line CPU(s) list:   0-103
> > Vendor ID:               GenuineIntel
> >   BIOS Vendor ID:        Intel
> >   Model name:            Intel(R) Xeon(R) Gold 6230R CPU @ 2.10GHz
> >     BIOS Model name:     Intel(R) Xeon(R) Gold 6230R CPU @ 2.10GHz
> > Operating System:
> > cat /etc/os-release
> > NAME="Red Hat Enterprise Linux CoreOS"
> > ID="rhcos"
> > ID_LIKE="rhel fedora"
> > VERSION="415.92.202402201450-0"
> > VERSION_ID="4.15"
> > VARIANT="CoreOS"
> > VARIANT_ID=coreos
> > PLATFORM_ID="platform:el9"
> > PRETTY_NAME="Red Hat Enterprise Linux CoreOS 415.92.202402201450-0 (Plow)"
> > ANSI_COLOR="0;31"
> > CPE_NAME="cpe:/o:redhat:enterprise_linux:9::coreos"
> > HOME_URL="https://www.redhat.com/";
> > DOCUMENTATION_URL="https://docs.openshift.com/container-platform/4.15/";
> > BUG_REPORT_URL="https://bugzilla.redhat.com/";
> > REDHAT_BUGZILLA_PRODUCT="OpenShift Container Platform"
> > REDHAT_BUGZILLA_PRODUCT_VERSION="4.15"
> > REDHAT_SUPPORT_PRODUCT="OpenShift Container Platform"
> > REDHAT_SUPPORT_PRODUCT_VERSION="4.15"
> > OPENSHIFT_VERSION="4.15"
> > RHEL_VERSION="9.2"
> > OSTREE_VERSION="415.92.202402201450-0"
> > OCP Cluster
> > oc version
> > Client Version: 4.15.0-202402070507.p0.g48dcf59.assembly.stream-48dcf59
> > Kustomize Version: v5.0.4-0.20230601165947-6ce0bf390ce3

Reply via email to