> On Sep 28, 2017, at 9:40 PM, Andrew Theurer <[email protected]> wrote: > > In our tests, ~36Mpps is the maximum we can get. We usually run a test with > TRex, bidirectional, 2 pci cards, 1 port per x8 gen3 PCI adapter, with a > device under test using same HW config but running testomd with 2 or more > queues per port. Bidirectional aggregate traffic is in the 72Mpps range. > So, in that test, each active port is transmitting and receiving ~36Mpps, > however, I don't believe the received packets are DMA'd to memory, just > counted on the adapter. I have never observed the Fortville doing higher > than that.
40Gbits is the limit and I think 36Mpps is the MAX for the PCI I think, if I remember correctly. The t-rex must be counting differently as you stated. I need to ask some folks here. I have two 40G NICs, but at this time I do not have enough slots to put in the other 40 and keep my 10Gs in the system. I need to fix the problem below, but have not had the chance. > > -Andrew > > On Thu, Sep 28, 2017 at 3:59 PM, Wiles, Keith <[email protected]> wrote: > > > On Sep 28, 2017, at 6:06 AM, Mauricio Valdueza <[email protected]> wrote: > > > > Hi Guys; > > > > I am testing a Fortville 40Gb nic with PKTgen > > > > I see linerate in 40Gb with 156B packet size, but once I decrease size, > > linerate is far away > > In Pktgen the packet count is taken from the hardware registers on the NIC > and the bit rate is calculated using those values. Not all NICs flush the TX > done queue and from one start command to the next the numbers can be off as > the old packets are being recycled with the new size packets. Please try the > different sizes and bring down pktgen between runs just to see if that is the > problem. > > > > > WITH 158B > > Link State : <UP-40000-FD> ----TotalRate---- > > Pkts/s Max/Rx : 0/0 0/0 > > Max/Tx : 28090480/28089840 28090480/28089840 > > MBits/s Rx/Tx : 0/40000 0/40000 > > ----------------------------------------------------------------------------------------------- > > > > WITH 128B > > Link State : <UP-40000-FD> ----TotalRate---- > > Pkts/s Max/Rx : 0/0 0/0 > > Max/Tx : 33784179/33783908 33784179/33783908 > > MBits/s Rx/Tx : 0/40000 0/40000 > > ------------------------------------------------------------------------------------------------ > > > > With 64B > > Link State : <UP-40000-FD> ----TotalRate---- > > Pkts/s Max/Rx : 0/0 0/0 > > Max/Tx : 35944587/35941680 35944587/35941680 > > MBits/s Rx/Tx : 0/24152 0/24152 > > ---------------------------------------------------------------------------------------------- > > > > Should I run any optimization? > > > > My environment is: > > > > •VMware ESXi version: 6.5.0, 4887370 > > •Exact NIC version: Intel Corporation XL710 for 40GbE QSFP+ > > •NIC driver version: i40en version 1.3.1 > > •Server Vendor: Dell > > •Server Make: Dell Inc. PowerEdge R730 > > •CPU Model: I ntel(R) Xeon(R) CPU E5-2697 v3 @ > > 2.60GHz > > •Huge pages size: 2M > > •Test VM: What is it? Ubuntu 16.04 > > • DPDK is compiled there? dpdk-17.08 > > •Test traffic kind: IP/UDP? Both tested > > Traffic generator: Intel pktgen version? pktgen-3.4.1 > > > > > > I am executing: > > > > sudo ./app/x86_64-native-linuxapp-gcc/pktgen -c 0xff n 3 --proc-type auto > > --socket-mem 9096 -- -m "[1:2-7].0" --crc-strip > > > > > > Thanks in advance > > > > > > mauricio > > > > Regards, > Keith > > Regards, Keith
