On 02/17/2014 08:19 PM, Brandeburg, Jesse wrote: > Forgive my top post. > > With the new kernel you may be running into needing faster cleanup to > increase tx speed. try increasing the interrupt rate via ethtool -C ethX > rx-usecs 10, yes I said rx because there is only one rate control for the > interrupt. > > You can easily do line rate tx with 82599. The biggest limiter in tx only > tests is the amount of data in flight and the time it takes to get acks back. >
Thanks for the suggestions. So I tried upping the rx-usecs on the server install instance to 10 (originally 1) and saw a clear bump up to what I would consider line rate ~9.36Gbs. Switching to 10 usecs sounds like its a decrease in interrupt rate though. Interestingly I tested my live iso version of ubuntu 12.04.4 desktop again and see an ~9.39Gbs line rate with out any tuning (default ixgbe driver 3.13.10, default rx-usecs@1, same upstream iperf server). Switching to the rx-usecs=10 on this platform degraded the performance, to 8.69Gbs. Ubuntu no longer maintains separate desktop and server kernels, so I'm trusting all the core-kernel operation would be identical. Thus the live iso test is likely as pristine an experience as can be had, wrt stock performance. It'd take it if I could get it. ;) I'd take from this that there is some functionality or setting introduce in an actual system install that's introducing a hit on performance. Any thoughts? > Also please make sure you have run the set_irq_affinity script to bind > interrupts to CPUs. > I tried running `set_irg_affinity eth4` but it didn't appear to have any impact on performance. If anything it degraded. > -- > Jesse Brandeburg > > >> On Feb 17, 2014, at 5:42 PM, "Ben Greear" <gree...@candelatech.com> wrote: >> >>> On 02/17/2014 02:19 PM, John-Paul Robinson wrote: >>> Hi, >>> >>> I don't know if this topic is appropriate here, please direct me to a >>> better place if not. >>> >>> I've been spending considerable time trying to measure the performance >>> of our 10G fabric that uses Intel X520 cards. The primary test machine >>> has dual Intel(R) Xeon(R) CPU E5-2650 0 @ 2.00GHz chip 8-core chips and >>> 96GB RAM. >>> >>> The test machine is now running Ubuntu 12.04.4 server with kernel 3.11.0 >>> with latest ixgbe driver 3.19.1. >>> >>> Using iperf (2.0.5) I see about 9.39Gbs steady inbound transfers (there >>> are a few glitches where I've seen drop to 7Gbs but it recovers). My >>> outbound transfers, however, are about 8.83Gbs steady and tend to be >>> more variable. >>> >>> This is the best performance I can get on the server. >>> >>> Interestingly when I boot the machine off the live CDROM image for >>> Ubuntu 12.04.4 desktop, I see nice steady 9.39Gbs in both directions. >>> This is the best performance i have seen with this card to-date. >>> >>> I've spent a lot of time with these cards and in general they have be >>> very finicky, delivering inconsistent results from test to test, being >>> very sensitive to driver and kernel versions. >>> >>> I've taken them from extremely erratic performance on Ubuntu 12.04.1 >>> with the stock ixgbe 3.6.7 driver to much higher, more stable >>> performance simply by updating to ixgbe 3.11.33. It would be nice to >>> see a stable flatline performance at line speeds on kernel 3.11 with the >>> 3.19.1 driver. >>> >>> I'm wondering if there is a known configuration profile that allows >>> these cards to perform at line speeds or if there are known issues or >>> hardware incompatibilities. >>> >>> I know there are a lot of subtleties to performance tuning but >>> performance on other cards in our fabric (btw from Brocade) deliver very >>> consistent, stable, high performance line speed results over many tests. >>> >>> I've been scratching my head for a while and am looking for a fresh >>> perspective or deeper understanding. >> >> First, check 'dmesg' and make sure your NICs are using at least >> x8 pci with 5GT/s. >> >> Check BIOS and disable 'VT-d' if it is on...it hurts performance by >> 50% or so. >> >> Try using several (5-10) flows in iperf, maybe just use 5-10 >> instances of iperf so you get good usage of your cores. >> >> Thanks, >> Ben >> >> >> -- >> Ben Greear <gree...@candelatech.com> >> Candela Technologies Inc http://www.candelatech.com >> >> >> ------------------------------------------------------------------------------ >> Managing the Performance of Cloud-Based Applications >> Take advantage of what the Cloud has to offer - Avoid Common Pitfalls. >> Read the Whitepaper. >> http://pubads.g.doubleclick.net/gampad/clk?id=121054471&iu=/4140/ostg.clktrk >> _______________________________________________ >> E1000-devel mailing list >> E1000-devel@lists.sourceforge.net >> https://lists.sourceforge.net/lists/listinfo/e1000-devel >> To learn more about Intel® Ethernet, visit >> http://communities.intel.com/community/wired ------------------------------------------------------------------------------ Managing the Performance of Cloud-Based Applications Take advantage of what the Cloud has to offer - Avoid Common Pitfalls. Read the Whitepaper. http://pubads.g.doubleclick.net/gampad/clk?id=121054471&iu=/4140/ostg.clktrk _______________________________________________ E1000-devel mailing list E1000-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/e1000-devel To learn more about Intel® Ethernet, visit http://communities.intel.com/community/wired