I think this may be fixed with a newer NVM image. In any case, the newer NVM images will help with other issues including performance. Go to downloadcenter.intel.com and search for xl710 and you should find an updater and instructions.
Todd Fujinaka Software Application Engineer Networking Division (ND) Intel Corporation todd.fujin...@intel.com (503) 712-4565 -----Original Message----- From: Dmitry Mikhaylov [mailto:d...@in-solve.ru] Sent: Monday, March 09, 2015 6:55 AM To: e1000-devel@lists.sourceforge.net Subject: [E1000-devel] TX driver issue detected, PF reset issued - debug results Hi Team, I would like to report my experience so somebody may have a real debugging shortcut next time. There is also a practical wish to developers. We have moved to local 10G for storage based on X710 DA-2 two weeks ago. Right from the start we have lots of these events: kernel: [772391.372876] i40e 0000:07:00.1: TX driver issue detected, PF reset issued kernel: [772391.594984] i40e 0000:07:00.1: i40e_ptp_init: added PHC on enp7s0f1 kernel: [772391.741089] i40e 0000:07:00.1 enp7s0f1: NIC Link is Up 10 Gbps Full Duplex, Flow Control: None kernel: [772391.752696] i40e 0000:07:00.1 enp7s0f1: NIC Link is Down kernel: [772392.557943] i40e 0000:07:00.1 enp7s0f1: NIC Link is Up 10 Gbps Full Duplex, Flow Control: None The server works normally, but there can be 500+ of these events In a row, giving 5 seconds working / 5 seconds not working experience. While the next hour we may have zero of these. But as a rule, 100+ events daily. I shall to note that the driver resets just fine. No crashes, visible links or whatever. The reset method itself seems to be OK. The problem is that the network performance in such a mode is surely unacceptable. Initial setup was i40e <http://sourceforge.net/projects/e1000/files/i40e%20stable/1.1.23/> 1.1.23 as of Intel official downloads. Our next try was with your <http://sourceforge.net/projects/e1000/files/i40e%20stable/1.2.37/> 1.2.37 released a few weeks ago because you said in the list that "something of that sort was fixed". Zero behavior change as it seems. The problem however WAS solved with ethtool -K enp7s0f1 tso off that attempt was based on http://sourceforge.net/p/e1000/bugs/407/ thread, related to completely other setup with different card and different driver, but still the problem seems the same (for years?) and the solution remains. Working setup is here: l31 ~ # ethtool -k enp7s0f1 Offload parameters for enp7s0f1: rx-checksumming: on tx-checksumming: on scatter-gather: on tcp-segmentation-offload: off udp-fragmentation-offload: off generic-segmentation-offload: on generic-receive-offload: on large-receive-offload: off rx-vlan-offload: on tx-vlan-offload: on ntuple-filters: on receive-hashing: on Not working setup differs in this only: tcp-segmentation-offload: on I would really suggest this problem be either fixed at last (not so important) OR noted in some READMEs. That's really a show stopper and somebody with debugging karma and try-end-see luck may end up with non-working adapter. Really I think that in local unified networks with no segmentation this feature is irrelevant. But if someone is facing this 10G card to internet and thus expecting some real segmentation, real pain in the performance side may arise with lack of the feature. Maybe changing card defaults may be appropriate, although not sure about it as it is bad to change it in the middle of driver life cycle. But X710 is on the very start for now, so maybe?... Other aux. info that may help: We are using DA SFP+. One port of the card is attached. Traffic is somehow floating from 1 to ~7 Gbps and the hang is NOT related to traffic, just peer TCP behavior (as it seems after solution). We have only local network (used for iSCSI), JUMBO enabled with MTU9000; MTUs are fine, we have 0.0001% packet fragmentation. It seems that packet gets fragmented only on some really odd occasion. No VLANs. No teaming. Very simple plain setup with two 10G switches, 10 hosts (5 here and 5 there) and SFP+ DA all the way (some errors and drops in the links "as usual", but not perceptible to end apps). This exactly card in debug was in Intel R2312GZ4GCSAS server system. The card is installed on the PCI raiser, together with some LSI-based Intel RAID on the same raiser. enp7s0f1 Link encap:Ethernet HWaddr 68:05:ca:30:5b:c9 inet addr:10.21.0.87 Bcast:10.21.0.255 Mask:255.255.255.0 UP BROADCAST RUNNING MULTICAST MTU:9000 Metric:1 RX packets:520305857 errors:0 dropped:0 overruns:0 frame:0 TX packets:1450598509 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:1797022578878 (1.6 TiB) TX bytes:8995312721942 (8.1 TiB) l31 ~ # ethtool enp7s0f1 Settings for enp7s0f1: Supported ports: [ FIBRE ] Supported link modes: 10000baseT/Full Supported pause frame use: Symmetric Supports auto-negotiation: No Advertised link modes: Not reported Advertised pause frame use: No Advertised auto-negotiation: No Speed: 10000Mb/s Duplex: Full Port: Direct Attach Copper PHYAD: 0 Transceiver: external Auto-negotiation: off Supports Wake-on: d Wake-on: d Current message level: 0x0000000f (15) drv probe link timer Link detected: yes l31 ~ # ethtool -i enp7s0f1 driver: i40e version: 1.2.37 firmware-version: f4.22.27454 a1.2 n4.25 e143f bus-info: 0000:07:00.1 supports-statistics: yes supports-test: yes supports-eeprom-access: yes supports-register-dump: yes l31 ~ # ethtool -k enp7s0f1 Offload parameters for enp7s0f1: rx-checksumming: on tx-checksumming: on scatter-gather: on tcp-segmentation-offload: off udp-fragmentation-offload: off generic-segmentation-offload: on generic-receive-offload: on large-receive-offload: off rx-vlan-offload: on tx-vlan-offload: on ntuple-filters: on receive-hashing: on I will add some more info if required to help you to understand things better if you wish. Dmitry. ------------------------------------------------------------------------------ Dive into the World of Parallel Programming The Go Parallel Website, sponsored by Intel and developed in partnership with Slashdot Media, is your hub for all things parallel software development, from weekly thought leadership blogs to news, videos, case studies, tutorials and more. Take a look and join the conversation now. http://goparallel.sourceforge.net/ _______________________________________________ E1000-devel mailing list E1000-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/e1000-devel To learn more about Intel® Ethernet, visit http://communities.intel.com/community/wired