Hello,
Please, help me get rid of rx_fifo_errors.
I got couple of LVS servers, Intel I350-T4 adapter, all 4 interfaces bonded
together, 1 Gbit/s RX on peak load, 800k RX packets, 30% CPU load , 100
errors/s, kernel 2.6.39.4 , openSUSE 10.2
Usually I start seeing errors when traffic reach 600 Mbit/s & 500k packets.
I used iperf to test what is wrong - 1.5 Gbit/s RX and 1M was not enough to
produce at least one rx_fifo_error.
Found useful e-mail
http://www.mail-archive.com/[email protected]/msg04299.html ,
but it doesn't help me.
modinfo igb
filename: /lib/modules/2.6.39.4-net-07/kernel/drivers/net/igb/igb.ko
version: 4.1.2
license: GPL
description: Intel(R) Gigabit Ethernet Network Driver
author: Intel Corporation, <[email protected]>
srcversion: DBABE4876BBED027F305396
lspci | grep -i eth
02:00.0 Ethernet controller: Intel Corporation 82574L Gigabit Network Connection
03:00.0 Ethernet controller: Intel Corporation 82574L Gigabit Network Connection
08:00.0 Ethernet controller: Intel Corporation I350 Gigabit Network Connection
(rev 01)
08:00.1 Ethernet controller: Intel Corporation I350 Gigabit Network Connection
(rev 01)
08:00.2 Ethernet controller: Intel Corporation I350 Gigabit Network Connection
(rev 01)
08:00.3 Ethernet controller: Intel Corporation I350 Gigabit Network Connection
(rev 01)
cat /etc/modprobe.conf.local
#
# please add local extensions to this file
#
options igb RSS=6,6,6,6,6,6
08:00.0 Ethernet controller: Intel Corporation I350 Gigabit Network Connection
(rev 01)
Subsystem: Intel Corporation Ethernet Server Adapter I350-T4
Control: I/O- Mem+ BusMaster+ SpecCycle- MemWINV- VGASnoop- ParErr-
Stepping- SERR+ FastB2B-
Status: Cap+ 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort- <TAbort-
<MAbort- >SERR- <PERR-
Latency: 0, Cache Line Size: 256 bytes
Interrupt: pin A routed to IRQ 30
Region 0: Memory at fb800000 (32-bit, non-prefetchable) [size=1M]
Region 3: Memory at fb77c000 (32-bit, non-prefetchable) [size=16K]
Expansion ROM at fb780000 [disabled] [size=512K]
Capabilities: [40] Power Management version 3
Flags: PMEClk- DSI+ D1- D2- AuxCurrent=0mA
PME(D0+,D1-,D2-,D3hot+,D3cold+)
Status: D0 PME-Enable- DSel=0 DScale=1 PME-
Capabilities: [50] Message Signalled Interrupts: Mask+ 64bit+ Queue=0/0
Enable-
Address: 0000000000000000 Data: 0000
Masking: 00000000 Pending: 00000000
Capabilities: [70] MSI-X: Enable+ Mask- TabSize=10
Vector table: BAR=3 offset=00000000
PBA: BAR=3 offset=00002000
Capabilities: [a0] Express Endpoint IRQ 0
Device: Supported: MaxPayload 512 bytes, PhantFunc 0, ExtTag-
Device: Latency L0s <512ns, L1 <64us
Device: AtnBtn- AtnInd- PwrInd-
Device: Errors: Correctable- Non-Fatal- Fatal- Unsupported-
Device: RlxdOrd- ExtTag- PhantFunc- AuxPwr- NoSnoop+
Device: MaxPayload 256 bytes, MaxReadReq 512 bytes
Link: Supported Speed unknown, Width x4, ASPM L0s L1, Port 0
Link: Latency L0s <4us, L1 <32us
Link: ASPM Disabled RCB 64 bytes CommClk+ ExtSynch-
Link: Speed unknown, Width x4
ethtool -S eth2
NIC statistics:
rx_packets: 447885957564
tx_packets: 452967126934
rx_bytes: 67873379483451
tx_bytes: 68756350830290
rx_broadcast: 428921
tx_broadcast: 0
rx_multicast: 1551532
tx_multicast: 277100
multicast: 1551532
collisions: 0
rx_crc_errors: 0
rx_no_buffer_count: 22993
rx_missed_errors: 15207708
tx_aborted_errors: 0
tx_carrier_errors: 0
tx_window_errors: 0
tx_abort_late_coll: 0
tx_deferred_ok: 0
tx_single_coll_ok: 0
tx_multi_coll_ok: 0
tx_timeout_count: 0
rx_long_length_errors: 0
rx_short_length_errors: 0
rx_align_errors: 0
tx_tcp_seg_good: 120
tx_tcp_seg_failed: 0
rx_flow_control_xon: 0
rx_flow_control_xoff: 0
tx_flow_control_xon: 0
tx_flow_control_xoff: 0
rx_long_byte_count: 67873379483451
tx_dma_out_of_sync: 0
lro_aggregated: 0
lro_flushed: 0
lro_recycled: 0
tx_smbus: 0
rx_smbus: 0
dropped_smbus: 0
os2bmc_rx_by_bmc: 0
os2bmc_tx_by_bmc: 0
os2bmc_tx_by_host: 0
os2bmc_rx_by_host: 0
rx_errors: 0
tx_errors: 0
tx_dropped: 0
rx_length_errors: 0
rx_over_errors: 0
rx_frame_errors: 0
rx_fifo_errors: 15207708
tx_fifo_errors: 0
tx_heartbeat_errors: 0
tx_queue_0_packets: 90661746199
tx_queue_0_bytes: 13121025270748
tx_queue_0_restart: 7195
tx_queue_1_packets: 87116922206
tx_queue_1_bytes: 12606476905526
tx_queue_1_restart: 6373
tx_queue_2_packets: 87119613756
tx_queue_2_bytes: 12611199722139
tx_queue_2_restart: 6420
tx_queue_3_packets: 87117808078
tx_queue_3_bytes: 12613371344837
tx_queue_3_restart: 6600
tx_queue_4_packets: 87921390354
tx_queue_4_bytes: 12653843205269
tx_queue_4_restart: 7046
tx_queue_5_packets: 13029864072
tx_queue_5_bytes: 1885102419867
tx_queue_5_restart: 0
rx_queue_0_packets: 76985209776
rx_queue_0_bytes: 11360654634336
rx_queue_0_drops: 0
rx_queue_0_csum_err: 499288
rx_queue_0_alloc_failed: 0
rx_queue_1_packets: 76992526831
rx_queue_1_bytes: 11362461767970
rx_queue_1_drops: 0
rx_queue_1_csum_err: 495089
rx_queue_1_alloc_failed: 0
rx_queue_2_packets: 73491590861
rx_queue_2_bytes: 10841294846552
rx_queue_2_drops: 0
rx_queue_2_csum_err: 478361
rx_queue_2_alloc_failed: 0
rx_queue_3_packets: 73476036502
rx_queue_3_bytes: 10841575376811
rx_queue_3_drops: 0
rx_queue_3_csum_err: 470166
rx_queue_3_alloc_failed: 0
rx_queue_4_packets: 73472725484
rx_queue_4_bytes: 10839168689362
rx_queue_4_drops: 0
rx_queue_4_csum_err: 471490
rx_queue_4_alloc_failed: 0
rx_queue_5_packets: 73468408695
rx_queue_5_bytes: 10836758716694
rx_queue_5_drops: 0
rx_queue_5_csum_err: 479575
rx_queue_5_alloc_failed: 0
ethtool -k eth2
Offload parameters for eth2:
rx-checksumming: on
tx-checksumming: on
scatter-gather: on
tcp-segmentation-offload: on
udp-fragmentation-offload: off
generic-segmentation-offload: off
generic-receive-offload: off
large-receive-offload: off
rx-vlan-offload: on
tx-vlan-offload: on
ntuple-filters: off
receive-hashing: on
ethtool -g eth2
Ring parameters for eth2:
Pre-set maximums:
RX: 4096
RX Mini: 0
RX Jumbo: 0
TX: 4096
Current hardware settings:
RX: 4096
RX Mini: 0
RX Jumbo: 0
TX: 4096
cat /proc/net/bonding/bond0
Ethernet Channel Bonding Driver: v3.7.0 (June 2, 2010)
Bonding Mode: IEEE 802.3ad Dynamic link aggregation
Transmit Hash Policy: layer2+3 (2)
MII Status: up
MII Polling Interval (ms): 100
Up Delay (ms): 0
Down Delay (ms): 0
802.3ad info
LACP rate: slow
Aggregator selection policy (ad_select): stable
Active Aggregator Info:
Aggregator ID: 1
Number of ports: 4
Actor Key: 17
Partner Key: 9
Partner Mac Address: c4:71:fe:2b:ba:00
Slave Interface: eth2
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 1
Permanent HW addr: a0:36:9f:0c:b8:b8
Aggregator ID: 1
Slave queue ID: 0
Slave Interface: eth3
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 1
Permanent HW addr: a0:36:9f:0c:b8:b9
Aggregator ID: 1
Slave queue ID: 0
Slave Interface: eth4
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 1
Permanent HW addr: a0:36:9f:0c:b8:ba
Aggregator ID: 1
Slave queue ID: 0
Slave Interface: eth5
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 1
Permanent HW addr: a0:36:9f:0c:b8:bb
Aggregator ID: 1
Slave queue ID: 0
C-states:
<4>using polling idle threads.
<4>WARNING: polling idle and HT enabled, performance may degrade.
Thanks in advance!
Ilya Schanikov
odnoklassniki.ru
------------------------------------------------------------------------------
Minimize network downtime and maximize team effectiveness.
Reduce network management and security costs.Learn how to hire
the most talented Cisco Certified professionals. Visit the
Employer Resources Portal
http://www.cisco.com/web/learning/employer_resources/index.html
_______________________________________________
E1000-devel mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/e1000-devel
To learn more about Intel® Ethernet, visit
http://communities.intel.com/community/wired