So to close the loop on mail list, after offline discussions, this has to do 
with configuring bonding and VLAN's.  This was not an issue with the drivers or 
the HW.

Thanks.

Cheers,
John

From: Jekels, Donny [mailto:[email protected]]
Sent: Wednesday, January 30, 2013 9:47 AM
To: Ronciak, John
Cc: [email protected]
Subject: Re: [E1000-devel] ARPING does not work

John,

Here is what I am talking about, all basic network functionality works.

we also have set the arp_announce in the kernel settings to 1 that is supposed 
to fix this behavior and it does but not on these NIC's
the IP that I used for this testing is 10.8.198.250

when I plumb it up and arping from it does not respond to pings.

it ONLY starts working after I initate ping -I 10.8.198.250 10.11.11.11

then the IP on the interface works.

BUT arping fails completely


djekels@postgres19:~$ ip route
10.33.0.0/24<http://10.33.0.0/24>
          nexthop via 10.4.42.1  dev eth4 weight 1
          nexthop via 10.4.42.1  dev eth5 weight 1
          nexthop via 10.4.42.1  dev eth6 weight 1
          nexthop via 10.4.42.1  dev eth7 weight 1
10.8.198.0/24<http://10.8.198.0/24> dev vlan198  proto kernel  scope link  src 
10.8.198.119
10.8.199.0/24<http://10.8.199.0/24> dev vlan199  proto kernel  scope link  src 
10.8.199.39
10.32.0.0/24<http://10.32.0.0/24>
          nexthop via 10.4.42.1  dev eth4 weight 1
          nexthop via 10.4.42.1  dev eth5 weight 1
          nexthop via 10.4.42.1  dev eth6 weight 1
          nexthop via 10.4.42.1  dev eth7 weight 1
10.4.42.0/24<http://10.4.42.0/24> dev eth4  proto kernel  scope link  src 
10.4.42.96
10.4.42.0/24<http://10.4.42.0/24> dev eth5  proto kernel  scope link  src 
10.4.42.97
10.4.42.0/24<http://10.4.42.0/24> dev eth6  proto kernel  scope link  src 
10.4.42.98
10.4.42.0/24<http://10.4.42.0/24> dev eth7  proto kernel  scope link  src 
10.4.42.99
172.23.16.0/23<http://172.23.16.0/23> dev vlan4000  proto kernel  scope link  
src 172.23.16.211
172.23.0.0/16<http://172.23.0.0/16> via 172.23.16.1 dev vlan4000
default via 10.8.198.1 dev vlan198
djekels@postgres19:~$
djekels@postgres19:~$ ip addr
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN
    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
    inet 127.0.0.1/8<http://127.0.0.1/8> scope host lo
    inet6 ::1/128 scope host
       valid_lft forever preferred_lft forever
2: eth0: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc mq master bond0 
state UP qlen 1000
    link/ether bc:30:5b:f0:a3:80 brd ff:ff:ff:ff:ff:ff
3: eth1: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc mq master bond0 
state UP qlen 1000
    link/ether bc:30:5b:f0:a3:80 brd ff:ff:ff:ff:ff:ff
4: eth2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
    link/ether bc:30:5b:f0:a3:82 brd ff:ff:ff:ff:ff:ff
5: eth3: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
    link/ether bc:30:5b:f0:a3:83 brd ff:ff:ff:ff:ff:ff
6: eth4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP qlen 1000
    link/ether a0:36:9f:08:09:c8 brd ff:ff:ff:ff:ff:ff
    inet 10.4.42.96/24<http://10.4.42.96/24> brd 10.4.42.255 scope global eth4
    inet6 fe80::a236:9fff:fe08:9c8/64 scope link
       valid_lft forever preferred_lft forever
7: eth5: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP qlen 1000
    link/ether a0:36:9f:08:09:c9 brd ff:ff:ff:ff:ff:ff
    inet 10.4.42.97/24<http://10.4.42.97/24> brd 10.4.42.255 scope global eth5
    inet6 fe80::a236:9fff:fe08:9c9/64 scope link
       valid_lft forever preferred_lft forever
8: eth6: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP qlen 1000
    link/ether a0:36:9f:08:09:ca brd ff:ff:ff:ff:ff:ff
    inet 10.4.42.98/24<http://10.4.42.98/24> brd 10.4.42.255 scope global eth6
    inet6 fe80::a236:9fff:fe08:9ca/64 scope link
       valid_lft forever preferred_lft forever
9: eth7: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc mq state UP qlen 1000
    link/ether a0:36:9f:08:09:cb brd ff:ff:ff:ff:ff:ff
    inet 10.4.42.99/24<http://10.4.42.99/24> brd 10.4.42.255 scope global eth7
    inet6 fe80::a236:9fff:fe08:9cb/64 scope link
       valid_lft forever preferred_lft forever
10: bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc noqueue 
state UP
    link/ether bc:30:5b:f0:a3:80 brd ff:ff:ff:ff:ff:ff
    inet6 fe80::be30:5bff:fef0:a380/64 scope link
       valid_lft forever preferred_lft forever
11: vlan4000@bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc 
noqueue state UP
    link/ether bc:30:5b:f0:a3:80 brd ff:ff:ff:ff:ff:ff
    inet 172.23.16.211/23<http://172.23.16.211/23> brd 172.23.17.255 scope 
global vlan4000
    inet6 fe80::be30:5bff:fef0:a380/64 scope link
       valid_lft forever preferred_lft forever
12: vlan198@bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc 
noqueue state UP
    link/ether bc:30:5b:f0:a3:80 brd ff:ff:ff:ff:ff:ff
    inet 10.8.198.119/24<http://10.8.198.119/24> brd 10.8.198.255 scope global 
vlan198
    inet 10.8.198.134/24<http://10.8.198.134/24> brd 10.8.198.255 scope global 
secondary vlan198
    inet 10.8.198.167/24<http://10.8.198.167/24> brd 10.8.198.255 scope global 
secondary vlan198
    inet 10.8.198.250/24<http://10.8.198.250/24> brd 10.8.198.255 scope global 
secondary vlan198
    inet6 fe80::be30:5bff:fef0:a380/64 scope link
       valid_lft forever preferred_lft forever
13: vlan199@bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc 
noqueue state UP
    link/ether bc:30:5b:f0:a3:80 brd ff:ff:ff:ff:ff:ff
    inet 10.8.199.39/24<http://10.8.199.39/24> brd 10.8.199.255 scope global 
vlan199
    inet 10.8.199.105/24<http://10.8.199.105/24> brd 10.8.199.255 scope global 
secondary vlan199
    inet 10.8.199.108/24<http://10.8.199.108/24> brd 10.8.199.255 scope global 
secondary vlan199
    inet6 fe80::be30:5bff:fef0:a380/64 scope link
       valid_lft forever preferred_lft forever



djekels@postgres19:~$ netstat -s
Ip:
    71751543 total packets received
    166 with invalid addresses
    0 forwarded
    0 incoming packets discarded
    71706086 incoming packets delivered
    75075912 requests sent out
    10 dropped because of missing route
Icmp:
    3645 ICMP messages received
    6 input ICMP message failed.
    ICMP input histogram:
        destination unreachable: 1716
        echo requests: 1851
        echo replies: 78
    56066 ICMP messages sent
    0 ICMP messages failed
    ICMP output histogram:
        destination unreachable: 53169
        echo request: 1046
        echo replies: 1851
IcmpMsg:
        InType0: 78
        InType3: 1716
        InType8: 1851
        OutType0: 1851
        OutType3: 53169
        OutType8: 1046
Tcp:
    708305 active connections openings
    3975 passive connection openings
    431 failed connection attempts
    33564 connection resets received
    14 connections established
    68956581 segments received
    72313084 segments send out
    24468 segments retransmited
    0 bad segments received.
    30446 resets sent
Udp:
    2671621 packets received
    74239 packets to unknown port received.
    0 packet receive errors
    2682294 packets sent
UdpLite:
TcpExt:
    396 resets received for embryonic SYN_RECV sockets
    1075 packets pruned from receive queue because of socket buffer overrun
    174110 TCP sockets finished time wait in fast timer
    93 time wait sockets recycled by time stamp
    1547391 delayed acks sent
    47 delayed acks further delayed because of locked socket
    Quick ack mode was activated 887 times
    723398 packets directly queued to recvmsg prequeue.
    5496 bytes directly in process context from backlog
    4362730 bytes directly received in process context from prequeue
    49343700 packet headers predicted
    17131 packets header predicted and directly queued to user
    1638076 acknowledgments not containing data payload received
    31754543 predicted acknowledgments
    181 times recovered from packet loss by selective acknowledgements
    52 congestion windows recovered without slow start by DSACK
    1023 congestion windows recovered without slow start after partial ack
    191 TCP data loss events
    TCPLostRetransmit: 15
    114 timeouts after SACK recovery
    2 timeouts in loss state
    314 fast retransmits
    15 forward retransmits
    54 retransmits in slow start
    12748 other TCP timeouts
    13 SACK retransmits failed
    385169 packets collapsed in receive queue due to low socket buffer
    874 DSACKs sent for old packets
    236 DSACKs received
    8 connections reset due to unexpected data
    29985 connections reset due to early user close
    91 connections aborted due to timeout
    TCPDSACKIgnoredOld: 168
    TCPDSACKIgnoredNoUndo: 9
    TCPSackShifted: 259
    TCPSackMerged: 503
    TCPSackShiftFallback: 530
IpExt:
    InBcastPkts: 37803
    InOctets: -1264740945
    OutOctets: -592937989
    InBcastOctets: 9434134



On Wed, Jan 30, 2013 at 11:34 AM, Ronciak, John 
<[email protected]<mailto:[email protected]>> wrote:
So the NIC's are sending and receiving packets.  What is the stack doing 
('netstat -s')?  My guess is that it's not doing the right thing with broadcast 
packets.  Do you have a resolv.conf file configured to do what you want?  Maybe 
it's not correct?

Cheers,
John

From: Jekels, Donny [mailto:[email protected]<mailto:[email protected]>]
Sent: Wednesday, January 30, 2013 9:10 AM
To: Ronciak, John
Cc: [email protected]<mailto:[email protected]>
Subject: Re: [E1000-devel] ARPING does not work

John,

The NIC works for TCP and UDP sessions however no broadcast and arp seems to 
work.

we move IP's around between servers and part of this move is to garp the source 
IP so the switches can update their mac tables.
without this functionality the NIC's are useless to us.

please help

root@postgres19:~# ethtool -S eth7
NIC statistics:
     rx_packets: 11
     tx_packets: 6
     rx_bytes: 704
     tx_bytes: 492
     rx_broadcast: 11
     tx_broadcast: 0
     rx_multicast: 0
     tx_multicast: 6
     multicast: 0
     collisions: 0
     rx_crc_errors: 0
     rx_no_buffer_count: 0
     rx_missed_errors: 0
     tx_aborted_errors: 0
     tx_carrier_errors: 0
     tx_window_errors: 0
     tx_abort_late_coll: 0
     tx_deferred_ok: 0
     tx_single_coll_ok: 0
     tx_multi_coll_ok: 0
     tx_timeout_count: 0
     rx_long_length_errors: 0
     rx_short_length_errors: 0
     rx_align_errors: 0
     tx_tcp_seg_good: 0
     tx_tcp_seg_failed: 0
     rx_flow_control_xon: 0
     rx_flow_control_xoff: 0
     tx_flow_control_xon: 0
     tx_flow_control_xoff: 0
     rx_long_byte_count: 704
     tx_dma_out_of_sync: 0
     lro_aggregated: 0
     lro_flushed: 0
     lro_recycled: 0
     tx_smbus: 0
     rx_smbus: 0
     dropped_smbus: 0
     os2bmc_rx_by_bmc: 0
     os2bmc_tx_by_bmc: 0
     os2bmc_tx_by_host: 0
     os2bmc_rx_by_host: 0
     rx_errors: 0
     tx_errors: 0
     tx_dropped: 0
     rx_length_errors: 0
     rx_over_errors: 0
     rx_frame_errors: 0
     rx_fifo_errors: 0
     tx_fifo_errors: 0
     tx_heartbeat_errors: 0
     tx_queue_0_packets: 6
     tx_queue_0_bytes: 468
     tx_queue_0_restart: 0
     rx_queue_0_packets: 11
     rx_queue_0_bytes: 660
     rx_queue_0_drops: 0
     rx_queue_0_csum_err: 0
     rx_queue_0_alloc_failed: 0

root@postgres19:~# ethtool -i eth7
driver: igb
version: 4.1.2
firmware-version: 1.6, 0x80000816
bus-info: 0000:04:00.3

root@postgres19:~# arping 10.11.12.13
ARPING 10.11.12.13



On Wed, Jan 30, 2013 at 11:05 AM, Ronciak, John 
<[email protected]<mailto:[email protected]>> wrote:
Hi Donny,

Sorry to hear you are having problems.  This most likely is not a NIC/driver 
issue.  Something is mostly not configured correctly on these systems.  Did you 
use ethtool to look at the HW stats of the interfaces you are trying to ping 
out of?  If not, please do so.  Also, are each port/interface on a different 
subnet?  The stack does not do well (without extra configuration) with multiple 
ports on the same subnet.

The reason Debian Squeeze doesn't have support is that they didn't pick up an 
updated driver before they released it.  This is not a problem with the Intel 
drivers, it's a timing thing for Debian releases and how they update kernels 
and drivers.

Cheers,
John


> -----Original Message-----
> From: Jekels, Donny [mailto:[email protected]<mailto:[email protected]>]
> Sent: Tuesday, January 29, 2013 8:35 PM
> To: 
> [email protected]<mailto:[email protected]>
> Subject: [E1000-devel] ARPING does not work
>
> We have a handful of these Intel quad port NIC with the drive 4.1.7 and
> arping does not work on these NIC's.
>
> we run debian stock squeeze and had to compile igb driver from intel to
> get the NIC ordering correct.
>
> now arping is also not working.
>
> can you give me some advise how do I continue having this issue
> resolved?
>
> thanks,
> Donny


------------------------------------------------------------------------------
Everyone hates slow websites. So do we.
Make your web apps faster with AppDynamics
Download AppDynamics Lite for free today:
http://p.sf.net/sfu/appdyn_d2d_jan
_______________________________________________
E1000-devel mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/e1000-devel
To learn more about Intel&#174; Ethernet, visit 
http://communities.intel.com/community/wired

Reply via email to