Hi, I discovered an issue with my LACP configuration and i am having trouble figuring it out. I am running 2 Dell Powered 610's with 4 broadcomm nics. I am trying to bond them together, however only one of the nics goes active no mater how much traffic i push across the links.
I have spoken to my network admin, and says that the switch ports are configured and can only see one active link on the switch. Thanks Bryan
Ethernet Channel Bonding Driver: v3.7.1 (April 27, 2011) Bonding Mode: IEEE 802.3ad Dynamic link aggregation Transmit Hash Policy: layer2+3 (2) MII Status: up MII Polling Interval (ms): 100 Up Delay (ms): 0 Down Delay (ms): 0 802.3ad info LACP rate: slow Min links: 0 Aggregator selection policy (ad_select): stable System priority: 65535 System MAC address: f4:8e:38:c5:fc:a8 Active Aggregator Info: Aggregator ID: 1 Number of ports: 1 Actor Key: 9 Partner Key: 20 Partner Mac Address: a4:6c:2a:e5:30:00 Slave Interface: em1 MII Status: up Speed: 1000 Mbps Duplex: full Link Failure Count: 0 Permanent HW addr: f4:8e:38:c5:fc:a8 Slave queue ID: 0 Aggregator ID: 1 Actor Churn State: none Partner Churn State: none Actor Churned Count: 0 Partner Churned Count: 0 details actor lacp pdu: system priority: 65535 system mac address: f4:8e:38:c5:fc:a8 port key: 9 port priority: 255 port number: 1 port state: 61 details partner lacp pdu: system priority: 8192 system mac address: a4:6c:2a:e5:30:00 oper key: 20 port priority: 32768 port number: 25 port state: 61 Slave Interface: em2 MII Status: up Speed: 1000 Mbps Duplex: full Link Failure Count: 0 Permanent HW addr: f4:8e:38:c5:fc:a9 Slave queue ID: 0 Aggregator ID: 2 Actor Churn State: churned Actor Churned Count: 1 Partner Churned Count: 1 details actor lacp pdu: system priority: 65535 system mac address: f4:8e:38:c5:fc:a8 port key: 9 port priority: 255 port number: 2 port state: 5 details partner lacp pdu: system priority: 32768 system mac address: a4:6c:2a:e5:30:00 oper key: 20 port priority: 32768 port number: 73 port state: 5 Slave Interface: em3 MII Status: up Speed: 1000 Mbps Duplex: full Link Failure Count: 0 Permanent HW addr: f4:8e:38:c5:fc:aa Slave queue ID: 0 Aggregator ID: 2 Actor Churn State: churned Partner Churn State: churned Actor Churned Count: 1 Partner Churned Count: 1 details actor lacp pdu: system priority: 65535 system mac address: f4:8e:38:c5:fc:a8 port key: 9 port priority: 255 port number: 3 port state: 5 details partner lacp pdu: system priority: 32768 system mac address: a4:6c:2a:e5:30:00 oper key: 20 port priority: 32768 port number: 26 port state: 5 Slave Interface: em4 MII Status: up Speed: 1000 Mbps Duplex: full Link Failure Count: 0 Permanent HW addr: f4:8e:38:c5:fc:ab Slave queue ID: 0 Aggregator ID: 2 Actor Churn State: churned Partner Churn State: churned Actor Churned Count: 1 Partner Churned Count: 1 details actor lacp pdu: oper key: 20 port priority: 32768 port number: 73 port state: 5 Slave Interface: em3 MII Status: up Speed: 1000 Mbps Duplex: full Link Failure Count: 0 Permanent HW addr: f4:8e:38:c5:fc:aa Slave queue ID: 0 Aggregator ID: 2 Actor Churn State: churned Partner Churn State: churned Actor Churned Count: 1 Partner Churned Count: 1 details actor lacp pdu: system priority: 65535 system mac address: f4:8e:38:c5:fc:a8 port key: 9 port priority: 255 port number: 3 port state: 5 details partner lacp pdu: system priority: 32768 system mac address: a4:6c:2a:e5:30:00 oper key: 20 port priority: 32768 port number: 26 port state: 5 Slave Interface: em4 MII Status: up Speed: 1000 Mbps Duplex: full Link Failure Count: 0 Permanent HW addr: f4:8e:38:c5:fc:ab Slave queue ID: 0 Aggregator ID: 2 Actor Churn State: churned Partner Churn State: churned Actor Churned Count: 1 Partner Churned Count: 1 details actor lacp pdu: system priority: 65535 system mac address: f4:8e:38:c5:fc:a8 port key: 9 port priority: 255 port number: 4 port state: 5 details partner lacp pdu: system priority: 32768 system mac address: a4:6c:2a:e5:30:00 oper key: 20 port priority: 32768 port number: 74 port state: 5
Apr 20 10:51:04 vm-host-colo-2 systemd: Stopping LSB: Bring up/down networking... Apr 20 10:51:04 vm-host-colo-2 kernel: DMZ: port 1(bond0.10) entered disabled state Apr 20 10:51:04 vm-host-colo-2 network: Shutting down interface DMZ: [ OK ] Apr 20 10:51:04 vm-host-colo-2 kernel: Internal-Dev: port 1(bond0.30) entered disabled state Apr 20 10:51:04 vm-host-colo-2 network: Shutting down interface Internal-Dev: [ OK ] Apr 20 10:51:05 vm-host-colo-2 kernel: Lab: port 1(bond0.40) entered disabled state Apr 20 10:51:05 vm-host-colo-2 network: Shutting down interface Lab: [ OK ] Apr 20 10:51:05 vm-host-colo-2 kernel: Server-Net: port 1(bond0.20) entered disabled state Apr 20 10:51:05 vm-host-colo-2 network: Shutting down interface Server-Net: [ OK ] Apr 20 10:51:05 vm-host-colo-2 kernel: Workstation: port 1(bond0.50) entered disabled state Apr 20 10:51:05 vm-host-colo-2 network: Shutting down interface Workstation: [ OK ] Apr 20 10:51:05 vm-host-colo-2 kernel: ovirtmgmt: port 1(bond0) entered disabled state Apr 20 10:51:05 vm-host-colo-2 network: Shutting down interface ovirtmgmt: [ OK ] Apr 20 10:51:06 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): bond0.10: link becomes ready Apr 20 10:51:06 vm-host-colo-2 kernel: device bond0.10 left promiscuous mode Apr 20 10:51:06 vm-host-colo-2 kernel: DMZ: port 1(bond0.10) entered disabled state Apr 20 10:51:06 vm-host-colo-2 network: Shutting down interface bond0.10: [ OK ] Apr 20 10:51:06 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): bond0.102: link becomes ready Apr 20 10:51:06 vm-host-colo-2 network: Shutting down interface bond0.102: [ OK ] Apr 20 10:51:06 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): bond0.20: link becomes ready Apr 20 10:51:06 vm-host-colo-2 kernel: device bond0.20 left promiscuous mode Apr 20 10:51:06 vm-host-colo-2 kernel: Server-Net: port 1(bond0.20) entered disabled state Apr 20 10:51:06 vm-host-colo-2 network: Shutting down interface bond0.20: [ OK ] Apr 20 10:51:06 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): bond0.30: link becomes ready Apr 20 10:51:06 vm-host-colo-2 kernel: device bond0.30 left promiscuous mode Apr 20 10:51:06 vm-host-colo-2 kernel: Internal-Dev: port 1(bond0.30) entered disabled state Apr 20 10:51:06 vm-host-colo-2 network: Shutting down interface bond0.30: [ OK ] Apr 20 10:51:07 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): bond0.40: link becomes ready Apr 20 10:51:07 vm-host-colo-2 kernel: device bond0.40 left promiscuous mode Apr 20 10:51:07 vm-host-colo-2 kernel: Lab: port 1(bond0.40) entered disabled state Apr 20 10:51:07 vm-host-colo-2 network: Shutting down interface bond0.40: [ OK ] Apr 20 10:51:07 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): bond0.50: link becomes ready Apr 20 10:51:07 vm-host-colo-2 kernel: device bond0.50 left promiscuous mode Apr 20 10:51:07 vm-host-colo-2 kernel: Workstation: port 1(bond0.50) entered disabled state Apr 20 10:51:07 vm-host-colo-2 network: Shutting down interface bond0.50: [ OK ] Apr 20 10:51:07 vm-host-colo-2 kernel: bond0: Removing slave em1 Apr 20 10:51:07 vm-host-colo-2 kernel: bond0: Removing an active aggregator Apr 20 10:51:07 vm-host-colo-2 kernel: bond0: Releasing active interface em1 Apr 20 10:51:07 vm-host-colo-2 kernel: bond0: the permanent HWaddr of em1 - f4:8e:38:c5:fc:a8 - is still in use by bond0 - set the HWaddr of em1 to a different address to avoid conflicts Apr 20 10:51:07 vm-host-colo-2 kernel: bond0: first active interface up! Apr 20 10:51:07 vm-host-colo-2 kernel: device em1 left promiscuous mode Apr 20 10:51:08 vm-host-colo-2 kernel: bond0: Removing slave em2 Apr 20 10:51:08 vm-host-colo-2 kernel: bond0: Releasing backup interface em2 Apr 20 10:51:08 vm-host-colo-2 kernel: bond0: first active interface up! Apr 20 10:51:08 vm-host-colo-2 kernel: device em2 left promiscuous mode Apr 20 10:51:08 vm-host-colo-2 kernel: bond0: Removing slave em3 Apr 20 10:51:08 vm-host-colo-2 kernel: bond0: Releasing backup interface em3 Apr 20 10:51:08 vm-host-colo-2 kernel: bond0: first active interface up! Apr 20 10:51:08 vm-host-colo-2 kernel: device em3 left promiscuous mode Apr 20 10:51:08 vm-host-colo-2 kernel: bond0: Removing slave em4 Apr 20 10:51:08 vm-host-colo-2 kernel: bond0: Removing an active aggregator Apr 20 10:51:08 vm-host-colo-2 kernel: bond0: Releasing backup interface em4 Apr 20 10:51:08 vm-host-colo-2 kernel: device em4 left promiscuous mode Apr 20 10:51:08 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): bond0: link becomes ready Apr 20 10:51:08 vm-host-colo-2 kernel: device bond0 left promiscuous mode Apr 20 10:51:08 vm-host-colo-2 kernel: ovirtmgmt: port 1(bond0) entered disabled state Apr 20 10:51:08 vm-host-colo-2 network: Shutting down interface bond0: [ OK ] Apr 20 10:51:09 vm-host-colo-2 network: Shutting down loopback interface: [ OK ] Apr 20 10:51:09 vm-host-colo-2 systemd: Starting LSB: Bring up/down networking... Apr 20 10:51:09 vm-host-colo-2 network: Bringing up loopback interface: [ OK ] Apr 20 10:51:09 vm-host-colo-2 kernel: bond0: Setting MII monitoring interval to 100 Apr 20 10:51:09 vm-host-colo-2 kernel: bond0: Setting xmit hash policy to layer2+3 (2) Apr 20 10:51:09 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_UP): bond0: link is not ready Apr 20 10:51:09 vm-host-colo-2 kernel: 8021q: adding VLAN 0 to HW filter on device bond0 Apr 20 10:51:10 vm-host-colo-2 kernel: bond0: Setting MII monitoring interval to 100 Apr 20 10:51:10 vm-host-colo-2 kernel: bond0: Setting xmit hash policy to layer2+3 (2) Apr 20 10:51:10 vm-host-colo-2 kernel: bond0: Adding slave em1 Apr 20 10:51:10 vm-host-colo-2 kernel: bond0: Enslaving em1 as a backup interface with a down link Apr 20 10:51:10 vm-host-colo-2 kernel: bond0: Adding slave em2 Apr 20 10:51:10 vm-host-colo-2 kernel: bond0: Enslaving em2 as a backup interface with a down link Apr 20 10:51:10 vm-host-colo-2 kernel: bond0: Adding slave em3 Apr 20 10:51:10 vm-host-colo-2 kernel: bond0: Enslaving em3 as a backup interface with a down link Apr 20 10:51:10 vm-host-colo-2 kernel: bond0: Adding slave em4 Apr 20 10:51:10 vm-host-colo-2 kernel: bond0: Enslaving em4 as a backup interface with a down link Apr 20 10:51:10 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_UP): bond0: link is not ready Apr 20 10:51:10 vm-host-colo-2 kernel: 8021q: adding VLAN 0 to HW filter on device bond0 Apr 20 10:51:10 vm-host-colo-2 kernel: device bond0 entered promiscuous mode Apr 20 10:51:10 vm-host-colo-2 kernel: device em1 entered promiscuous mode Apr 20 10:51:10 vm-host-colo-2 kernel: device em2 entered promiscuous mode Apr 20 10:51:10 vm-host-colo-2 kernel: device em3 entered promiscuous mode Apr 20 10:51:10 vm-host-colo-2 kernel: device em4 entered promiscuous mode Apr 20 10:51:10 vm-host-colo-2 network: Bringing up interface bond0: [ OK ] Apr 20 10:51:11 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_UP): bond0.10: link is not ready Apr 20 10:51:11 vm-host-colo-2 kernel: device bond0.10 entered promiscuous mode Apr 20 10:51:11 vm-host-colo-2 network: Bringing up interface bond0.10: [ OK ] Apr 20 10:51:11 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_UP): bond0.102: link is not ready Apr 20 10:51:13 vm-host-colo-2 kernel: tg3 0000:01:00.0 em1: Link is up at 1000 Mbps, full duplex Apr 20 10:51:13 vm-host-colo-2 kernel: tg3 0000:01:00.0 em1: Flow control is on for TX and on for RX Apr 20 10:51:13 vm-host-colo-2 kernel: tg3 0000:01:00.0 em1: EEE is disabled Apr 20 10:51:13 vm-host-colo-2 kernel: bond0: link status definitely up for interface em1, 1000 Mbps full duplex Apr 20 10:51:13 vm-host-colo-2 kernel: bond0: Warning: No 802.3ad response from the link partner for any adapters in the bond Apr 20 10:51:13 vm-host-colo-2 kernel: bond0: first active interface up! Apr 20 10:51:13 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): bond0: link becomes ready Apr 20 10:51:13 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): bond0.10: link becomes ready Apr 20 10:51:13 vm-host-colo-2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): bond0.102: link becomes ready Apr 20 10:51:13 vm-host-colo-2 kernel: tg3 0000:01:00.1 em2: Link is up at 1000 Mbps, full duplex Apr 20 10:51:13 vm-host-colo-2 kernel: tg3 0000:01:00.1 em2: Flow control is on for TX and on for RX Apr 20 10:51:13 vm-host-colo-2 kernel: tg3 0000:01:00.1 em2: EEE is disabled Apr 20 10:51:13 vm-host-colo-2 kernel: bond0: link status definitely up for interface em2, 1000 Mbps full duplex Apr 20 10:51:13 vm-host-colo-2 kernel: tg3 0000:02:00.0 em3: Link is up at 1000 Mbps, full duplex Apr 20 10:51:13 vm-host-colo-2 kernel: tg3 0000:02:00.0 em3: Flow control is on for TX and on for RX Apr 20 10:51:13 vm-host-colo-2 kernel: tg3 0000:02:00.0 em3: EEE is disabled Apr 20 10:51:13 vm-host-colo-2 kernel: bond0: link status definitely up for interface em3, 1000 Mbps full duplex Apr 20 10:51:13 vm-host-colo-2 kernel: tg3 0000:02:00.1 em4: Link is up at 1000 Mbps, full duplex Apr 20 10:51:13 vm-host-colo-2 kernel: tg3 0000:02:00.1 em4: Flow control is on for TX and on for RX Apr 20 10:51:13 vm-host-colo-2 kernel: tg3 0000:02:00.1 em4: EEE is disabled Apr 20 10:51:13 vm-host-colo-2 kernel: bond0: link status definitely up for interface em4, 1000 Mbps full duplex Apr 20 10:51:15 vm-host-colo-2 network: Bringing up interface bond0.102: [ OK ] Apr 20 10:51:15 vm-host-colo-2 kernel: device bond0.20 entered promiscuous mode Apr 20 10:51:15 vm-host-colo-2 network: Bringing up interface bond0.20: [ OK ] Apr 20 10:51:15 vm-host-colo-2 kernel: device bond0.30 entered promiscuous mode Apr 20 10:51:15 vm-host-colo-2 network: Bringing up interface bond0.30: [ OK ] Apr 20 10:51:15 vm-host-colo-2 kernel: device bond0.40 entered promiscuous mode Apr 20 10:51:15 vm-host-colo-2 network: Bringing up interface bond0.40: [ OK ] Apr 20 10:51:15 vm-host-colo-2 kernel: device bond0.50 entered promiscuous mode Apr 20 10:51:15 vm-host-colo-2 network: Bringing up interface bond0.50: [ OK ] Apr 20 10:51:15 vm-host-colo-2 kernel: DMZ: port 1(bond0.10) entered forwarding state Apr 20 10:51:15 vm-host-colo-2 kernel: DMZ: port 1(bond0.10) entered forwarding state Apr 20 10:51:16 vm-host-colo-2 network: Bringing up interface DMZ: [ OK ] Apr 20 10:51:16 vm-host-colo-2 kernel: Internal-Dev: port 1(bond0.30) entered forwarding state Apr 20 10:51:16 vm-host-colo-2 kernel: Internal-Dev: port 1(bond0.30) entered forwarding state Apr 20 10:51:16 vm-host-colo-2 network: Bringing up interface Internal-Dev: [ OK ] Apr 20 10:51:16 vm-host-colo-2 kernel: Lab: port 1(bond0.40) entered forwarding state Apr 20 10:51:16 vm-host-colo-2 kernel: Lab: port 1(bond0.40) entered forwarding state Apr 20 10:51:16 vm-host-colo-2 network: Bringing up interface Lab: [ OK ] Apr 20 10:51:16 vm-host-colo-2 kernel: Server-Net: port 1(bond0.20) entered forwarding state Apr 20 10:51:16 vm-host-colo-2 kernel: Server-Net: port 1(bond0.20) entered forwarding state Apr 20 10:51:16 vm-host-colo-2 network: Bringing up interface Server-Net: [ OK ] Apr 20 10:51:16 vm-host-colo-2 kernel: Workstation: port 1(bond0.50) entered forwarding state Apr 20 10:51:16 vm-host-colo-2 kernel: Workstation: port 1(bond0.50) entered forwarding state Apr 20 10:51:16 vm-host-colo-2 network: Bringing up interface Workstation: [ OK ] Apr 20 10:51:17 vm-host-colo-2 kernel: ovirtmgmt: port 1(bond0) entered forwarding state Apr 20 10:51:17 vm-host-colo-2 kernel: ovirtmgmt: port 1(bond0) entered forwarding state Apr 20 10:51:21 vm-host-colo-2 network: Bringing up interface ovirtmgmt: [ OK ] Apr 20 10:51:21 vm-host-colo-2 systemd: Started LSB: Bring up/down networking.
-- ifcfg-bond0 # Generated by VDSM version 4.19.10.1-1.el7.centos DEVICE=bond0 BONDING_OPTS='mode=4 miimon=100 xmit_hash_policy=2' BRIDGE=ovirtmgmt ONBOOT=yes MTU=1500 DEFROUTE=no NM_CONTROLLED=no IPV6INIT=no -- ifcfg-em1 # Generated by VDSM version 4.19.10.1-1.el7.centos DEVICE=em1 MASTER=bond0 SLAVE=yes ONBOOT=yes BOOTPROTO=none MTU=1500 DEFROUTE=no NM_CONTROLLED=no IPV6INIT=no --ifcfg-em2 # Generated by VDSM version 4.19.10.1-1.el7.centos DEVICE=em2 MASTER=bond0 SLAVE=yes ONBOOT=yes BOOTPROTO=none MTU=1500 DEFROUTE=no NM_CONTROLLED=no IPV6INIT=no --ifcfg-em3 # Generated by VDSM version 4.19.10.1-1.el7.centos DEVICE=em3 MASTER=bond0 SLAVE=yes ONBOOT=yes BOOTPROTO=none MTU=1500 DEFROUTE=no NM_CONTROLLED=no IPV6INIT=no --ifcfg-em4 # Generated by VDSM version 4.19.10.1-1.el7.centos DEVICE=em4 MASTER=bond0 SLAVE=yes ONBOOT=yes BOOTPROTO=none MTU=1500 DEFROUTE=no NM_CONTROLLED=no IPV6INIT=no
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users