hi Marcus,
El 4/12/18 a las 16:09, Marcus Haarmann escribió:
Hi,
you did not provide details about your configuration.
How is the network card set up ? Bonding ?
Send your /etc/network/interfaces details.
If bonding is active, check if the mode is correct in /proc/net/bonding.
We encountered differences between /etc/network/interfaces setup and resulting
mode.
Also, check your switch configuration, VLAN setup, MTU etc.
Yes, sorry about that. I have double checked the switch and all 3 node
SFP+ port have the same configuration.
/etc/network/interfaces in proxmox1 node:
auto lo
iface lo inet loopback
iface eth0 inet manual
iface eth1 inet manual
iface eth2 inet manual
iface eth3 inet manual
iface eth4 inet manual
iface eth5 inet manual
auto vmbr10
iface vmbr10 inet static
address 192.168.10.201
netmask 255.255.255.0
bridge_ports eth3
bridge_stp off
bridge_fd 0
auto vmbr0
iface vmbr0 inet static
address 192.168.0.201
netmask 255.255.255.0
gateway 192.168.0.100
bridge_ports eth4
bridge_stp off
bridge_fd 0
auto eth4.100
iface eth4.100 inet static
address 10.0.2.1
netmask 255.255.255.0
up ip addr add 10.0.3.1/24 dev eth4.100
Cluster is running on vmbr0 network (192.168.0.0/24)
Cheers
Marcus Haarmann
Von: "Eneko Lacunza" <elacu...@binovo.es>
An: "pve-user" <pve-user@pve.proxmox.com>
Gesendet: Dienstag, 4. Dezember 2018 15:57:10
Betreff: [PVE-User] Multicast problems with Intel X540 - 10Gtek network card?
Hi all,
We have just updated a 3-node Proxmox cluster from 3.4 to 5.2, Ceph
hammer to Luminous and the network from 1 Gbit to 10Gbit... one of the
three Proxmox nodes is new too :)
Generally all was good and VMs are working well. :-)
BUT, we have some problems with the cluster; promxox1 node joins and
then after about 4 minutes drops from the cluster.
All multicast tests
https://pve.proxmox.com/wiki/Multicast_notes#Using_omping_to_test_multicast
run fine except the last one:
*** proxmox1:
root@proxmox1:~# omping -c 600 -i 1 -F -q proxmox1 proxmox3 proxmox4
proxmox3 : waiting for response msg
proxmox4 : waiting for response msg
proxmox3 : joined (S,G) = (*, 232.43.211.234), pinging
proxmox4 : joined (S,G) = (*, 232.43.211.234), pinging
proxmox3 : given amount of query messages was sent
proxmox4 : given amount of query messages was sent
proxmox3 : unicast, xmt/rcv/%loss = 600/600/0%, min/avg/max/std-dev =
0.073/0.184/0.390/0.061
proxmox3 : multicast, xmt/rcv/%loss = 600/262/56%, min/avg/max/std-dev =
0.092/0.207/0.421/0.068
proxmox4 : unicast, xmt/rcv/%loss = 600/600/0%, min/avg/max/std-dev =
0.049/0.167/0.369/0.059
proxmox4 : multicast, xmt/rcv/%loss = 600/262/56%, min/avg/max/std-dev =
0.063/0.185/0.386/0.064
*** proxmox3:
root@proxmox3:/etc# omping -c 600 -i 1 -F -q proxmox1 proxmox3 proxmox4
proxmox1 : waiting for response msg
proxmox4 : waiting for response msg
proxmox4 : joined (S,G) = (*, 232.43.211.234), pinging
proxmox1 : waiting for response msg
proxmox1 : joined (S,G) = (*, 232.43.211.234), pinging
proxmox4 : given amount of query messages was sent
proxmox1 : given amount of query messages was sent
proxmox1 : unicast, xmt/rcv/%loss = 600/600/0%, min/avg/max/std-dev =
0.083/0.193/1.030/0.055
proxmox1 : multicast, xmt/rcv/%loss = 600/600/0%, min/avg/max/std-dev =
0.102/0.209/1.050/0.054
proxmox4 : unicast, xmt/rcv/%loss = 600/600/0%, min/avg/max/std-dev =
0.041/0.108/0.172/0.026
proxmox4 : multicast, xmt/rcv/%loss = 600/600/0%, min/avg/max/std-dev =
0.048/0.123/0.190/0.030
*** root@proxmox4:~# omping -c 600 -i 1 -F -q proxmox1 proxmox3 proxmox4
proxmox1 : waiting for response msg
proxmox3 : waiting for response msg
proxmox1 : waiting for response msg
proxmox3 : waiting for response msg
proxmox3 : joined (S,G) = (*, 232.43.211.234), pinging
proxmox1 : joined (S,G) = (*, 232.43.211.234), pinging
proxmox1 : given amount of query messages was sent
proxmox3 : given amount of query messages was sent
proxmox1 : unicast, xmt/rcv/%loss = 600/600/0%, min/avg/max/std-dev =
0.085/0.188/0.356/0.040
proxmox1 : multicast, xmt/rcv/%loss = 600/600/0%, min/avg/max/std-dev =
0.114/0.208/0.377/0.041
proxmox3 : unicast, xmt/rcv/%loss = 600/600/0%, min/avg/max/std-dev =
0.048/0.117/0.289/0.023
proxmox3 : multicast, xmt/rcv/%loss = 600/600/0%, min/avg/max/std-dev =
0.064/0.134/0.290/0.026
Ok, so it seems we have a network problem on proxmox1 node. Network
cards are as follows:
- proxmox1: Intel X540 (10Gtek)
- proxmox3: Intel X710 (Intel)
- proxmox4: Intel X710 (Intel)
Switch is Dell N1224T-ON.
Does anyone have experience with Intel X540 chip network cards or Linux
ixgbe network driver or 10Gtek manufacturer?
If we change corosync communication to 1 Gbit network cards (broadcom)
connected to an old HPE 1800-24G switch, cluster is stable...
We also have a running cluster with Dell n1224T-ON switch and X710
network cards without issues.
Thanks a lot
Eneko
--
Zuzendari Teknikoa / Director Técnico
Binovo IT Human Project, S.L.
Telf. 943569206
Astigarraga bidea 2, 2º izq. oficina 11; 20180 Oiartzun (Gipuzkoa)
www.binovo.es
_______________________________________________
pve-user mailing list
pve-user@pve.proxmox.com
https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user