Hi
I am using a two node pacemaker cluster with teaming enabled. The cluster has
1. Two team interfaces with different subents.
2. The team1 has a NFS VIP plumbed to it.
3. The VirtualIP from pacemaker is configured to plumb to team0(Corosync
ring number is 0)
In this case the corosync takes the NFS IP as its ring address and checks the
same in the corosync.conf. Since conf file has team0 hostname the corosync
start fails.
Outputs:
$ip a output:
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN qlen 1
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master team0 state
UP qlen 1000
link/ether 38:63:bb:3f:a4:ac brd ff:ff:ff:ff:ff:ff
3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master team1 state
UP qlen 1000
link/ether 38:63:bb:3f:a4:ad brd ff:ff:ff:ff:ff:ff
4: eth2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen 1000
link/ether 38:63:bb:3f:a4:ae brd ff:ff:ff:ff:ff:ff
5: eth3: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq state DOWN qlen
1000
link/ether 38:63:bb:3f:a4:af brd ff:ff:ff:ff:ff:ff
6: eth4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master team0 state
UP qlen 1000
link/ether 36:f7:05:1f:b3:b1 brd ff:ff:ff:ff:ff:ff
7: eth5: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master team1 state
UP qlen 1000
link/ether 38:63:bb:3f:a4:ad brd ff:ff:ff:ff:ff:ff
8: eth6: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen 1000
link/ether 28:80:23:a7:dd:fe brd ff:ff:ff:ff:ff:ff
9: eth7: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq state DOWN qlen
1000
link/ether 28:80:23:a7:dd:ff brd ff:ff:ff:ff:ff:ff
10: team1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP
qlen 1000
link/ether 38:63:bb:3f:a4:ad brd ff:ff:ff:ff:ff:ff
inet 10.64.23.117/28 brd 10.64.23.127 scope global team1
valid_lft forever preferred_lft forever
inet 10.64.23.121/24 scope global secondary team1:~m0
valid_lft forever preferred_lft forever
inet6 fe80::3a63:bbff:fe3f:a4ad/64 scope link
valid_lft forever preferred_lft forever
11: team0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP
qlen 1000
link/ether 38:63:bb:3f:a4:ac brd ff:ff:ff:ff:ff:ff
inet 10.64.23.103/28 brd 10.64.23.111 scope global team0
valid_lft forever preferred_lft forever
inet6 fe80::3a63:bbff:fe3f:a4ac/64 scope link
valid_lft forever preferred_lft forever
Corosync Conf File:
cat /etc/corosync/corosync.conf
totem {
version: 2
secauth: off
cluster_name: DES
transport: udp
rrp_mode: passive
interface {
ringnumber: 0
bindnetaddr: 10.64.23.96
mcastaddr: 224.1.1.1
mcastport: 6860
}
}
nodelist {
node {
ring0_addr: dl380x4415
nodeid: 1
}
node {
ring0_addr: dl360x4405
nodeid: 2
}
}
quorum {
provider: corosync_votequorum
two_node: 1
}
logging {
to_logfile: yes
logfile: /var/log/cluster/corosync.log
to_syslog: yes
}
/etc/hosts:
$ cat /etc/hosts
127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4
::1 localhost localhost.localdomain localhost6 localhost6.localdomain6
10.64.23.101 dl380x4416
10.64.23.104 dl380x4389
10.64.23.106 dl360x4387
10.64.23.103 dl380x4415
10.64.23.105 dl360x4405
10.64.23.115 dl380x4416-int
10.64.23.117 dl380x4415-int
10.64.23.119 dl360x4405-int
10.64.23.120 dl360x4387-int
10.64.23.118 dl380x4389-int
10.64.23.102 dl380x4414
Logs:
[3029] dl380x4415 corosyncerror [MAIN ] Corosync Cluster Engine exiting with
status 20 at service.c:356.
[19040] dl380x4415 corosyncnotice [MAIN ] Corosync Cluster Engine ('2.4.0'):
started and ready to provide service.
[19040] dl380x4415 corosyncinfo [MAIN ] Corosync built-in features: dbus
systemd xmlconf qdevices qnetd snmp pie relro bindnow
[19040] dl380x4415 corosyncnotice [TOTEM ] Initializing transport (UDP/IP
Multicast).
[19040] dl380x4415 corosyncnotice [TOTEM ] Initializing transmit/receive
security (NSS) crypto: none hash: none
[19040] dl380x4415 corosyncnotice [TOTEM ] The network interface
[10.64.23.121] is now up.
[19040] dl380x4415 corosyncnotice [SERV ] Service engine loaded: corosync
configuration map access [0]
[19040] dl380x4415 corosyncinfo [QB ] server name: cmap
[19040] dl380x4415 corosyncnotice [SERV ] Service engine loaded: corosync
configuration service [1]
[19040] dl380x4415 corosyncinfo [QB ] server name: cfg
[19040] dl380x4415 corosyncnotice [SERV ] Service engine loaded: corosync
cluster closed process group service v1.01 [2]
[19040] dl380x4415 corosyncinfo [QB ] server name: cpg
[19040] dl380x4415 corosyncnotice [SERV ] Service engine loaded: corosync
profile loading service [4]
[19040] dl380x4415 corosyncnotice [QUORUM] Using quorum provider
corosync_votequorum
[19040] dl380x4415 corosynccrit [QUORUM] Quorum provider:
corosync_votequorum failed to initialize.
[19040] dl380x4415 corosyncerror [SERV ] Service engine 'corosync_quorum'
failed to load for reason 'configuration error: nodelist or
quorum.expected_votes must be configured!'
With Regards
Somanath Thilak J
_______________________________________________
Users mailing list: [email protected]
http://lists.clusterlabs.org/mailman/listinfo/users
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org