Hi guys,

Apologies for reposting this query, it inadvertently got added to an existing 
topic!


We have a three node cluster deployed in a customer's network:
- 2 nodes are on the same switch
- 3rd node on the same subnet but there's a router in between.
- IP Multicast is enabled and has been tested using omping as follows..

On each node ran..

omping node01 node02 node3


ON node 3

Node01 :   unicast, xmt/rcv/%loss = 23/23/0%, min/avg/max/std-dev = 
0.128/0.181/0.255/0.025
Node01 : multicast, xmt/rcv/%loss = 23/23/0%, min/avg/max/std-dev = 
0.140/0.187/0.219/0.021
Node02 :   unicast, xmt/rcv/%loss = 8/8/0%, min/avg/max/std-dev = 
0.115/0.150/0.168/0.021
Node02 : multicast, xmt/rcv/%loss = 8/8/0%, min/avg/max/std-dev = 
0.134/0.162/0.177/0.014


On node 2


Node01 :   unicast, xmt/rcv/%loss = 9/9/0%, min/avg/max/std-dev = 
0.168/0.191/0.205/0.014
Node01 : multicast, xmt/rcv/%loss = 9/8/11% (seq>=2 0%), min/avg/max/std-dev = 
0.138/0.179/0.206/0.028
Node03 :   unicast, xmt/rcv/%loss = 9/9/0%, min/avg/max/std-dev = 
0.112/0.149/0.175/0.022
Node03 : multicast, xmt/rcv/%loss = 9/8/11% (seq>=2 0%), min/avg/max/std-dev = 
0.124/0.167/0.178/0.018



On node 1

Node02 :   unicast, xmt/rcv/%loss = 8/8/0%, min/avg/max/std-dev = 
0.154/0.185/0.208/0.019
Node02 : multicast, xmt/rcv/%loss = 8/8/0%, min/avg/max/std-dev = 
0.175/0.198/0.214/0.015
Node03 :   unicast, xmt/rcv/%loss = 23/23/0%, min/avg/max/std-dev = 
0.114/0.160/0.185/0.019
Node03 : multicast, xmt/rcv/%loss = 23/22/4% (seq>=2 0%), min/avg/max/std-dev = 
0.124/0.172/0.197/0.019


- Problem is intermittent but frequent. Occasionally starts fine when started 
from scratch.

We suspect the problem is related to node 3 as we can see lrmd failures as per 
the attached log. We've checked permissions are ok as per 
https://bugs.launchpad.net/ubuntu/+source/cluster-glue/+bug/676391



stonith-ng[1437]:    error: ais_dispatch: AIS connection failed
stonith-ng[1437]:    error: stonith_peer_ais_destroy: AIS connection terminated
corosync[1430]:   [SERV  ] Service engine unloaded: Pacemaker Cluster Manager 
1.1.6
corosync[1430]:   [SERV  ] Service engine unloaded: corosync extended virtual 
synchrony service
corosync[1430]:   [SERV  ] Service engine unloaded: corosync configuration 
service
corosync[1430]:   [SERV  ] Service engine unloaded: corosync cluster closed 
process group service v1.01
corosync[1430]:   [SERV  ] Service engine unloaded: corosync cluster config 
database access v1.01
corosync[1430]:   [SERV  ] Service engine unloaded: corosync profile loading 
service
corosync[1430]:   [SERV  ] Service engine unloaded: corosync cluster quorum 
service v0.1
corosync[1430]:   [MAIN  ] Corosync Cluster Engine exiting with status 0 at 
main.c:1894.

corosync[4931]:   [MAIN  ] Corosync built-in features: nss dbus rdma snmp
corosync[4931]:   [MAIN  ] Successfully read main configuration file 
'/etc/corosync/corosync.conf'.
corosync[4931]:   [TOTEM ] Initializing transport (UDP/IP Multicast).
corosync[4931]:   [TOTEM ] Initializing transmit/receive security: libtomcrypt 
SOBER128/SHA1HMAC (mode 0).
corosync[4931]:   [TOTEM ] The network interface [10.87.79.59] is now up.
corosync[4931]:   [pcmk  ] Logging: Initialized pcmk_startup
corosync[4931]:   [SERV  ] Service engine loaded: Pacemaker Cluster Manager 
1.1.6
corosync[4931]:   [pcmk  ] Logging: Initialized pcmk_startup
corosync[4931]:   [SERV  ] Service engine loaded: Pacemaker Cluster Manager 
1.1.6
corosync[4931]:   [SERV  ] Service engine loaded: corosync extended virtual 
synchrony service
corosync[4931]:   [SERV  ] Service engine loaded: corosync configuration service
orosync[4931]:   [SERV  ] Service engine loaded: corosync cluster closed 
process group service v1.01
corosync[4931]:   [SERV  ] Service engine loaded: corosync cluster config 
database access v1.01
corosync[4931]:   [SERV  ] Service engine loaded: corosync profile loading 
service
corosync[4931]:   [SERV  ] Service engine loaded: corosync cluster quorum 
service v0.1
corosync[4931]:   [MAIN  ] Compatibility mode set to whitetank.  Using V1 and 
V2 of the synchronization engine.
corosync[4931]:   [TOTEM ] A processor joined or left the membership and a new 
membership was formed.
corosync[4931]:   [CPG   ] chosen downlist: sender r(0) ip(10.87.79.59) ; 
members(old:0 left:0)
corosync[4931]:   [MAIN  ] Completed service synchronization, ready to provide 
service.
cib[4937]:     info: crm_log_init_worker: Changed active directory to 
/var/lib/heartbeat/cores/hacluster
cib[4937]:     info: retrieveCib: Reading cluster configuration from: 
/var/lib/heartbeat/crm/cib.xml (digest: /var/lib/heartbeat/crm/cib.xml.sig)
 cib[4937]:     info: validate_with_relaxng: Creating RNG parser context
 stonith-ng[4945]:     info: crm_log_init_worker: Changed active directory to 
/var/lib/heartbeat/cores/root
 stonith-ng[4945]:     info: get_cluster_type: Cluster type is: 'openais'
 stonith-ng[4945]:   notice: crm_cluster_connect: Connecting to cluster 
infrastructure: classic openais (with plugin)
 stonith-ng[4945]:     info: init_ais_connection_classic: Creating connection 
to our Corosync plugin
 cib[4944]:     info: crm_log_init_worker: Changed active directory to 
/var/lib/heartbeat/cores/hacluster
 cib[4944]:     info: retrieveCib: Reading cluster configuration from: 
/var/lib/heartbeat/crm/cib.xml (digest: /var/lib/heartbeat/crm/cib.xml.sig)
 stonith-ng[4945]:     info: init_ais_connection_classic: AIS connection 
established
 stonith-ng[4945]:     info: get_ais_nodeid: Server details: id=1003428268 
uname=w0110Danmtapp03 cname=pcmk
 stonith-ng[4945]:     info: init_ais_connection_once: Connection to 'classic 
openais (with plugin)': established
 stonith-ng[4945]:     info: crm_new_peer: Node node03 now has id: 1003428268
 stonith-ng[4945]:     info: crm_new_peer: Node 1003428268 is now known as 
node03
 cib[4944]:     info: validate_with_relaxng: Creating RNG parser context
 lrmd: [4939]: info: G_main_add_SignalHandler: Added signal handler for signal 
15
 lrmd: [4946]: info: Signal sent to pid=4939, waiting for process to exit
 lrmd: [4939]: info: G_main_add_SignalHandler: Added signal handler for signal 
17
 lrmd: [4939]: info: enabling coredumps
 stonith-ng[4938]:     info: crm_log_init_worker: Changed active directory to 
/var/lib/heartbeat/cores/root
 lrmd: [4939]: info: G_main_add_SignalHandler: Added signal handler for signal 
10
 lrmd: [4939]: info: G_main_add_SignalHandler: Added signal handler for signal 
12
 lrmd: [4939]: info: Started.
 stonith-ng[4938]:     info: get_cluster_type: Cluster type is: 'openais'
 lrmd: [4939]: info: lrmd is shutting down
 stonith-ng[4938]:   notice: crm_cluster_connect: Connecting to cluster 
infrastructure: classic openais (with plugin)
 stonith-ng[4938]:     info: init_ais_connection_classic: Creating connection 
to our Corosync plugin
 attrd[4940]:     info: crm_log_init_worker: Changed active directory to 
/var/lib/heartbeat/cores/hacluster
 pengine[4941]:     info: crm_log_init_worker: Changed active directory to 
/var/lib/heartbeat/cores/hacluster
 attrd[4940]:     info: main: Starting up
 attrd[4940]:     info: get_cluster_type: Cluster type is: 'openais'
 attrd[4940]:   notice: crm_cluster_connect: Connecting to cluster 
infrastructure: classic openais (with plugin)
 attrd[4940]:     info: init_ais_connection_classic: Creating connection to our 
Corosync plugin
 crmd[4942]:     info: crm_log_init_worker: Changed active directory to 
/var/lib/heartbeat/cores/hacluster
 pengine[4941]:     info: main: Starting pengine
 crmd[4942]:   notice: main: CRM Hg Version: 
148fccfd5985c5590cc601123c6c16e966b85d14
 pengine[4948]:     info: crm_log_init_worker: Changed active directory to 
/var/lib/heartbeat/cores/hacluster
 pengine[4948]:  warning: main: Terminating previous PE instance
 attrd[4947]:     info: crm_log_init_worker: Changed active directory to 
/var/lib/heartbeat/cores/hacluster
 pengine[4941]:  warning: process_pe_message: Received quit message, terminating
 attrd[4947]:     info: main: Starting up
 attrd[4947]:     info: get_cluster_type: Cluster type is: 'openais'
 attrd[4947]:   notice: crm_cluster_connect: Connecting to cluster 
infrastructure: classic openais (with plugin)
 attrd[4947]:     info: init_ais_connection_classic: Creating connection to our 
Corosync plugin
 crmd[4949]:     info: crm_log_init_worker: Changed active directory to 
/var/lib/heartbeat/cores/hacluster
 crmd[4949]:   notice: main: CRM Hg Version: 
148fccfd5985c5590cc601123c6c16e966b85d14
 stonith-ng[4938]:     info: init_ais_connection_classic: AIS connection 
established
 stonith-ng[4938]:     info: get_ais_nodeid: Server details: id=1003428268 
uname=node03 cname=pcmk
 stonith-ng[4938]:     info: init_ais_connection_once: Connection to 'classic 
openais (with plugin)': established
 stonith-ng[4938]:     info: crm_new_peer: Node node03 now has id: 1003428268
 stonith-ng[4938]:     info: crm_new_peer: Node 1003428268 is now known as 
node03
 attrd[4940]:     info: init_ais_connection_classic: AIS connection established
 attrd[4940]:     info: get_ais_nodeid: Server details: id=1003428268 
uname=node03 cname=pcmk
 attrd[4940]:     info: init_ais_connection_once: Connection to 'classic 
openais (with plugin)': established
 attrd[4940]:     info: crm_new_peer: Node node03 now has id: 1003428268
 attrd[4940]:     info: crm_new_peer: Node 1003428268 is now known as node03
 attrd[4940]:     info: main: Cluster connection active
 attrd[4940]:     info: main: Accepting attribute updates
 attrd[4940]:   notice: main: Starting mainloop...
 attrd[4947]:     info: init_ais_connection_classic: AIS connection established
 attrd[4947]:     info: get_ais_nodeid: Server details: id=1003428268 
uname=node03 cname=pcmk
 attrd[4947]:     info: init_ais_connection_once: Connection to 'classic 
openais (with plugin)': established
 attrd[4947]:     info: crm_new_peer: Node node03 now has id: 1003428268
 attrd[4947]:     info: crm_new_peer: Node 1003428268 is now known as node03
 attrd[4947]:     info: main: Cluster connection active
 attrd[4947]:     info: main: Accepting attribute updates
 attrd[4947]:   notice: main: Starting mainloop...
 cib[4937]:     info: startCib: CIB Initialization completed successfully
 cib[4937]:     info: get_cluster_type: Cluster type is: 'openais'
 cib[4937]:   notice: crm_cluster_connect: Connecting to cluster 
infrastructure: classic openais (with plugin)
 cib[4937]:     info: init_ais_connection_classic: Creating connection to our 
Corosync plugin
 cib[4944]:     info: startCib: CIB Initialization completed successfully
 cib[4944]:     info: get_cluster_type: Cluster type is: 'openais'
 cib[4944]:   notice: crm_cluster_connect: Connecting to cluster 
infrastructure: classic openais (with plugin)
 cib[4944]:     info: init_ais_connection_classic: Creating connection to our 
Corosync plugin
 cib[4937]:     info: init_ais_connection_classic: AIS connection established
 cib[4937]:     info: get_ais_nodeid: Server details: id=1003428268 
uname=node03 cname=pcmk
 cib[4937]:     info: init_ais_connection_once: Connection to 'classic openais 
(with plugin)': established
 cib[4937]:     info: crm_new_peer: Node node03 now has id: 1003428268
 cib[4937]:     info: crm_new_peer: Node 1003428268 is now known as node03
 cib[4937]:     info: cib_init: Starting cib mainloop
 cib[4937]:     info: ais_dispatch_message: Membership 6892: quorum still lost
 cib[4937]:     info: crm_update_peer: Node node03: id=1003428268 state=member 
(new) addr=r(0) ip(10.87.79.59)  (new) votes=1 (new) born=0 seen=6892 
proc=00000000000000000000000000111312 (new)
 cib[4944]:     info: init_ais_connection_classic: AIS connection established
 cib[4944]:     info: get_ais_nodeid: Server details: id=1003428268 
uname=node03 cname=pcmk
 cib[4944]:     info: init_ais_connection_once: Connection to 'classic openais 
(with plugin)': established
 cib[4944]:     info: crm_new_peer: Node node03 now has id: 1003428268
 cib[4944]:     info: crm_new_peer: Node 1003428268 is now known as node03
 cib[4944]:     info: cib_init: Starting cib mainloop
 stonith-ng[4945]:   notice: setup_cib: Watching for stonith topology changes
 stonith-ng[4945]:     info: main: Starting stonith-ng mainloop
 cib[4937]:     info: ais_dispatch_message: Membership 6896: quorum still lost
 corosync[4931]:   [TOTEM ] A processor joined or left the membership and a new 
membership was formed.
 cib[4937]:     info: crm_new_peer: Node <null> now has id: 969873836
 cib[4937]:     info: crm_update_peer: Node (null): id=969873836 state=member 
(new) addr=r(0) ip(172.25.207.57)  votes=0 born=0 seen=6896 
proc=00000000000000000000000000000000
 cib[4937]:     info: crm_new_peer: Node <null> now has id: 986651052
 cib[4937]:     info: crm_update_peer: Node (null): id=986651052 state=member 
(new) addr=r(0) ip(172.25.207.58)  votes=0 born=0 seen=6896 
proc=00000000000000000000000000000000
 cib[4937]:   notice: ais_dispatch_message: Membership 6896: quorum acquired
 cib[4937]:     info: crm_get_peer: Node 986651052 is now known as node02
 cib[4937]:     info: crm_update_peer: Node node02: id=986651052 state=member 
addr=r(0) ip(172.25.207.58)  votes=1 (new) born=6812 seen=6896 
proc=00000000000000000000000000111312 (new)
 cib[4937]:     info: ais_dispatch_message: Membership 6896: quorum retained
 cib[4937]:     info: crm_get_peer: Node 969873836 is now known as node01
 cib[4937]:     info: crm_update_peer: Node node01: id=969873836 state=member 
addr=r(0) ip(172.25.207.57)  votes=1 (new) born=6848 seen=6896 
proc=00000000000000000000000000111312 (new)
 rsyslogd-2177: imuxsock begins to drop messages from pid 4931 due to 
rate-limiting
 crmd[4942]:     info: do_cib_control: CIB connection established
 crmd[4942]:     info: get_cluster_type: Cluster type is: 'openais'
 crmd[4942]:   notice: crm_cluster_connect: Connecting to cluster 
infrastructure: classic openais (with plugin)
 crmd[4942]:     info: init_ais_connection_classic: Creating connection to our 
Corosync plugin
 cib[4937]:     info: cib_process_diff: Diff 1.249.28 -> 1.249.29 not applied 
to 1.249.0: current "num_updates" is less than required
 cib[4937]:     info: cib_server_process_diff: Requesting re-sync from peer
 crmd[4949]:     info: do_cib_control: CIB connection established
 crmd[4949]:     info: get_cluster_type: Cluster type is: 'openais'
 crmd[4949]:   notice: crm_cluster_connect: Connecting to cluster 
infrastructure: classic openais (with plugin)
 crmd[4949]:     info: init_ais_connection_classic: Creating connection to our 
Corosync plugin
 stonith-ng[4938]:   notice: setup_cib: Watching for stonith topology changes
 stonith-ng[4938]:     info: main: Starting stonith-ng mainloop
 cib[4937]:   notice: cib_server_process_diff: Not applying diff 1.249.29 -> 
1.249.30 (sync in progress)
 crmd[4942]:     info: init_ais_connection_classic: AIS connection established
 crmd[4942]:     info: get_ais_nodeid: Server details: id=1003428268 
uname=node03 cname=pcmk
 crmd[4942]:     info: init_ais_connection_once: Connection to 'classic openais 
(with plugin)': established
 crmd[4942]:     info: crm_new_peer: Node node03 now has id: 1003428268
 crmd[4942]:     info: crm_new_peer: Node 1003428268 is now known as node03
 crmd[4942]:     info: ais_status_callback: status: node03 is now unknown
 crmd[4942]:     info: do_ha_control: Connected to the cluster
 crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 1 (30 max) 
times
 crmd[4949]:     info: init_ais_connection_classic: AIS connection established
 crmd[4949]:     info: get_ais_nodeid: Server details: id=1003428268 
uname=node03 cname=pcmk
 crmd[4949]:     info: init_ais_connection_once: Connection to 'classic openais 
(with plugin)': established
 crmd[4942]:   notice: ais_dispatch_message: Membership 6896: quorum acquired
 crmd[4949]:     info: crm_new_peer: Node node03 now has id: 1003428268
 crmd[4949]:     info: crm_new_peer: Node 1003428268 is now known as node03
 crmd[4942]:     info: crm_new_peer: Node node01 now has id: 969873836
 crmd[4949]:     info: ais_status_callback: status: node03 is now unknown
 crmd[4942]:     info: crm_new_peer: Node 969873836 is now known as node01
 crmd[4949]:     info: do_ha_control: Connected to the cluster
 crmd[4942]:     info: ais_status_callback: status: node01 is now unknown
 crmd[4942]:     info: ais_status_callback: status: node01 is now member (was 
unknown)
 crmd[4942]:     info: crm_update_peer: Node node01: id=969873836 state=member 
(new) addr=r(0) ip(172.25.207.57)  votes=1 born=6848 seen=6896 
proc=00000000000000000000000000111312
 crmd[4942]:     info: crm_new_peer: Node node02 now has id: 986651052
 crmd[4942]:     info: crm_new_peer: Node 986651052 is now known as node02
 crmd[4942]:     info: ais_status_callback: status: node02 is now unknown
 crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 1 (30 max) 
times
 crmd[4942]:     info: ais_status_callback: status: node02 is now member (was 
unknown)
 crmd[4942]:     info: crm_update_peer: Node node02: id=986651052 state=member 
(new) addr=r(0) ip(172.25.207.58)  votes=1 born=6812 seen=6896 
proc=00000000000000000000000000111312
 crmd[4942]:   notice: crmd_peer_update: Status update: Client node03/crmd now 
has status [online] (DC=<null>)
 crmd[4942]:     info: ais_status_callback: status: node03 is now member (was 
unknown)
 crmd[4942]:     info: crm_update_peer: Node node03: id=1003428268 state=member 
(new) addr=r(0) ip(10.87.79.59)  (new) votes=1 (new) born=6896 seen=6896 
proc=00000000000000000000000000111312 (new)
 crmd[4942]:     info: ais_dispatch_message: Membership 6896: quorum retained
 cib[4937]:   notice: cib_server_process_diff: Not applying diff 1.249.30 -> 
1.249.31 (sync in progress)
 crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 2 (30 max) 
times
 crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 3 (30 max) 
times
crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 2 (30 max) 
times
crmd[4949]:   notice: ais_dispatch_message: Membership 6896: quorum acquired
rsyslogd-2177: imuxsock begins to drop messages from pid 4937 due to 
rate-limiting
crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 4 (30 max) 
times
crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 5 (30 max) 
times
pengine[4948]:     info: main: Starting pengine
crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
warning: do_lrm_control: Failed to sign on to the LRM 6 (30 max) times
crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 3 (30 max) 
times
attrd[4940]:     info: cib_connect: Connected to the CIB after 1 signon attempts
attrd[4940]:     info: cib_connect: Sending full refresh
crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 7 (30 max) 
times
attrd[4947]:     info: cib_connect: Connected to the CIB after 1 signon attempts
attrd[4947]:     info: cib_connect: Sending full refresh
crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 4 (30 max) 
times
crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 8 (30 max) 
times
crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 5 (30 max) 
times
crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 9 (30 max) 
times
crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 6 (30 max) 
times
crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 10 (30 max) 
times
crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 7 (30 max) 
times
crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 11 (30 max) 
times
crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 8 (30 max) 
times
crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 12 (30 max) 
times
crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 9 (30 max) 
times
crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 13 (30 max) 
times
crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 10 (30 max) 
times
crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 14 (30 max) 
times
 crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 11 (30 max) 
times
crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
 crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 12 (30 max) 
times
 crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
(2000ms)
 crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 15 (30 max) 
times
crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 13 (30 max) 
times
crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 16 (30 max) 
times
 crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
(2000ms)
 crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 14 (30 max) 
times
 crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
(2000ms)
 crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 17 (30 max) 
times
 crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
(2000ms)
 crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 15 (30 max) 
times
 crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
(2000ms)
 crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 18 (30 max) 
times


We have the following components installed..


corosynclib-1.4.1-15.el6.x86_64
corosync-1.4.1-15.el6.x86_64
cluster-glue-libs-1.0.5-6.el6.x86_64
clusterlib-3.0.12.1-49.el6.x86_64
pacemaker-cluster-libs-1.1.7-6.el6.x86_64
cluster-glue-1.0.5-6.el6.x86_64
resource-agents-3.9.2-12.el6.x86_64



We'd appreciate assistance on how to debug what the issue may be and some 
possible causes.

Cheers,
Jimmy
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to