Hi everyone,
I am experiencing strange problems on a two node cluster. The nodes are virtual
machines on a XEN 5.6 SP2 host. I have set up a simple cluster with no
resources defined. After I setup the configuration and commit it to the cluster
the message [TOTEM ] FAILED TO RECEIVE comes up and shortly after that the AIS
connection is lost and I cannot access the crm anymore.
Here is the output:
Sep 21 13:41:04 node1 cib: [11499]: info: cib_replace_notify: Replaced: 0.5.23
-> 0.6.1 from <null>
Sep 21 13:41:04 node1 crmd: [11503]: info: abort_transition_graph:
te_update_diff:124 - Triggered transition abort (complete=1, tag=diff,
id=(null), magic=NA, cib=0.6.1) : Non-status change
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff- <cib admin_epoch="0"
epoch="5" num_updates="23" />
Sep 21 13:41:04 node1 crmd: [11503]: info: do_state_transition: State
transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL
origin=abort_transition_graph ]
Sep 21 13:41:04 node1 attrd: [11501]: info: do_cib_replaced: Sending full
refresh
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ <cib epoch="6"
num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2"
crm_feature_set="3.0.5" have-quorum="1" dc-uuid="node1" >
Sep 21 13:41:04 node1 crmd: [11503]: info: do_state_transition: All 2 cluster
nodes are eligible to run resources.
Sep 21 13:41:04 node1 attrd: [11501]: info: attrd_trigger_update: Sending flush
op to all hosts for: probe_complete (true)
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ <configuration >
Sep 21 13:41:04 node1 crmd: [11503]: info: do_pe_invoke: Query 70: Requesting
the current CIB: S_POLICY_ENGINE
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ <crm_config >
Sep 21 13:41:04 node1 crmd: [11503]: info: do_state_transition: State
transition S_POLICY_ENGINE -> S_ELECTION [ input=I_ELECTION
cause=C_FSA_INTERNAL origin=do_cib_replaced ]
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ <cluster_property_set
id="cib-bootstrap-options" >
Sep 21 13:41:04 node1 crmd: [11503]: info: update_dc: Unset DC node1
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ <nvpair
id="cib-bootstrap-options-stonith-enabled" name="stonith-enabled" value="false"
__crm_diff_marker__="added:top" />
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ <nvpair
id="cib-bootstrap-options-no-quorum-policy" name="no-quorum-policy"
value="ignore" __crm_diff_marker__="added:top" />
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+
</cluster_property_set>
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ </crm_config>
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ <rsc_defaults
__crm_diff_marker__="added:top" >
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ <meta_attributes
id="rsc-options" >
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ <nvpair
id="rsc-options-resource-stickiness" name="resource-stickiness" value="100" />
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ </meta_attributes>
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ </rsc_defaults>
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ </configuration>
Sep 21 13:41:04 node1 cib: [11499]: info: cib:diff+ </cib>
Sep 21 13:41:04 node1 cib: [11499]: info: cib_process_request: Operation
complete: op cib_replace for section 'all' (origin=local/cibadmin/2,
version=0.6.1): ok (rc=0)
Sep 21 13:41:04 node1 cib: [11499]: info: cib_process_request: Operation
complete: op cib_modify for section nodes (origin=local/crmd/68,
version=0.6.2): ok (rc=0)
Sep 21 13:41:04 node1 crmd: [11503]: info: config_query_callback: Shutdown
escalation occurs after: 1200000ms
Sep 21 13:41:04 node1 crmd: [11503]: info: config_query_callback: Checking for
expired actions every 900000ms
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c
Sep 21 13:41:04 node1 crmd: [11503]: info: config_query_callback: Sending
expected-votes=2 to corosync
Sep 21 13:41:04 node1 crmd: [11503]: info: ais_dispatch_message: Membership
208: quorum retained
Sep 21 13:41:04 node1 crmd: [11503]: info: crmd_ais_dispatch: Setting expected
votes to 2
Sep 21 13:41:04 node1 cib: [11499]: info: cib_process_request: Operation
complete: op cib_modify for section crm_config (origin=local/crmd/73,
version=0.6.5): ok (rc=0)
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:04 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:05 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:05 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:05 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:06 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:06 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:07 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:07 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:07 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:08 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:08 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:09 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:09 node1 corosync[11485]: [TOTEM ] Retransmit List: 7a 7b 7c 7d
7e 7f 80 81 82 83 84
Sep 21 13:41:17 node1 corosync[11485]: [TOTEM ] FAILED TO RECEIVE
Sep 21 13:41:26 node1 stonith-ng: [11498]: ERROR: ais_dispatch: Receiving
message body failed: (2) Library error: Resource temporarily unavailable (11)
Sep 21 13:41:26 node1 crmd: [11503]: ERROR: ais_dispatch: Receiving message
body failed: (2) Library error: Resource temporarily unavailable (11)
Sep 21 13:41:26 node1 attrd: [11501]: ERROR: ais_dispatch: Receiving message
body failed: (2) Library error: Resource temporarily unavailable (11)
Sep 21 13:41:26 node1 cib: [11499]: ERROR: ais_dispatch: Receiving message body
failed: (2) Library error: Resource temporarily unavailable (11)
Sep 21 13:41:26 node1 stonith-ng: [11498]: ERROR: ais_dispatch: AIS connection
failed
Sep 21 13:41:26 node1 crmd: [11503]: ERROR: ais_dispatch: AIS connection failed
Sep 21 13:41:26 node1 attrd: [11501]: ERROR: ais_dispatch: AIS connection failed
Sep 21 13:41:26 node1 cib: [11499]: ERROR: ais_dispatch: AIS connection failed
Sep 21 13:41:26 node1 stonith-ng: [11498]: ERROR: stonith_peer_ais_destroy: AIS
connection terminated
Sep 21 13:41:26 node1 crmd: [11503]: info: crmd_ais_destroy: connection closed
Sep 21 13:41:26 node1 attrd: [11501]: CRIT: attrd_ais_destroy: Lost connection
to OpenAIS service!
Sep 21 13:41:26 node1 cib: [11499]: ERROR: cib_ais_destroy: AIS connection
terminated
Sep 21 13:41:26 node1 crmd: [11503]: info: cib_native_msgready: Lost connection
to the CIB service [11499].
Sep 21 13:41:26 node1 attrd: [11501]: notice: main: Exiting...
Sep 21 13:41:26 node1 crmd: [11503]: CRIT: cib_native_dispatch: Lost connection
to the CIB service [11499/callback].
Sep 21 13:41:26 node1 attrd: [11501]: ERROR: attrd_cib_connection_destroy:
Connection to the CIB terminated...
Sep 21 13:41:26 node1 crmd: [11503]: CRIT: cib_native_dispatch: Lost connection
to the CIB service [11499/command].
Sep 21 13:41:26 node1 crmd: [11503]: ERROR: crmd_cib_connection_destroy:
Connection to the CIB terminated...
Sep 21 13:41:26 node1 crmd: [11503]: info: stonith_msgready: Lost connection to
the STONITH service [11498].
Sep 21 13:41:26 node1 crmd: [11503]: CRIT: stonith_dispatch: Lost connection to
the STONITH service [11498/callback].
Sep 21 13:41:26 node1 crmd: [11503]: CRIT: stonith_dispatch: Lost connection to
the STONITH service [11498/command].
Sep 21 13:41:26 node1 crmd: [11503]: CRIT: tengine_stonith_connection_destroy:
Fencing daemon connection failed
Sep 21 13:41:26 node1 crmd: [11503]: ERROR: do_log: FSA: Input I_ERROR from
crmd_cib_connection_destroy() received in state S_ELECTION
Sep 21 13:41:26 node1 crmd: [11503]: info: do_state_transition: State
transition S_ELECTION -> S_RECOVERY [ input=I_ERROR cause=C_FSA_INTERNAL
origin=crmd_cib_connection_destroy ]
Sep 21 13:41:26 node1 crmd: [11503]: ERROR: do_recover: Action A_RECOVER
(0000000001000000) not supported
Sep 21 13:41:26 node1 crmd: [11503]: info: do_dc_release: DC role released
Sep 21 13:41:26 node1 crmd: [11503]: info: pe_connection_destroy: Connection to
the Policy Engine released
Sep 21 13:41:26 node1 crmd: [11503]: info: do_te_control: Transitioner is now
inactive
Sep 21 13:41:26 node1 crmd: [11503]: ERROR: do_log: FSA: Input I_TERMINATE from
do_recover() received in state S_RECOVERY
Sep 21 13:41:26 node1 crmd: [11503]: info: do_state_transition: State
transition S_RECOVERY -> S_TERMINATE [ input=I_TERMINATE cause=C_FSA_INTERNAL
origin=do_recover ]
Sep 21 13:41:26 node1 crmd: [11503]: info: do_shutdown: Disconnecting STONITH...
Sep 21 13:41:26 node1 crmd: [11503]: info: do_lrm_control: Disconnected from
the LRM
Sep 21 13:41:26 node1 crmd: [11503]: info: do_ha_control: Disconnected from
OpenAIS
Sep 21 13:41:26 node1 crmd: [11503]: info: do_cib_control: Disconnecting CIB
Sep 21 13:41:26 node1 crmd: [11503]: info: do_exit: Performing A_EXIT_0 -
gracefully exiting the CRMd
Sep 21 13:41:26 node1 crmd: [11503]: ERROR: do_exit: Could not recover from
internal error
Sep 21 13:41:26 node1 crmd: [11503]: info: free_mem: Dropping
I_RELEASE_SUCCESS: [ state=S_TERMINATE cause=C_FSA_INTERNAL
origin=do_dc_release ]
Sep 21 13:41:26 node1 crmd: [11503]: info: free_mem: Dropping I_TERMINATE: [
state=S_TERMINATE cause=C_FSA_INTERNAL origin=do_stop ]
Sep 21 13:41:26 node1 crmd: [11503]: info: crm_xml_cleanup: Cleaning up memory
from libxml2
Sep 21 13:41:26 node1 crmd: [11503]: info: do_exit: [crmd] stopped (2)
This is the configuration, nothing fancy here:
node node1
node node2
property $id="cib-bootstrap-options" \
dc-version="1.1.5-5bd2b9154d7d9f86d7f56fe0a74072a5a6590c60" \
cluster-infrastructure="openais" \
expected-quorum-votes="2" \
stonith-enabled="false" \
no-quorum-policy="ignore"
rsc_defaults $id="rsc-options" \
resource-stickiness="100"
I am running corosync 1.3.1, openais 1.1.4 and pacemaker 1.1.5.
I see this a lot too on a different cluster on this XEN host.
What does "[TOTEM ] FAILED TO RECEIVE" exactly mean and why does the cluster
get so unstable after that?
Thank you,
Sascha
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems