On 8 Oct 2014, at 2:09 am, Brian J. Murrell (brian) <br...@interlinx.bc.ca> wrote:
> Given a 2 node pacemaker-1.1.10-14.el6_5.3 cluster with nodes "node5" > and "node6" I saw an "unknown" third node being added to the cluster, > but only on node5: Is either node using dhcp? I would guess node6 got a new IP address (or that corosync decided to bind to a different one) > > Sep 18 22:52:16 node5 corosync[17321]: [pcmk ] notice: pcmk_peer_update: > Transitional membership event on ring 12: memb=2, new=0, lost=0 > Sep 18 22:52:16 node5 corosync[17321]: [pcmk ] info: pcmk_peer_update: > memb: node6 3713011210 > Sep 18 22:52:16 node5 corosync[17321]: [pcmk ] info: pcmk_peer_update: > memb: node5 3729788426 > Sep 18 22:52:16 node5 corosync[17321]: [pcmk ] notice: pcmk_peer_update: > Stable membership event on ring 12: memb=3, new=1, lost=0 > Sep 18 22:52:16 node5 corosync[17321]: [pcmk ] info: update_member: > Creating entry for node 2085752330 born on 12 > Sep 18 22:52:16 node5 corosync[17321]: [pcmk ] info: update_member: Node > 2085752330/unknown is now: member > Sep 18 22:52:16 node5 corosync[17321]: [pcmk ] info: pcmk_peer_update: > NEW: .pending. 2085752330 > Sep 18 22:52:16 node5 corosync[17321]: [pcmk ] info: pcmk_peer_update: > MEMB: node6 3713011210 > Sep 18 22:52:16 node5 corosync[17321]: [pcmk ] info: pcmk_peer_update: > MEMB: node5 3729788426 > Sep 18 22:52:16 node5 corosync[17321]: [pcmk ] info: pcmk_peer_update: > MEMB: .pending. 2085752330 > > Above is where this third node seems to appear. > > Sep 18 22:52:16 node5 corosync[17321]: [pcmk ] info: > send_member_notification: Sending membership update 12 to 2 children > Sep 18 22:52:16 node5 corosync[17321]: [TOTEM ] A processor joined or left > the membership and a new membership was formed. > Sep 18 22:52:16 node5 cib[17371]: notice: crm_update_peer_state: > plugin_handle_membership: Node (null)[2085752330] - state is now member (was > (null)) > Sep 18 22:52:16 node5 crmd[17376]: notice: crm_update_peer_state: > plugin_handle_membership: Node (null)[2085752330] - state is now member (was > (null)) > Sep 18 22:52:16 node5 crmd[17376]: notice: do_state_transition: State > transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL > origin=abort_transition_graph ] > Sep 18 22:52:16 node5 crmd[17376]: error: join_make_offer: No recipient > for welcome message > Sep 18 22:52:16 node5 crmd[17376]: warning: do_state_transition: Only 2 of 3 > cluster nodes are eligible to run resources - continue 0 > Sep 18 22:52:16 node5 attrd[17374]: notice: attrd_local_callback: Sending > full refresh (origin=crmd) > Sep 18 22:52:16 node5 attrd[17374]: notice: attrd_trigger_update: Sending > flush op to all hosts for: probe_complete (true) > Sep 18 22:52:16 node5 stonith-ng[17372]: notice: unpack_config: On loss of > CCM Quorum: Ignore > Sep 18 22:52:16 node5 cib[17371]: notice: cib:diff: Diff: --- 0.31.2 > Sep 18 22:52:16 node5 cib[17371]: notice: cib:diff: Diff: +++ 0.32.1 > 4a679012144955c802557a39707247a2 > Sep 18 22:52:16 node5 cib[17371]: notice: cib:diff: -- <nvpair > value="Stopped" id="res1-meta_attributes-target-role"/> > Sep 18 22:52:16 node5 cib[17371]: notice: cib:diff: ++ <nvpair > name="target-role" id="res1-meta_attributes-target-role" value="Started"/> > Sep 18 22:52:16 node5 pengine[17375]: notice: unpack_config: On loss of CCM > Quorum: Ignore > Sep 18 22:52:16 node5 pengine[17375]: notice: LogActions: Start > res1#011(node5) > Sep 18 22:52:16 node5 crmd[17376]: notice: te_rsc_command: Initiating > action 7: start res1_start_0 on node5 (local) > Sep 18 22:52:16 node5 pengine[17375]: notice: process_pe_message: > Calculated Transition 22: /var/lib/pacemaker/pengine/pe-input-165.bz2 > Sep 18 22:52:16 node5 stonith-ng[17372]: notice: stonith_device_register: > Device 'st-fencing' already existed in device list (1 active devices) > > On node6 at the same time the following was in the log: > > Sep 18 22:52:15 node6 corosync[11178]: [TOTEM ] Incrementing problem > counter for seqid 5 iface 10.128.0.221 to [1 of 10] > Sep 18 22:52:16 node6 corosync[11178]: [TOTEM ] Incrementing problem > counter for seqid 8 iface 10.128.0.221 to [2 of 10] > Sep 18 22:52:17 node6 corosync[11178]: [TOTEM ] Decrementing problem > counter for iface 10.128.0.221 to [1 of 10] > Sep 18 22:52:19 node6 corosync[11178]: [TOTEM ] ring 1 active with no faults > > Any idea what's going on here? > > Cheers, > b. > > > > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org
signature.asc
Description: Message signed with OpenPGP using GPGMail
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org