Hi, Thanks Andrew for giving me advise for upgrading heartbeat 2.1.3 to pacemaker 1.0.4 anlong with heartbeat 2.99.2.
After Installing and configuring I run couple of tests, like turning off the nodes, starting and rebooting them. Working fine what I desired that moment I've assigned private IPs on the nodes i.e. node, ha1 IP: 192.168.0.184 and node ha2 IP:192.168.0.185. But when I assigned ha1 and ha2 as a public IP, I'm still facing probelm like previous version of heartbeat 2.1.3. When I turn off ha2 and power it up it shows following status; ============ Last updated: Thu Jul 30 17:46:12 2009 Stack: Heartbeat Current DC: ha2 (70503c2e-bb4a-48f8-aab3-53696656a4d0) - partition with quorum Version: 1.0.4-6dede86d6105786af3a5321ccf66b44b6914f0aa 2 Nodes configured, unknown expected votes 4 Resources configured. ============ OFFLINE: [ ha1 ] Online: [ ha2 ] IPaddr_1 (ocf::heartbeat:IPaddr): Started ha2 IPaddr_2 (ocf::heartbeat:IPaddr): Started ha2 OpenSips_1 (ocf::heartbeat:OpenSips): Started ha2 OpenSips_2 (ocf::heartbeat:OpenSips): Started ha2 On the other hand when I check the node ha1 which was failover to ha2, shows its status as listed below; ============ Last updated: Thu Jul 30 17:46:12 2009 Stack: Heartbeat Current DC: ha1 (e651c120-b9a1-489a-baf7-caf0028ad540) - partition with quorum Version: 1.0.4-6dede86d6105786af3a5321ccf66b44b6914f0aa 2 Nodes configured, unknown expected votes 4 Resources configured. ============ OFFLINE: [ ha2 ] Online: [ ha1 ] IPaddr_1 (ocf::heartbeat:IPaddr): Started ha1 IPaddr_2 (ocf::heartbeat:IPaddr): Started ha1 OpenSips_1 (ocf::heartbeat:OpenSips): Started ha1 OpenSips_2 (ocf::heartbeat:OpenSips): Started ha1 When I checked the logs it shows me that ha2 is not member of ha1, as I'm listing down the logs below; Jul 30 18:10:34 ha1 crmd: [2841]: WARN: crmd_ha_msg_callback: Ignoring HA message (op=vote) from ha2: not in our membership list (size=1) Jul 30 18:10:37 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_slave_all message (38) from ha2: not in our membership Jul 30 18:10:37 ha1 attrd: [2840]: info: attrd_ha_callback: flush message from ha2 Jul 30 18:10:39 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_replace message (3b) from ha2: not in our membership Jul 30 18:10:39 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (3f) from ha2: not in our membership Jul 30 18:10:41 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (41) from ha2: not in our membership Jul 30 18:10:41 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (43) from ha2: not in our membership Jul 30 18:10:41 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (45) from ha2: not in our membership Jul 30 18:10:41 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (46) from ha2: not in our membership Jul 30 18:10:41 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (47) from ha2: not in our membership Jul 30 18:10:41 ha1 attrd: [2840]: info: attrd_ha_callback: flush message from ha2 Jul 30 18:10:42 ha1 last message repeated 3 times Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (4a) from ha2: not in our membership Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (4b) from ha2: not in our membership Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (4c) from ha2: not in our membership Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (4d) from ha2: not in our membership Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (4e) from ha2: not in our membership Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (4f) from ha2: not in our membership Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (50) from ha2: not in our membership Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (52) from ha2: not in our membership Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding cib_apply_diff message (53) from ha2: not in our membership Jul 30 18:10:55 ha1 ccm: [2836]: info: Break tie for 2 nodes cluster Jul 30 18:10:55 ha1 crmd: [2841]: info: mem_handle_event: Got an event OC_EV_MS_INVALID from ccm Jul 30 18:10:55 ha1 cib: [2837]: info: mem_handle_event: Got an event OC_EV_MS_INVALID from ccm Jul 30 18:10:55 ha1 crmd: [2841]: info: mem_handle_event: no mbr_track info Jul 30 18:10:55 ha1 cib: [2837]: info: mem_handle_event: no mbr_track info Jul 30 18:10:55 ha1 crmd: [2841]: info: mem_handle_event: Got an event OC_EV_MS_NEW_MEMBERSHIP from ccm Jul 30 18:10:55 ha1 cib: [2837]: info: mem_handle_event: Got an event OC_EV_MS_NEW_MEMBERSHIP from ccm Jul 30 18:10:55 ha1 crmd: [2841]: info: mem_handle_event: instance=31, nodes=1, new=0, lost=0, n_idx=0, new_idx=1, old_idx=3 Jul 30 18:10:55 ha1 cib: [2837]: info: mem_handle_event: instance=31, nodes=1, new=0, lost=0, n_idx=0, new_idx=1, old_idx=3 Jul 30 18:10:55 ha1 crmd: [2841]: info: crmd_ccm_msg_callback: Quorum (re)attained after event=NEW MEMBERSHIP (id=31) Jul 30 18:10:55 ha1 cib: [2837]: info: cib_ccm_msg_callback: Processing CCM event=NEW MEMBERSHIP (id=31) Jul 30 18:10:55 ha1 crmd: [2841]: info: ccm_event_detail: NEW MEMBERSHIP: trans=31, nodes=1, new=0, lost=0 n_idx=0, new_idx=1, old_idx=3 Jul 30 18:10:55 ha1 crmd: [2841]: info: ccm_event_detail: CURRENT: ha1 [nodeid=0, born=31] Jul 30 18:10:55 ha1 crmd: [2841]: info: populate_cib_nodes_ha: Requesting the list of configured nodes Jul 30 18:10:57 ha1 cib: [2837]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/182, version=0.37.66): ok (rc=0) Jul 30 18:11:00 ha1 crmd: [2841]: WARN: crmd_ha_msg_callback: Ignoring HA message (op=noop) from ha2: not in our membership list (size=1) Kindly review my problem. I'm attaching my ha.cf and cib.xml along with and please do reply. -- Regards, Ahmed Munir
<cib validate-with="pacemaker-1.0" crm_feature_set="3.0.1" have-quorum="1" admin_epoch="0" epoch="42" num_updates="0" cib-last-written="Thu Jul 30 18:25:59 2009" dc-uuid="e651c120-b9a1-489a-baf7-caf0028ad540">
<configuration>
<crm_config>
<cluster_property_set id="cib-bootstrap-options">
<nvpair id="cib-bootstrap-options-symmetric-cluster" name="symmetric-cluster" value="true"/>
<nvpair id="cib-bootstrap-options-no-quorum-policy" name="no-quorum-policy" value="stop"/>
<nvpair id="cib-bootstrap-options-default-resource-stickiness" name="default-resource-stickiness" value="0"/>
<nvpair id="cib-bootstrap-options-default-resource-failure-stickiness" name="default-resource-failure-stickiness" value="0"/>
<nvpair id="cib-bootstrap-options-stonith-enabled" name="stonith-enabled" value="false"/>
<nvpair id="cib-bootstrap-options-stonith-action" name="stonith-action" value="reboot"/>
<nvpair id="cib-bootstrap-options-startup-fencing" name="startup-fencing" value="true"/>
<nvpair id="cib-bootstrap-options-stop-orphan-resources" name="stop-orphan-resources" value="true"/>
<nvpair id="cib-bootstrap-options-stop-orphan-actions" name="stop-orphan-actions" value="true"/>
<nvpair id="cib-bootstrap-options-remove-after-stop" name="remove-after-stop" value="false"/>
<nvpair id="cib-bootstrap-options-short-resource-names" name="short-resource-names" value="true"/>
<nvpair id="cib-bootstrap-options-transition-idle-timeout" name="transition-idle-timeout" value="5min"/>
<nvpair id="cib-bootstrap-options-default-action-timeout" name="default-action-timeout" value="20s"/>
<nvpair id="cib-bootstrap-options-is-managed-default" name="is-managed-default" value="true"/>
<nvpair id="cib-bootstrap-options-cluster-delay" name="cluster-delay" value="60s"/>
<nvpair id="cib-bootstrap-options-pe-error-series-max" name="pe-error-series-max" value="-1"/>
<nvpair id="cib-bootstrap-options-pe-warn-series-max" name="pe-warn-series-max" value="-1"/>
<nvpair id="cib-bootstrap-options-pe-input-series-max" name="pe-input-series-max" value="-1"/>
<nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="1.0.4-6dede86d6105786af3a5321ccf66b44b6914f0aa"/>
<nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="Heartbeat"/>
</cluster_property_set>
</crm_config>
<nodes>
<node id="70503c2e-bb4a-48f8-aab3-53696656a4d0" uname="ha2" type="normal"/>
<node id="e651c120-b9a1-489a-baf7-caf0028ad540" uname="ha1" type="normal"/>
</nodes>
<resources>
<primitive class="ocf" provider="heartbeat" type="IPaddr" id="IPaddr_1">
<operations>
<op id="IPaddr_1_mon" interval="10s" name="monitor" timeout="8s"/>
</operations>
<instance_attributes id="IPaddr_1_inst_attr">
<nvpair name="ip" value="(Public IP 1)" id="IPaddr_1_machine_1"/>
<nvpair id="IP1_attr_1" name="netmask" value="27"/>
<nvpair id="IP1_attr_2" name="nic" value="eth0:0"/>
</instance_attributes>
</primitive>
<primitive class="ocf" provider="heartbeat" type="IPaddr" id="IPaddr_2">
<operations>
<op id="IPaddr_2_mon" interval="10s" name="monitor" timeout="8s"/>
</operations>
<instance_attributes id="IPaddr_2_inst_attr">
<nvpair name="ip" value="(Public IP 2)" id="IPaddr_2_machine_2"/>
<nvpair id="IP2_attr_1" name="netmask" value="27"/>
<nvpair id="IP2_attr_2" name="nic" value="eth0:0"/>
</instance_attributes>
</primitive>
<primitive class="ocf" provider="heartbeat" id="OpenSips_1" type="OpenSips">
<operations>
<op id="opensips-start-ha1" name="start" interval="0" timeout="15s"/>
<op id="opensips-stop-ha1" name="stop" interval="0" timeout="10s"/>
<op name="monitor" timeout="10s" id="opensips-monitor-ha1" interval="12s">
<instance_attributes id="ha1-monitor_10s">
<nvpair name="ip" id="ha1-opensips-monitor-ip" value="(Public IP 1)"/>
</instance_attributes>
</op>
</operations>
</primitive>
<primitive class="ocf" provider="heartbeat" id="OpenSips_2" type="OpenSips">
<operations>
<op id="opensips-start-ha2" name="start" interval="0" timeout="15s"/>
<op id="opensips-stop-ha2" name="stop" interval="0" timeout="10s"/>
<op name="monitor" timeout="10s" id="opensips-monitor-ha2" interval="12s">
<instance_attributes id="ha2-monitor_10s">
<nvpair name="ip" id="ha2-opensips-monitor-ip" value="(Public IP 2)"/>
</instance_attributes>
</op>
</operations>
</primitive>
</resources>
<constraints>
<rsc_location id="OpenSips_resource_location_1" rsc="OpenSips_1">
<rule id="rule_ha1" score="100">
<expression id="expression_uname_eq_ha1" attribute="#uname" operation="eq" value="ha1"/>
</rule>
</rsc_location>
<rsc_location id="OpenSips_resource_location_2" rsc="OpenSips_2">
<rule id="rule_ha2" score="100">
<expression id="expression_uname_eq_ha2" attribute="#uname" operation="eq" value="ha2"/>
</rule>
</rsc_location>
<rsc_location id="rsc_location_IPaddr_1" rsc="IPaddr_1">
<rule id="prefered_location_IPaddr_1" score="200">
<expression attribute="#uname" id="prefered_location_IPaddr_1_expr" operation="eq" value="ha1"/>
</rule>
</rsc_location>
<rsc_location id="rsc_location_IPaddr_2" rsc="IPaddr_2">
<rule id="prefered_location_IPaddr_2" score="200">
<expression attribute="#uname" id="prefered_location_IPaddr_2_expr" operation="eq" value="ha2"/>
</rule>
</rsc_location>
<rsc_location id="my1_resource1_connected" rsc="IPaddr_1">
<rule id="my1_resource1_connected_rule" score-attribute="pingd">
<expression id="my1_resource1_connected_expr_defined" attribute="pingd" operation="defined"/>
</rule>
</rsc_location>
<rsc_location id="my2_resource2_connected" rsc="IPaddr_2">
<rule id="my2_resource2_connected_rule" score-attribute="pingd">
<expression id="my2_resource2_connected_expr_defined" attribute="pingd" operation="defined"/>
</rule>
</rsc_location>
</constraints>
</configuration>
</cib>
ha.cf
Description: Binary data
_______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
