Hi,

Thanks Andrew for giving me advise for upgrading heartbeat 2.1.3 to
pacemaker 1.0.4 anlong with heartbeat 2.99.2.

After Installing and configuring I run couple of tests, like turning off the
nodes, starting  and rebooting them. Working fine what I desired that moment
I've assigned private
IPs on the nodes i.e. node, ha1 IP: 192.168.0.184 and node ha2
IP:192.168.0.185.

But when I assigned ha1 and ha2 as a public IP, I'm still facing probelm
like previous version of heartbeat 2.1.3.
When I turn off ha2 and power it up it shows following status;

============
Last updated: Thu Jul 30 17:46:12 2009
Stack: Heartbeat
Current DC: ha2 (70503c2e-bb4a-48f8-aab3-53696656a4d0) - partition with
quorum
Version: 1.0.4-6dede86d6105786af3a5321ccf66b44b6914f0aa
2 Nodes configured, unknown expected votes
4 Resources configured.
============
OFFLINE: [ ha1  ]
Online: [ ha2  ]

IPaddr_1        (ocf::heartbeat:IPaddr):        Started ha2
IPaddr_2        (ocf::heartbeat:IPaddr):        Started ha2
OpenSips_1      (ocf::heartbeat:OpenSips):      Started ha2
OpenSips_2      (ocf::heartbeat:OpenSips):      Started ha2

On the other hand when I check the node ha1 which was failover to ha2, shows
its status as listed below;

============
Last updated: Thu Jul 30 17:46:12 2009
Stack: Heartbeat
Current DC: ha1 (e651c120-b9a1-489a-baf7-caf0028ad540) - partition with
quorum
Version: 1.0.4-6dede86d6105786af3a5321ccf66b44b6914f0aa
2 Nodes configured, unknown expected votes
4 Resources configured.
============
OFFLINE: [ ha2  ]
Online: [ ha1  ]

IPaddr_1        (ocf::heartbeat:IPaddr):        Started ha1
IPaddr_2        (ocf::heartbeat:IPaddr):        Started ha1
OpenSips_1      (ocf::heartbeat:OpenSips):      Started ha1
OpenSips_2      (ocf::heartbeat:OpenSips):      Started ha1

When I checked the logs it shows me that ha2 is not member of ha1, as I'm
listing down the logs below;

Jul 30 18:10:34 ha1 crmd: [2841]: WARN: crmd_ha_msg_callback: Ignoring HA
message (op=vote) from ha2: not in our membership list (size=1)
Jul 30 18:10:37 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_slave_all message (38) from ha2: not in our membership
Jul 30 18:10:37 ha1 attrd: [2840]: info: attrd_ha_callback: flush message
from ha2
Jul 30 18:10:39 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_replace message (3b) from ha2: not in our membership
Jul 30 18:10:39 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (3f) from ha2: not in our membership
Jul 30 18:10:41 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (41) from ha2: not in our membership
Jul 30 18:10:41 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (43) from ha2: not in our membership
Jul 30 18:10:41 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (45) from ha2: not in our membership
Jul 30 18:10:41 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (46) from ha2: not in our membership
Jul 30 18:10:41 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (47) from ha2: not in our membership
Jul 30 18:10:41 ha1 attrd: [2840]: info: attrd_ha_callback: flush message
from ha2
Jul 30 18:10:42 ha1 last message repeated 3 times
Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (4a) from ha2: not in our membership
Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (4b) from ha2: not in our membership
Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (4c) from ha2: not in our membership
Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (4d) from ha2: not in our membership
Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (4e) from ha2: not in our membership
Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (4f) from ha2: not in our membership
Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (50) from ha2: not in our membership
Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (52) from ha2: not in our membership
Jul 30 18:10:42 ha1 cib: [2837]: WARN: cib_peer_callback: Discarding
cib_apply_diff message (53) from ha2: not in our membership
Jul 30 18:10:55 ha1 ccm: [2836]: info: Break tie for 2 nodes cluster
Jul 30 18:10:55 ha1 crmd: [2841]: info: mem_handle_event: Got an event
OC_EV_MS_INVALID from ccm
Jul 30 18:10:55 ha1 cib: [2837]: info: mem_handle_event: Got an event
OC_EV_MS_INVALID from ccm
Jul 30 18:10:55 ha1 crmd: [2841]: info: mem_handle_event: no mbr_track info
Jul 30 18:10:55 ha1 cib: [2837]: info: mem_handle_event: no mbr_track info
Jul 30 18:10:55 ha1 crmd: [2841]: info: mem_handle_event: Got an event
OC_EV_MS_NEW_MEMBERSHIP from ccm
Jul 30 18:10:55 ha1 cib: [2837]: info: mem_handle_event: Got an event
OC_EV_MS_NEW_MEMBERSHIP from ccm
Jul 30 18:10:55 ha1 crmd: [2841]: info: mem_handle_event: instance=31,
nodes=1, new=0, lost=0, n_idx=0, new_idx=1, old_idx=3
Jul 30 18:10:55 ha1 cib: [2837]: info: mem_handle_event: instance=31,
nodes=1, new=0, lost=0, n_idx=0, new_idx=1, old_idx=3
Jul 30 18:10:55 ha1 crmd: [2841]: info: crmd_ccm_msg_callback: Quorum
(re)attained after event=NEW MEMBERSHIP (id=31)
Jul 30 18:10:55 ha1 cib: [2837]: info: cib_ccm_msg_callback: Processing CCM
event=NEW MEMBERSHIP (id=31)
Jul 30 18:10:55 ha1 crmd: [2841]: info: ccm_event_detail: NEW MEMBERSHIP:
trans=31, nodes=1, new=0, lost=0 n_idx=0, new_idx=1, old_idx=3
Jul 30 18:10:55 ha1 crmd: [2841]: info: ccm_event_detail:       CURRENT: ha1
[nodeid=0, born=31]
Jul 30 18:10:55 ha1 crmd: [2841]: info: populate_cib_nodes_ha: Requesting
the list of configured nodes
Jul 30 18:10:57 ha1 cib: [2837]: info: cib_process_request: Operation
complete: op cib_modify for section nodes (origin=local/crmd/182,
version=0.37.66): ok (rc=0)
Jul 30 18:11:00 ha1 crmd: [2841]: WARN: crmd_ha_msg_callback: Ignoring HA
message (op=noop) from ha2: not in our membership list (size=1)


Kindly review my problem. I'm attaching my ha.cf and cib.xml along with and
please do reply.

-- 
Regards,

Ahmed Munir
<cib validate-with="pacemaker-1.0" crm_feature_set="3.0.1" have-quorum="1" admin_epoch="0" epoch="42" num_updates="0" cib-last-written="Thu Jul 30 18:25:59 2009" dc-uuid="e651c120-b9a1-489a-baf7-caf0028ad540">
  <configuration>
    <crm_config>
      <cluster_property_set id="cib-bootstrap-options">
        <nvpair id="cib-bootstrap-options-symmetric-cluster" name="symmetric-cluster" value="true"/>
        <nvpair id="cib-bootstrap-options-no-quorum-policy" name="no-quorum-policy" value="stop"/>
        <nvpair id="cib-bootstrap-options-default-resource-stickiness" name="default-resource-stickiness" value="0"/>
        <nvpair id="cib-bootstrap-options-default-resource-failure-stickiness" name="default-resource-failure-stickiness" value="0"/>
        <nvpair id="cib-bootstrap-options-stonith-enabled" name="stonith-enabled" value="false"/>
        <nvpair id="cib-bootstrap-options-stonith-action" name="stonith-action" value="reboot"/>
        <nvpair id="cib-bootstrap-options-startup-fencing" name="startup-fencing" value="true"/>
        <nvpair id="cib-bootstrap-options-stop-orphan-resources" name="stop-orphan-resources" value="true"/>
        <nvpair id="cib-bootstrap-options-stop-orphan-actions" name="stop-orphan-actions" value="true"/>
        <nvpair id="cib-bootstrap-options-remove-after-stop" name="remove-after-stop" value="false"/>
        <nvpair id="cib-bootstrap-options-short-resource-names" name="short-resource-names" value="true"/>
        <nvpair id="cib-bootstrap-options-transition-idle-timeout" name="transition-idle-timeout" value="5min"/>
        <nvpair id="cib-bootstrap-options-default-action-timeout" name="default-action-timeout" value="20s"/>
        <nvpair id="cib-bootstrap-options-is-managed-default" name="is-managed-default" value="true"/>
        <nvpair id="cib-bootstrap-options-cluster-delay" name="cluster-delay" value="60s"/>
        <nvpair id="cib-bootstrap-options-pe-error-series-max" name="pe-error-series-max" value="-1"/>
        <nvpair id="cib-bootstrap-options-pe-warn-series-max" name="pe-warn-series-max" value="-1"/>
        <nvpair id="cib-bootstrap-options-pe-input-series-max" name="pe-input-series-max" value="-1"/>
        <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="1.0.4-6dede86d6105786af3a5321ccf66b44b6914f0aa"/>
        <nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="Heartbeat"/>
      </cluster_property_set>
    </crm_config>
    <nodes>
      <node id="70503c2e-bb4a-48f8-aab3-53696656a4d0" uname="ha2" type="normal"/>
      <node id="e651c120-b9a1-489a-baf7-caf0028ad540" uname="ha1" type="normal"/>
    </nodes>
    <resources>
      <primitive class="ocf" provider="heartbeat" type="IPaddr" id="IPaddr_1">
        <operations>
          <op id="IPaddr_1_mon" interval="10s" name="monitor" timeout="8s"/>
        </operations>
        <instance_attributes id="IPaddr_1_inst_attr">
          <nvpair name="ip" value="(Public IP 1)" id="IPaddr_1_machine_1"/>
          <nvpair id="IP1_attr_1" name="netmask" value="27"/>
          <nvpair id="IP1_attr_2" name="nic" value="eth0:0"/>
        </instance_attributes>
      </primitive>
      <primitive class="ocf" provider="heartbeat" type="IPaddr" id="IPaddr_2">
        <operations>
          <op id="IPaddr_2_mon" interval="10s" name="monitor" timeout="8s"/>
        </operations>
        <instance_attributes id="IPaddr_2_inst_attr">
          <nvpair name="ip" value="(Public IP 2)" id="IPaddr_2_machine_2"/>
          <nvpair id="IP2_attr_1" name="netmask" value="27"/>
          <nvpair id="IP2_attr_2" name="nic" value="eth0:0"/>
        </instance_attributes>
      </primitive>
 <primitive class="ocf" provider="heartbeat" id="OpenSips_1" type="OpenSips">
        <operations>
          <op id="opensips-start-ha1" name="start" interval="0" timeout="15s"/>
          <op id="opensips-stop-ha1" name="stop" interval="0" timeout="10s"/>
          <op name="monitor" timeout="10s" id="opensips-monitor-ha1" interval="12s">
            <instance_attributes id="ha1-monitor_10s">
              <nvpair name="ip" id="ha1-opensips-monitor-ip" value="(Public IP 1)"/>
            </instance_attributes>
          </op>
        </operations>
      </primitive>
      <primitive class="ocf" provider="heartbeat" id="OpenSips_2" type="OpenSips">
        <operations>
          <op id="opensips-start-ha2" name="start" interval="0" timeout="15s"/>
          <op id="opensips-stop-ha2" name="stop" interval="0" timeout="10s"/>
          <op name="monitor" timeout="10s" id="opensips-monitor-ha2" interval="12s">
            <instance_attributes id="ha2-monitor_10s">
              <nvpair name="ip" id="ha2-opensips-monitor-ip" value="(Public IP 2)"/>
            </instance_attributes>
          </op>
        </operations>
      </primitive>
    </resources>
    <constraints>
      <rsc_location id="OpenSips_resource_location_1" rsc="OpenSips_1">
        <rule id="rule_ha1" score="100">
          <expression id="expression_uname_eq_ha1" attribute="#uname" operation="eq" value="ha1"/>
        </rule>
      </rsc_location>
      <rsc_location id="OpenSips_resource_location_2" rsc="OpenSips_2">
        <rule id="rule_ha2" score="100">
          <expression id="expression_uname_eq_ha2" attribute="#uname" operation="eq" value="ha2"/>
        </rule>
      </rsc_location>
      <rsc_location id="rsc_location_IPaddr_1" rsc="IPaddr_1">
        <rule id="prefered_location_IPaddr_1" score="200">
          <expression attribute="#uname" id="prefered_location_IPaddr_1_expr" operation="eq" value="ha1"/>
        </rule>
      </rsc_location>
      <rsc_location id="rsc_location_IPaddr_2" rsc="IPaddr_2">
        <rule id="prefered_location_IPaddr_2" score="200">
          <expression attribute="#uname" id="prefered_location_IPaddr_2_expr" operation="eq" value="ha2"/>
        </rule>
      </rsc_location>
      <rsc_location id="my1_resource1_connected" rsc="IPaddr_1">
        <rule id="my1_resource1_connected_rule" score-attribute="pingd">
          <expression id="my1_resource1_connected_expr_defined" attribute="pingd" operation="defined"/>
        </rule>
      </rsc_location>
      <rsc_location id="my2_resource2_connected" rsc="IPaddr_2">
        <rule id="my2_resource2_connected_rule" score-attribute="pingd">
          <expression id="my2_resource2_connected_expr_defined" attribute="pingd" operation="defined"/>
        </rule>
      </rsc_location>
    </constraints>
  </configuration>
</cib>

Attachment: ha.cf
Description: Binary data

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to