Hello list,

I have a two node cluster configured with a resource named "resource_ip1".
It's of type "IPaddr" and of class "ocf".

When the node1 hosting "resource_ip1" is rebooted it failsover to node2.
However the fail-count-resource for "resource_ip1" on node1 does not get
incremented and still shows up as 0.

         crm_failcount -G -U node1 -r resource_ip1
         name=fail-count-resource_ip1 value=0

Further more eventhough i have "Default Resource Stickiness" set to 100000,
and Default Failure Resource Stickiness set to -INFINITY [with no override
stickiness on resources] the resource "resource_ip1" fails back on to node 1
after it's back on-line.

Any ideas on why this is happening is very much appreciated?

The same config works fine for resource_xinetd. Please see below for the
current cib.xml

Regards
--Raghu

CIB.XML

 <cib generated="true" admin_epoch="0" have_quorum="true" ignore_dtd="false"
num_peers="2" cib_feature_revision="2.0" crm_feature_set="2.0"
ccm_transition="112" dc_uuid="1f0ffb39-b275-4e92-b6ca-8a3d00f2fb44"
epoch="187" num_updates="3" cib-last-written="Tue Jul  8 13:24:57 2008">
   <configuration>
     <crm_config>
       <cluster_property_set id="cib-bootstrap-options">
         <attributes>
           <nvpair id="cib-bootstrap-options-dc-version" name="dc-version"
value="2.1.3-node: a3184d5240c6e7032aef9cce6e5b7752ded544b3"/>
           <nvpair id="cib-bootstrap-options-stonith-enabled"
name="stonith-enabled" value="false"/>
           <nvpair name="default-resource-stickiness"
id="cib-bootstrap-options-default-resource-stickiness" value="100000"/>
           <nvpair name="default-resource-failure-stickiness"
id="cib-bootstrap-options-default-resource-failure-stickiness"
value="-INFINITY"/>
           <nvpair id="cib-bootstrap-options-last-lrm-refresh"
name="last-lrm-refresh" value="1215547282"/>
           <nvpair id="cib-bootstrap-options-no-quorum-policy"
name="no-quorum-policy" value="stop"/>
         </attributes>
       </cluster_property_set>
     </crm_config>
     <nodes>
       <node id="1f0ffb39-b275-4e92-b6ca-8a3d00f2fb44" uname="vcs9473"
type="normal"/>
       <node id="03fd6ad6-e7b1-4722-96d8-54e3be84a59c" uname="vcs9472"
type="normal"/>
     </nodes>
     <resources>
       <primitive id="resource_ip1" class="ocf" type="IPaddr"
provider="heartbeat">
         <instance_attributes id="58501e02-21f5-49d3-aebf-fca5e378ae70">
           <attributes>
             <nvpair name="ip" value="172.25.52.245"
id="430bc62c-d2d2-4054-87a3-a9fe041f0ecc"/>
           </attributes>
         </instance_attributes>
       </primitive>
       <primitive id="resource_xinetd" class="lsb" type="xinetd"
provider="heartbeat">
         <meta_attributes id="resource_xinetd_meta_attrs">
           <attributes/>
         </meta_attributes>
         <operations>
           <op id="b33c816d-d85a-47f2-bb35-0edc300da907" name="monitor"
interval="15" timeout="15" start_delay="15" disabled="false" role="Started"
on_fail="restart"/>
         </operations>
       </primitive>
     </resources>
     <constraints>
       <rsc_colocation id="colocation_ftp" from="resource_ip1"
to="resource_xinetd" score="INFINITY"/>
     </constraints>
   </configuration>
 </cib>
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to