On Wed, Jul 9, 2008 at 19:47, Raghuram Bondalapati
<[EMAIL PROTECTED]> wrote:
> Andrew, the Score for Node vcs9472 for Resource resource_ip1 is always set
> to -INFINITY.
>
> This is preventing the resource_ip1 from sticking to vcs9472, when nodes
> vcs9473 reboots and come back on-line. [resource_ip1 always migrates back to
> vcs9473].
>
> Is there a way to reset the score for resource_ip1 on vcs9472?

try crm_resource -C

>
> I am running hearbeat 2.1.3 with pacemaker 0.6.5 extensions.
>
> Resource            Score     Node            Stickiness #Fail
> Fail-Stickiness
> resource_ip1        -INFINITY vcs9472         100000     0
> -INFINITY
> resource_ip1        100000    vcs9473         100000     0
> -INFINITY
> resource_xinetd     -INFINITY vcs9472         100000     0
> -INFINITY
> resource_xinetd     -INFINITY vcs9473         100000     0
> -INFINITY
> resource_xinetd1    0         vcs9472         100000     0
> -INFINITY
> resource_xinetd1    200000    vcs9473         100000     0
> -INFINITY
>
>
>
> On 7/9/08, Andrew Beekhof <[EMAIL PROTECTED]> wrote:
>>
>> Known bug in 2.1.3
>>
>> Please grab the latest Pacemaker release (0.6.5) for you distro from
>>   http://download.opensuse.org/repositories/server:/ha-clustering/
>>
>> On Tue, Jul 8, 2008 at 22:51, Raghuram Bondalapati
>> <[EMAIL PROTECTED]> wrote:
>> > Hello list,
>> >
>> > I have a two node cluster configured with a resource named
>> "resource_ip1".
>> > It's of type "IPaddr" and of class "ocf".
>> >
>> > When the node1 hosting "resource_ip1" is rebooted it failsover to node2.
>> > However the fail-count-resource for "resource_ip1" on node1 does not get
>> > incremented and still shows up as 0.
>> >
>> >         crm_failcount -G -U node1 -r resource_ip1
>> >         name=fail-count-resource_ip1 value=0
>> >
>> > Further more eventhough i have "Default Resource Stickiness" set to
>> 100000,
>> > and Default Failure Resource Stickiness set to -INFINITY [with no
>> override
>> > stickiness on resources] the resource "resource_ip1" fails back on to
>> node 1
>> > after it's back on-line.
>> >
>> > Any ideas on why this is happening is very much appreciated?
>> >
>> > The same config works fine for resource_xinetd. Please see below for the
>> > current cib.xml
>> >
>> > Regards
>> > --Raghu
>> >
>> > CIB.XML
>> >
>> >  <cib generated="true" admin_epoch="0" have_quorum="true"
>> ignore_dtd="false"
>> > num_peers="2" cib_feature_revision="2.0" crm_feature_set="2.0"
>> > ccm_transition="112" dc_uuid="1f0ffb39-b275-4e92-b6ca-8a3d00f2fb44"
>> > epoch="187" num_updates="3" cib-last-written="Tue Jul  8 13:24:57 2008">
>> >   <configuration>
>> >     <crm_config>
>> >       <cluster_property_set id="cib-bootstrap-options">
>> >         <attributes>
>> >           <nvpair id="cib-bootstrap-options-dc-version" name="dc-version"
>> > value="2.1.3-node: a3184d5240c6e7032aef9cce6e5b7752ded544b3"/>
>> >           <nvpair id="cib-bootstrap-options-stonith-enabled"
>> > name="stonith-enabled" value="false"/>
>> >           <nvpair name="default-resource-stickiness"
>> > id="cib-bootstrap-options-default-resource-stickiness" value="100000"/>
>> >           <nvpair name="default-resource-failure-stickiness"
>> > id="cib-bootstrap-options-default-resource-failure-stickiness"
>> > value="-INFINITY"/>
>> >           <nvpair id="cib-bootstrap-options-last-lrm-refresh"
>> > name="last-lrm-refresh" value="1215547282"/>
>> >           <nvpair id="cib-bootstrap-options-no-quorum-policy"
>> > name="no-quorum-policy" value="stop"/>
>> >         </attributes>
>> >       </cluster_property_set>
>> >     </crm_config>
>> >     <nodes>
>> >       <node id="1f0ffb39-b275-4e92-b6ca-8a3d00f2fb44" uname="vcs9473"
>> > type="normal"/>
>> >       <node id="03fd6ad6-e7b1-4722-96d8-54e3be84a59c" uname="vcs9472"
>> > type="normal"/>
>> >     </nodes>
>> >     <resources>
>> >       <primitive id="resource_ip1" class="ocf" type="IPaddr"
>> > provider="heartbeat">
>> >         <instance_attributes id="58501e02-21f5-49d3-aebf-fca5e378ae70">
>> >           <attributes>
>> >             <nvpair name="ip" value="172.25.52.245"
>> > id="430bc62c-d2d2-4054-87a3-a9fe041f0ecc"/>
>> >           </attributes>
>> >         </instance_attributes>
>> >       </primitive>
>> >       <primitive id="resource_xinetd" class="lsb" type="xinetd"
>> > provider="heartbeat">
>> >         <meta_attributes id="resource_xinetd_meta_attrs">
>> >           <attributes/>
>> >         </meta_attributes>
>> >         <operations>
>> >           <op id="b33c816d-d85a-47f2-bb35-0edc300da907" name="monitor"
>> > interval="15" timeout="15" start_delay="15" disabled="false"
>> role="Started"
>> > on_fail="restart"/>
>> >         </operations>
>> >       </primitive>
>> >     </resources>
>> >     <constraints>
>> >       <rsc_colocation id="colocation_ftp" from="resource_ip1"
>> > to="resource_xinetd" score="INFINITY"/>
>> >     </constraints>
>> >   </configuration>
>> >  </cib>
>> > _______________________________________________
>> > Linux-HA mailing list
>> > [email protected]
>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> > See also: http://linux-ha.org/ReportingProblems
>> >
>> _______________________________________________
>> Linux-HA mailing list
>> [email protected]
>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> See also: http://linux-ha.org/ReportingProblems
>>
> _______________________________________________
> Linux-HA mailing list
> [email protected]
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
>
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to