On 10/26/07, Terry L. Inzauro <[EMAIL PROTECTED]> wrote:
> Andrew Beekhof wrote:
> > On 10/25/07, Terry L. Inzauro <[EMAIL PROTECTED]> wrote:
> >> Andrew Beekhof wrote:
> >>> On 10/23/07, Terry L. Inzauro <[EMAIL PROTECTED]> wrote:
> >>>> list,
> >>>>
> >>>> i think i have pingd working properly.
> >>>>
> >>>> /etc/ha.d/ha.cf
> >>>>         apiauth ping gid=root uid=root
> >>>>         respawn root /usr/lib/heartbeat/pingd -m 1000 -d 5s -a 
> >>>> default_ping_set
> >>> here use tell pingd to define "default_ping_set"
> >>>
> >>>> cib.xml locational constraint:
> >>>> <constraints>
> >>>>        <rsc_location id="afp_cl_vpn_loc" rsc="afp_cl_vpn">
> >>>>          <rule id="afp_cl_vpn_pref_1" score="100">
> >>>>            <expression id="afp_cl_vpn_loc_attr_1" attribute="#uname" 
> >>>> operation="eq" value="clvpn01"/>
> >>>>          </rule>
> >>>>          <rule id="afp_cl_vpn_loc_pingd_rule" 
> >>>> score_attribute="afp_cl_vpn">
> >>>>            <expression id="afp_cl_vpn_loc_pingd_rule_defined" 
> >>>> attribute="pingd" operation="defined"/>
> >>> but here you tell the PE to look for "pingd"
> >>>
> >>> s/pingd/default_ping_set/
> >>>
> >>>>          </rule>
> >>>>        </rsc_location>
> >>>> </constraints>
> >>>>
> >>>>
> >>>> my question is in regards to resource fail back.  how does one get the 
> >>>> resource to move back to the
> >>>> primary node in the event that the resource moves from its primary 
> >>>> cluster node to the secondary
> >>>> node as a result of its pingd score?
> >>> it depends on how default-resource-stickiness is set
> >>> if its > 0, then you need to move it manually with crm_resource -M
> >>>
> >>>> is this a manual administrtive process or 'should' hb take care
> >>>> of that for you?
> >>>>
> >>>>
> >>>> regards,
> >>>>
> >>>>
> >>>> _Terry
> >>>>
> >>
> >> ok, after resolving my mistakes in naming, i began to test.  the way i 
> >> understand this to work is
> >> that the resource will
> >> be placed on the node with the highest score.  i'm simulating network 
> >> failure by placing black hole
> >> routes on the node
> >> i wish to test (route add -host 172.16.1.2 gw 127.0.0.1). i would 'assume' 
> >> that the resource would
> >> be moved to the
> >> node with the best connectivity aka score...it does, but it also offlines 
> >> the node that i added the
> >> black hole route to.
> >
> > in what way did the node go offline?
> > (in heartbeat, the crm, physically?)
> >
> > if just in the crm, then please upgrade.  there was a bug about this
> > some time ago.
> >
> >> how does one reverse the effects of the off lining, short of restarting 
> >> heartbeat of both nodes?
> >>
> >> is that what ping is supposed to do? if so, what if i had other resources 
> >> configured to run on that
> >> node?
>
> yes, the node goes off line in crm, but heartbeat remains running but hangs 
> when an attempting to
> shut down.
>
> i was hoping you weren't going to say that.  i'm using debian etch(stable)
> and i like keeping the packages as close to their release cycle as possible.

not sure i understand what you mean there... which release cycle are
you talking about?

anyway, 2.0.7 is horribly old and i believe horms updated the official
debian heartbeat packages a week or two ago.

>
> clvpn01:/var/lib/heartbeat/crm# /usr/lib/heartbeat/heartbeat -V
> 2.0.7
>
> can you recommend which version would be the most stable/recommended? ;) in 
> sid, the latest one is:
> http://packages.debian.org/sid/heartbeat/i386/download

yep, thats a pretty good version.
or try the ones here which have a few extra bug fixes:
   http://download.opensuse.org/repositories/server:/ha-clustering/Debian_Etch/
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to