On Fri, Jun 4, 2010 at 11:25 PM, Tony Hunter <[email protected]> wrote:
> On Thu, Jun 03, 2010 at 07:18:16PM -0300, Diego Woitasen wrote:
>> On Wed, Jun 2, 2010 at 7:43 AM, Andrew Beekhof <[email protected]> wrote:
>>
>> > On Sat, May 29, 2010 at 3:54 AM, Diego Woitasen <[email protected]>
>> > wrote:
>> > > Hi,
>> > >  * I have three nodes: "ha1", "ha2" y "ha3".
>> > >  * Three resources: "sfex", "xfs_fs", "ip".
>> > >  * "sfex" and "xfs_fs" are members of a group called "xfs_grp".
>> > >  * "xfs_grp" can run on any node but "ip" resource can run on "ha1" or
>> > > "ha2" only.
>> > >  * When "xfs_grp" is running on "ha1" or "ha2", "ip" must run on the same
>> > > node.
>> > >  * One last thing, I need manual failback.
>> > >
>> > > My current configuration works except for the "manual failback" (a.k.a.
>> > > auto_failback off).
>> > >
>> > > node $id="0ace77ab-600a-4541-a682-ab0534bb3fc4" ha3
>> > > node $id="3d1f07b5-a79b-478f-b07c-02a7a5c5106c" ha2
>> > > node $id="c44a3a26-35d4-476e-a1e6-49f03f068f12" ha1
>> > > primitive ip ocf:heartbeat:IPaddr \
>> > >        params ip="192.168.1.147"
>> > > primitive sfex ocf:heartbeat:sfex \
>> > >        params device="/dev/sdb1" \
>> > >        op monitor interval="10" timeout="10" depth="0"
>> > > primitive xfs_fs ocf:heartbeat:Filesystem \
>> > >        params device="/dev/sdb2" directory="/shared" fstype="xfs" \
>> > >        op monitor interval="20" timeout="40" depth="0"
>> > > group xfs_grp sfex xfs_fs
>> > > location srv_loc ip -inf: ha3
>> > > colocation srv_col inf: ip xfs_grp
>> > > property $id="cib-bootstrap-options" \
>> > >        no-quorum-policy="ignore" \
>> > >        expected-quorum-votes="1" \
>> > >        stonith-enabled="0" \
>> > >        default-resource-stickiness="INFINITY"
>> > >
>> > > When "xfs_grp" is running in "ha3" and "ha1" or "ha2" are alive again,
>> > the
>> > > resources ("xfs_grp" and "ip") move to any of them.
>> > >
>> > > Any ideas?
>> >
>> > Not really, I don't understand what the problem is.
>> > ip can only run on ha1 or ha2, so its not surprising that it gets
>> > stopped occasionally (ie. when you shut down one node and make the
>> > other standby) while the group remains running.
>> > _______________________________________________
>> > Linux-HA mailing list
>> > [email protected]
>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> > See also: http://linux-ha.org/ReportingProblems
>> >
>>
>> May be my explanation was wrong.
>>
>> xfs_grp can run on ha1, ha2 or ha3.
>> ip can run on ha1 or ha2.
>>
>> If I shutdown ha1 and ha2, xfs_grp moves to ha3 without "ip". If ha1 (or
>> ha2) returns back, xfs_grp moves to ha1 and "ip" are started. I have
>> default-resource-stickiness="
>> INFINITY" so I think that xfs_grp should stays in ha3 until manual failback.
>
> I'm fairly new to pacemaker/corosync but I haven't seen a reply to your mail,
> so I'll take a shot. It seems your colocation rule below prevents xfs_grp
> from running on ha3 unless ip is also running there:
> colocation srv_col inf: ip xfs_grp
>
> And this rule seems to suggest the ip resource should _never_ run on ha3.
> location srv_loc ip -inf: ha3
>
> So, as far as I can see, the cluster is behaving as configured - ha1 or ha2
> takes over ip when one of them comes back online, since ip is not running
> anywhere. And of course xfs_grp is migrated because of the colocation
> constraint.

exactly
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to