Because that would be involve too much code change and testing working
and our time table is very tight

On Jan 16, 2008 1:12 AM, Dejan Muhamedagic <[EMAIL PROTECTED]> wrote:

> Hi,
>
> On Tue, Jan 15, 2008 at 10:51:48PM +0800, Tonglu Yi wrote:
> > Hi Thomas
> > Thanks.
> >
> > But I am using heartbeat 2.1.2 with release 1 style configuration, and
> it is
> > impossible to change to use release 2 configuration.
> >
> > Actually what I expect heartbeat to do is that at startup time,
> heartbeat
> > goes to stanby if there is any resource failed to start, and let the
> other
> > node to take over resources when the other server starts up.
> >
> > Currently the problem is that heartbeat does give up all resources, but
> > after giving up all resources, it still thinks itself holding all
> resources,
> > thus the peer is unable to take over resources when starting up.
>
> Could you provide the logs.
>
> Thanks,
>
> Dejan
>  >
> >
> > On Jan 15, 2008 10:39 PM, Thomas Glanzmann <[EMAIL PROTECTED]> wrote:
> >
> > > Hello Tonglu,
> > > the following setup works with heartbeat-2.1.3 and drbd-8.1.3:
> > >
> > > cibadmin -U -X '
> > > <configuration>
> > >        <resources>
> > >                <master_slave id="ms-drbd0">
> > >                        <meta_attributes id="ma-ms-drbd0">
> > >                                <attributes>
> > >                                        <nvpair id="ma-ms-drbd0-1"
> > > name="clone_max" value="2"/>
> > >                                        <nvpair id="ma-ms-drbd0-2"
> > > name="clone_node_max" value="1"/>
> > >                                        <nvpair id="ma-ms-drbd0-3"
> > > name="master_max" value="1"/>
> > >                                        <nvpair id="ma-ms-drbd0-4"
> > > name="master_node_max" value="1"/>
> > >                                        <nvpair id="ma-ms-drbd0-5"
> > > name="notify" value="yes"/>
> > >                                        <nvpair id="ma-ms-drbd0-6"
> > > name="globally_unique" value="false"/>
> > >                                </attributes>
> > >                        </meta_attributes>
> > >                        <primitive id="drbd0" class="ocf"
> > > provider="heartbeat" type="drbd">
> > >                                <instance_attributes id="ia-drbd0">
> > >                                        <attributes>
> > >                                                <nvpair id="ia-drbd0-1"
> > > name="drbd_resource" value="postgres"/>
> > >                                        </attributes>
> > >                                </instance_attributes>
> > >                                <operations>
> > >                                        <op id="op-ms-drbd2-1"
> > > name="monitor" interval="60s" timeout="60s" start_delay="30s"
> > > role="Master"/>
> > >                                        <op id="op-ms-drbd2-2"
> > > name="monitor" interval="61s" timeout="60s" start_delay="30s"
> role="Slave"/>
> > >                                </operations>
> > >                        </primitive>
> > >                </master_slave>
> > >
> > >                <group id="nfs-cluster">
> > >                        <primitive class="ocf" provider="heartbeat"
> > > type="Filesystem" id="fs0">
> > >                                <instance_attributes id="ia-fs0">
> > >                                        <attributes>
> > >                                                <nvpair id="ia-fs0-1"
> > > name="fstype" value="ext3"/>
> > >                                                <nvpair
> name="directory"
> > > id="ia-fs0-2" value="/srv/gcl"/>
> > >                                                <nvpair id="ia-fs0-3"
> > > name="device" value="/dev/drbd0"/>
> > >                                        </attributes>
> > >                                </instance_attributes>
> > >                                <operations>
> > >                                        <op id="fs0-monitor0"
> > > name="monitor" interval="60s" timeout="120s" start_delay="1m"/>
> > >                                </operations>
> > >                        </primitive>
> > >                </group>
> > >        </resources>
> > >
> > >        <constraints>
> > >                <rsc_location id="drbd0-placement-1" rsc="ms-drbd0">
> > >                        <rule id="drbd0-rule-1" score="-INFINITY">
> > >                                <expression id="exp-01" value="ha-1"
> > > attribute="#uname" operation="ne"/>
> > >                                <expression id="exp-02" value="ha-2"
> > > attribute="#uname" operation="ne"/>
> > >                        </rule>
> > >                </rsc_location>
> > >
> > >                <rsc_order id="nfs_promotes_ms-drbd0"
> from="nfs-cluster"
> > > action="start" to="ms-drbd0" to_action="promote"/>
> > >                <rsc_colocation id="nfs_on_drbd0" to="ms-drbd0"
> > > to_role="master" from="nfs-cluster" score="infinity"/>
> > >        </constraints>
> > > </configuration>
> > > '
> > >
> > >        Thomas
> > > _______________________________________________
> > > Linux-HA mailing list
> > > [email protected]
> > > http://lists.linux-ha.org/mailman/listinfo/linux-ha
> > > See also: http://linux-ha.org/ReportingProblems
> > >
> > _______________________________________________
> > Linux-HA mailing list
> > [email protected]
> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
> > See also: http://linux-ha.org/ReportingProblems
> _______________________________________________
> Linux-HA mailing list
> [email protected]
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
>
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to