Hi, On Wed, Jan 16, 2008 at 10:09:37AM +0800, Tonglu Yi wrote: > Because that would be involve too much code change and testing working > and our time table is very tight
A reply to a wrong message perhaps? Thanks, Dejan > On Jan 16, 2008 1:12 AM, Dejan Muhamedagic <[EMAIL PROTECTED]> wrote: > > > Hi, > > > > On Tue, Jan 15, 2008 at 10:51:48PM +0800, Tonglu Yi wrote: > > > Hi Thomas > > > Thanks. > > > > > > But I am using heartbeat 2.1.2 with release 1 style configuration, and > > it is > > > impossible to change to use release 2 configuration. > > > > > > Actually what I expect heartbeat to do is that at startup time, > > heartbeat > > > goes to stanby if there is any resource failed to start, and let the > > other > > > node to take over resources when the other server starts up. > > > > > > Currently the problem is that heartbeat does give up all resources, but > > > after giving up all resources, it still thinks itself holding all > > resources, > > > thus the peer is unable to take over resources when starting up. > > > > Could you provide the logs. > > > > Thanks, > > > > Dejan > > > > > > > > > On Jan 15, 2008 10:39 PM, Thomas Glanzmann <[EMAIL PROTECTED]> wrote: > > > > > > > Hello Tonglu, > > > > the following setup works with heartbeat-2.1.3 and drbd-8.1.3: > > > > > > > > cibadmin -U -X ' > > > > <configuration> > > > > <resources> > > > > <master_slave id="ms-drbd0"> > > > > <meta_attributes id="ma-ms-drbd0"> > > > > <attributes> > > > > <nvpair id="ma-ms-drbd0-1" > > > > name="clone_max" value="2"/> > > > > <nvpair id="ma-ms-drbd0-2" > > > > name="clone_node_max" value="1"/> > > > > <nvpair id="ma-ms-drbd0-3" > > > > name="master_max" value="1"/> > > > > <nvpair id="ma-ms-drbd0-4" > > > > name="master_node_max" value="1"/> > > > > <nvpair id="ma-ms-drbd0-5" > > > > name="notify" value="yes"/> > > > > <nvpair id="ma-ms-drbd0-6" > > > > name="globally_unique" value="false"/> > > > > </attributes> > > > > </meta_attributes> > > > > <primitive id="drbd0" class="ocf" > > > > provider="heartbeat" type="drbd"> > > > > <instance_attributes id="ia-drbd0"> > > > > <attributes> > > > > <nvpair id="ia-drbd0-1" > > > > name="drbd_resource" value="postgres"/> > > > > </attributes> > > > > </instance_attributes> > > > > <operations> > > > > <op id="op-ms-drbd2-1" > > > > name="monitor" interval="60s" timeout="60s" start_delay="30s" > > > > role="Master"/> > > > > <op id="op-ms-drbd2-2" > > > > name="monitor" interval="61s" timeout="60s" start_delay="30s" > > role="Slave"/> > > > > </operations> > > > > </primitive> > > > > </master_slave> > > > > > > > > <group id="nfs-cluster"> > > > > <primitive class="ocf" provider="heartbeat" > > > > type="Filesystem" id="fs0"> > > > > <instance_attributes id="ia-fs0"> > > > > <attributes> > > > > <nvpair id="ia-fs0-1" > > > > name="fstype" value="ext3"/> > > > > <nvpair > > name="directory" > > > > id="ia-fs0-2" value="/srv/gcl"/> > > > > <nvpair id="ia-fs0-3" > > > > name="device" value="/dev/drbd0"/> > > > > </attributes> > > > > </instance_attributes> > > > > <operations> > > > > <op id="fs0-monitor0" > > > > name="monitor" interval="60s" timeout="120s" start_delay="1m"/> > > > > </operations> > > > > </primitive> > > > > </group> > > > > </resources> > > > > > > > > <constraints> > > > > <rsc_location id="drbd0-placement-1" rsc="ms-drbd0"> > > > > <rule id="drbd0-rule-1" score="-INFINITY"> > > > > <expression id="exp-01" value="ha-1" > > > > attribute="#uname" operation="ne"/> > > > > <expression id="exp-02" value="ha-2" > > > > attribute="#uname" operation="ne"/> > > > > </rule> > > > > </rsc_location> > > > > > > > > <rsc_order id="nfs_promotes_ms-drbd0" > > from="nfs-cluster" > > > > action="start" to="ms-drbd0" to_action="promote"/> > > > > <rsc_colocation id="nfs_on_drbd0" to="ms-drbd0" > > > > to_role="master" from="nfs-cluster" score="infinity"/> > > > > </constraints> > > > > </configuration> > > > > ' > > > > > > > > Thomas > > > > _______________________________________________ > > > > Linux-HA mailing list > > > > [email protected] > > > > http://lists.linux-ha.org/mailman/listinfo/linux-ha > > > > See also: http://linux-ha.org/ReportingProblems > > > > > > > _______________________________________________ > > > Linux-HA mailing list > > > [email protected] > > > http://lists.linux-ha.org/mailman/listinfo/linux-ha > > > See also: http://linux-ha.org/ReportingProblems > > _______________________________________________ > > Linux-HA mailing list > > [email protected] > > http://lists.linux-ha.org/mailman/listinfo/linux-ha > > See also: http://linux-ha.org/ReportingProblems > > > _______________________________________________ > Linux-HA mailing list > [email protected] > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
