Hi, On Mon, Dec 10, 2007 at 04:27:37PM +0100, Franck Ganachaud wrote: > Ok, here is where I am now. > > I built a clone set and colocated the group1 with the clone. > But when I stop mysql, the clone monitor operation (mysql_orb) return that > DB is down but group1 isn't relocated to serveur B. > > Anyone has a clue?
Can you post logs? If you have hb_report you can use that. > I copy the cib : > > <configuration> > <crm_config> > <cluster_property_set id="cib-bootstrap-options"> > <attributes> > <nvpair id="cib-bootstrap-options-symmetric-cluster" > name="symmetric-cluster" value="true"/> > <nvpair id="cib-bootstrap-options-no_quorum-policy" > name="no_quorum-policy" value="stop"/> > <nvpair id="cib-bootstrap-options-default-resource-stickiness" > name="default-resource-stickiness" value="0"/> > <nvpair > id="cib-bootstrap-options-default-resource-failure-stickiness" > name="default-resource-failure-stickiness" value="0"/> > <nvpair id="cib-bootstrap-options-stonith-enabled" > name="stonith-enabled" value="false"/> > <nvpair id="cib-bootstrap-options-stonith-action" > name="stonith-action" value="reboot"/> > <nvpair id="cib-bootstrap-options-stop-orphan-resources" > name="stop-orphan-resources" value="true"/> > <nvpair id="cib-bootstrap-options-stop-orphan-actions" > name="stop-orphan-actions" value="true"/> > <nvpair id="cib-bootstrap-options-remove-after-stop" > name="remove-after-stop" value="false"/> > <nvpair id="cib-bootstrap-options-short-resource-names" > name="short-resource-names" value="true"/> > <nvpair id="cib-bootstrap-options-transition-idle-timeout" > name="transition-idle-timeout" value="5min"/> > <nvpair id="cib-bootstrap-options-default-action-timeout" > name="default-action-timeout" value="5s"/> > <nvpair id="cib-bootstrap-options-is-managed-default" > name="is-managed-default" value="true"/> > </attributes> > </cluster_property_set> > </crm_config> > <nodes> > <node id="0e8b2fa4-983b-4e56-a4a5-72dbb2aeaeec" type="normal" > uname="server_a"> > <instance_attributes > id="nodes-0e8b2fa4-983b-4e56-a4a5-72dbb2aeaeec"> > <attributes> > <nvpair id="standby-0e8b2fa4-983b-4e56-a4a5-72dbb2aeaeec" > name="standby" value="off"/> > </attributes> > </instance_attributes> > </node> > <node id="5cdd04e8-035a-44cf-ab60-3065840109db" type="normal" > uname="server_b"> > <instance_attributes > id="nodes-5cdd04e8-035a-44cf-ab60-3065840109db"> > <attributes> > <nvpair id="standby-5cdd04e8-035a-44cf-ab60-3065840109db" > name="standby" value="off"/> > </attributes> > </instance_attributes> > </node> > <node id="9e05d57a-ae9c-430d-a210-d03b9f37739e" type="normal" > uname="server_b"/> > <node id="352f29b5-f0ed-4866-a839-71dbdbfd491d" type="normal" > uname="server_a"/> > </nodes> Nodes are listed twice. > <resources> > <clone id="MySQL_ORB" interleave="false" is_managed="true" > notify="false" ordered="false"> > <instance_attributes id="MySQL_ORB_inst_attr"> > <attributes> > <nvpair id="MySQL_ORB_attr_0" name="clone_max" value="2"/> > <nvpair id="MySQL_ORB_attr_1" name="clone_node_max" value="1"/> > </attributes> > </instance_attributes> > <primitive class="ocf" id="mysql_orb1" is_managed="false" > provider="heartbeat" type="mysql_orb"> > <operations> > <op id="mysql_orb_mon" interval="30s" name="monitor" > on_fail="stop" timeout="30s"/> > </operations> > </primitive> > </clone> > <group id="group_1" restart_type="restart"> > <primitive class="ocf" id="IPaddr_Cluster" provider="heartbeat" > type="IPaddr"> > <operations> > <op id="IPaddr_Cluster_mon" interval="5s" name="monitor" > timeout="5s"/> > </operations> > <instance_attributes id="IPaddr_Cluster_inst_attr"> > <attributes> > <nvpair id="IPaddr_Cluster_attr_0" name="ip" > value="192.168.87.100"/> > </attributes> > </instance_attributes> > </primitive> > <primitive class="ocf" id="apache_2" provider="heartbeat" > type="apache"> > <operations> > <op id="apache_2_mon" interval="30s" name="monitor" > timeout="30s"/> > </operations> > <instance_attributes id="apache_2_inst_attr"> > <attributes> > <nvpair id="apache_2_attr_0" name="configfile" > value="/usr/local/apache/conf/httpd.conf"/> > </attributes> > </instance_attributes> > </primitive> > </group> > </resources> > <constraints> > <rsc_location id="rsc_location_group_1" rsc="group_1"> > <rule id="prefered_location_group_1" score="100"> > <expression attribute="#uname" id="prefered_location_group_1_expr" > operation="eq" value="server_a"/> > </rule> > </rsc_location> > <rsc_colocation from="group_1" id="web_if_mysql" score="INFINITY" > to="mysql_orb1"/> This constraint looks strange. Not sure how should the cluster behave, because you have two clones and the web group is bound to both. Thanks, Dejan > </constraints> > </configuration> > > > > Franck Ganachaud a ?crit : >> Thanks for the tips both of you. >> >> I'm going to work on that the next few days. >> >> Andrew Beekhof a ?crit : >>> >>> On Dec 5, 2007, at 1:44 PM, Dejan Muhamedagic wrote: >>> >>>> Hi, >>>> >>>> On Wed, Dec 05, 2007 at 11:51:20AM +0100, Franck Ganachaud wrote: >>>>> Well I don't want hearbeat to stop or start mysql. >>>> >>>> You should be better off if you do. Otherwise, you'll probably >>>> end up with an unmaintainable and complex configuration. >>>> >>>>> And if I colocate R1 with mysql, will a R1 move from server A to B >>>>> implie mysql move also? >>>> >>>> Not necessarily. Colocations don't have to be symmetrical. >>> >>> particularly not if its a clone (ie. a resource that has a copy running >>> on each node) >>> >>> if you really object to having heartbeat manage mysql, use >>> is_managed=false for just the mysql resource. >>> with this setting, the cluster will never ever modify the state of your >>> resource (only check it's health and make decisions for R1 based on that) >>> >>>> >>>> >>>> Thanks, >>>> >>>> Dejan >>>> >>>>> Franck. >>>>> >>>>> Andrew Beekhof a ?crit : >>>>>> >>>>>> On Dec 5, 2007, at 9:42 AM, Franck Ganachaud wrote: >>>>>> >>>>>>> I try to be more explicit. >>>>>>> I use v2 configuration >>>>>>> >>>>>>> I got 2 servers A and B. A set (group) of resource R1. >>>>>>> R1 is running by default on server A and jump to server B in case of >>>>>>> trouble. >>>>>>> This is currently setup and working fine. >>>>>>> >>>>>>> Now, i need to check mysql running on both server A and B. >>>>>>> >>>>>>> If mysql isn't ok on server running R1, I need to stop >>>>>>> whateverdaemond and move R1 to the other server >>>>>>> If mysql isn't ok on server not running R1, I just need to stop >>>>>>> whateverdaemond. >>>>>>> >>>>>>> I made a OCF agent that does the "stop whateverdaemond if mysql is >>>>>>> down" job >>>>>> >>>>>> dont do that >>>>>> write a proper agent for whateverdaemond (maybe you can use an >>>>>> init-script instead) >>>>>> >>>>>> add a clone resource for mysql >>>>>> add a clone resource for whateverdaemond >>>>>> colocate whateverdaemond with mysql >>>>>> colocate R1 with mysql >>>>>> >>>>>>> >>>>>>> This where I get puzzled, how to translate the relation between the >>>>>>> "if running R1" and the agent I wrote it in heartbeat groups and >>>>>>> constraints >>>>>>> >>>>>>> Hope it's clearer. >>>>>>> Franck. >>>>>>> >>>>>>> Dejan Muhamedagic a ?crit : >>>>>>>> Hi, >>>>>>>> >>>>>>>> On Tue, Dec 04, 2007 at 05:25:08PM +0100, Franck Ganachaud wrote: >>>>>>>> >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> I have a two 2 node cluster. >>>>>>>>> group1 is active/passive set of ressource. >>>>>>>>> group1 is currently hop-ing from one node to the other as intended >>>>>>>>> when one of the group ressource isn't available. >>>>>>>>> >>>>>>>>> Now I have a second task for heartbeat. >>>>>>>>> On each node, I have a mysql server, if it goes wrong, I must >>>>>>>>> shutdown a service and migrate the group1 (if it runs on this node) >>>>>>>>> to the other node. >>>>>>>>> And this something I don't know how to do. >>>>>>>>> >>>>>>>>> I was thinking about creating a group2 in an active/active >>>>>>>>> configuration testing mysql and if goes wrong just shutdown the >>>>>>>>> service in the stop process of the group2 but I don't know how to >>>>>>>>> force in this case the push of the group1 to the other node of it's >>>>>>>>> running on the group2 failing node. >>>>>>>>> >>>>>>>> >>>>>>>> Sorry, but I can't follow. Can you please rephrase. >>>>>>>> >>>>>>>> Which configuration do you use: v1 or v2? >>>>>>>> >>>>>>>> Thanks, >>>>>>>> >>>>>>>> Dejan >>>>>>>> >>>>>>>> >>>>>>>>> Any one can help me? >>>>>>>>> Hope it's clear. >>>>>>>>> >>>>>>>>> Franck. >>>>>>>>> _______________________________________________ >>>>>>>>> Linux-HA mailing list >>>>>>>>> [email protected] >>>>>>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >>>>>>>>> See also: http://linux-ha.org/ReportingProblems >>>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Linux-HA mailing list >>>>>>>> [email protected] >>>>>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >>>>>>>> See also: http://linux-ha.org/ReportingProblems >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> -- >>>>>>> GANACHAUD Franck >>>>>>> Consultant >>>>>>> Tel. : +33 (0)2 98 05 43 21 >>>>>>> http://www.altran.com >>>>>>> -- >>>>>>> Technop?le Brest Iroise >>>>>>> Site du Vernis - CS 23866 >>>>>>> 29238 Brest Cedex 3 - France >>>>>>> Tel. : +33 (0)2 98 05 43 21 >>>>>>> Fax. : +33 (0)2 98 05 20 34 >>>>>>> e-mail: [EMAIL PROTECTED] >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Linux-HA mailing list >>>>>>> [email protected] >>>>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >>>>>>> See also: http://linux-ha.org/ReportingProblems >>>>>> >>>>>> _______________________________________________ >>>>>> Linux-HA mailing list >>>>>> [email protected] >>>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >>>>>> See also: http://linux-ha.org/ReportingProblems >>>>>> >>>>> >>>>> -- >>>>> GANACHAUD Franck >>>>> Consultant >>>>> Tel. : +33 (0)2 98 05 43 21 >>>>> http://www.altran.com >>>>> -- >>>>> Technop?le Brest Iroise >>>>> Site du Vernis - CS 23866 >>>>> 29238 Brest Cedex 3 - France >>>>> Tel. : +33 (0)2 98 05 43 21 >>>>> Fax. : +33 (0)2 98 05 20 34 >>>>> e-mail: [EMAIL PROTECTED] >>>>> >>>>> _______________________________________________ >>>>> Linux-HA mailing list >>>>> [email protected] >>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >>>>> See also: http://linux-ha.org/ReportingProblems >>>> _______________________________________________ >>>> Linux-HA mailing list >>>> [email protected] >>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >>>> See also: http://linux-ha.org/ReportingProblems >>> >>> _______________________________________________ >>> Linux-HA mailing list >>> [email protected] >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >>> See also: http://linux-ha.org/ReportingProblems >>> >> > > -- > GANACHAUD Franck > Consultant > Tel. : +33 (0)2 98 05 43 21 > http://www.altran.com > -- > Technop?le Brest Iroise > Site du Vernis - CS 23866 > 29238 Brest Cedex 3 - France > Tel. : +33 (0)2 98 05 43 21 > Fax. : +33 (0)2 98 05 20 34 > e-mail: [EMAIL PROTECTED] > > _______________________________________________ > Linux-HA mailing list > [email protected] > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
