Hi,

On Mon, Dec 10, 2007 at 04:27:37PM +0100, Franck Ganachaud wrote:
> Ok, here is where I am now.
>
> I built a clone set and colocated the group1 with the clone.
> But when I stop mysql, the clone monitor operation (mysql_orb) return that 
> DB is down but group1 isn't relocated to serveur B.
> 
> Anyone has a clue?

Can you post logs? If you have hb_report you can use that.

> I copy the cib :
>
> <configuration>
>    <crm_config>
>      <cluster_property_set id="cib-bootstrap-options">
>        <attributes>
>          <nvpair id="cib-bootstrap-options-symmetric-cluster" 
> name="symmetric-cluster" value="true"/>
>          <nvpair id="cib-bootstrap-options-no_quorum-policy" 
> name="no_quorum-policy" value="stop"/>
>          <nvpair id="cib-bootstrap-options-default-resource-stickiness" 
> name="default-resource-stickiness" value="0"/>
>          <nvpair 
> id="cib-bootstrap-options-default-resource-failure-stickiness" 
> name="default-resource-failure-stickiness" value="0"/>
>          <nvpair id="cib-bootstrap-options-stonith-enabled" 
> name="stonith-enabled" value="false"/>
>          <nvpair id="cib-bootstrap-options-stonith-action" 
> name="stonith-action" value="reboot"/>
>          <nvpair id="cib-bootstrap-options-stop-orphan-resources" 
> name="stop-orphan-resources" value="true"/>
>          <nvpair id="cib-bootstrap-options-stop-orphan-actions" 
> name="stop-orphan-actions" value="true"/>
>          <nvpair id="cib-bootstrap-options-remove-after-stop" 
> name="remove-after-stop" value="false"/>
>          <nvpair id="cib-bootstrap-options-short-resource-names" 
> name="short-resource-names" value="true"/>
>          <nvpair id="cib-bootstrap-options-transition-idle-timeout" 
> name="transition-idle-timeout" value="5min"/>
>          <nvpair id="cib-bootstrap-options-default-action-timeout" 
> name="default-action-timeout" value="5s"/>
>          <nvpair id="cib-bootstrap-options-is-managed-default" 
> name="is-managed-default" value="true"/>
>        </attributes>
>      </cluster_property_set>
>    </crm_config>
>    <nodes>
>      <node id="0e8b2fa4-983b-4e56-a4a5-72dbb2aeaeec" type="normal" 
> uname="server_a">
>        <instance_attributes 
> id="nodes-0e8b2fa4-983b-4e56-a4a5-72dbb2aeaeec">
>          <attributes>
>            <nvpair id="standby-0e8b2fa4-983b-4e56-a4a5-72dbb2aeaeec" 
> name="standby" value="off"/>
>          </attributes>
>        </instance_attributes>
>      </node>
>      <node id="5cdd04e8-035a-44cf-ab60-3065840109db" type="normal" 
> uname="server_b">
>        <instance_attributes 
> id="nodes-5cdd04e8-035a-44cf-ab60-3065840109db">
>          <attributes>
>            <nvpair id="standby-5cdd04e8-035a-44cf-ab60-3065840109db" 
> name="standby" value="off"/>
>          </attributes>
>        </instance_attributes>
>      </node>
>      <node id="9e05d57a-ae9c-430d-a210-d03b9f37739e" type="normal" 
> uname="server_b"/>
>      <node id="352f29b5-f0ed-4866-a839-71dbdbfd491d" type="normal" 
> uname="server_a"/>
>    </nodes>

Nodes are listed twice.

>    <resources>
>      <clone id="MySQL_ORB" interleave="false" is_managed="true" 
> notify="false" ordered="false">
>        <instance_attributes id="MySQL_ORB_inst_attr">
>          <attributes>
>            <nvpair id="MySQL_ORB_attr_0" name="clone_max" value="2"/>
>            <nvpair id="MySQL_ORB_attr_1" name="clone_node_max" value="1"/>
>          </attributes>
>        </instance_attributes>
>        <primitive class="ocf" id="mysql_orb1" is_managed="false" 
> provider="heartbeat" type="mysql_orb">
>          <operations>
>            <op id="mysql_orb_mon" interval="30s" name="monitor" 
> on_fail="stop" timeout="30s"/>
>          </operations>
>        </primitive>
>      </clone>
>      <group id="group_1" restart_type="restart">
>        <primitive class="ocf" id="IPaddr_Cluster" provider="heartbeat" 
> type="IPaddr">
>          <operations>
>            <op id="IPaddr_Cluster_mon" interval="5s" name="monitor" 
> timeout="5s"/>
>          </operations>
>          <instance_attributes id="IPaddr_Cluster_inst_attr">
>            <attributes>
>              <nvpair id="IPaddr_Cluster_attr_0" name="ip" 
> value="192.168.87.100"/>
>            </attributes>
>          </instance_attributes>
>        </primitive>
>       <primitive class="ocf" id="apache_2" provider="heartbeat" 
> type="apache">
>          <operations>
>            <op id="apache_2_mon" interval="30s" name="monitor" 
> timeout="30s"/>
>          </operations>
>          <instance_attributes id="apache_2_inst_attr">
>            <attributes>
>              <nvpair id="apache_2_attr_0" name="configfile" 
> value="/usr/local/apache/conf/httpd.conf"/>
>            </attributes>
>          </instance_attributes>
>        </primitive>
>      </group>
>    </resources>
>    <constraints>
>      <rsc_location id="rsc_location_group_1" rsc="group_1">
>        <rule id="prefered_location_group_1" score="100">
>          <expression attribute="#uname" id="prefered_location_group_1_expr" 
> operation="eq" value="server_a"/>
>        </rule>
>      </rsc_location>
>      <rsc_colocation from="group_1" id="web_if_mysql" score="INFINITY" 
> to="mysql_orb1"/>

This constraint looks strange. Not sure how should the cluster
behave, because you have two clones and the web group is bound to
both.

Thanks,

Dejan

>    </constraints>
>  </configuration>
>
>
>
> Franck Ganachaud a ?crit :
>> Thanks for the tips both of you.
>>
>> I'm going to work on that the next few days.
>>
>> Andrew Beekhof a ?crit :
>>>
>>> On Dec 5, 2007, at 1:44 PM, Dejan Muhamedagic wrote:
>>>
>>>> Hi,
>>>>
>>>> On Wed, Dec 05, 2007 at 11:51:20AM +0100, Franck Ganachaud wrote:
>>>>> Well I don't want hearbeat to stop or start mysql.
>>>>
>>>> You should be better off if you do. Otherwise, you'll probably
>>>> end up with an unmaintainable and complex configuration.
>>>>
>>>>> And if I colocate R1 with mysql, will a R1 move from server A to B
>>>>> implie mysql move also?
>>>>
>>>> Not necessarily. Colocations don't have to be symmetrical.
>>>
>>> particularly not if its a clone (ie. a resource that has a copy running 
>>> on each node)
>>>
>>> if you really object to having heartbeat manage mysql, use 
>>> is_managed=false for just the mysql resource.
>>> with this setting, the cluster will never ever modify the state of your 
>>> resource (only check it's health and make decisions for R1 based on that)
>>>
>>>>
>>>>
>>>> Thanks,
>>>>
>>>> Dejan
>>>>
>>>>> Franck.
>>>>>
>>>>> Andrew Beekhof a ?crit :
>>>>>>
>>>>>> On Dec 5, 2007, at 9:42 AM, Franck Ganachaud wrote:
>>>>>>
>>>>>>> I try to be more explicit.
>>>>>>> I use v2 configuration
>>>>>>>
>>>>>>> I got 2 servers A and B. A set (group) of resource R1.
>>>>>>> R1 is running by default on server A and jump to server B in case of
>>>>>>> trouble.
>>>>>>> This is currently setup and working fine.
>>>>>>>
>>>>>>> Now, i need to check mysql running on both server A and B.
>>>>>>>
>>>>>>> If mysql isn't ok on server running R1, I need to stop
>>>>>>> whateverdaemond and move R1 to the other server
>>>>>>> If mysql isn't ok on server not running R1, I just need to stop
>>>>>>> whateverdaemond.
>>>>>>>
>>>>>>> I made a OCF agent that does the "stop whateverdaemond if mysql is
>>>>>>> down" job
>>>>>>
>>>>>> dont do that
>>>>>> write a proper agent for whateverdaemond (maybe you can use an
>>>>>> init-script instead)
>>>>>>
>>>>>> add a clone resource for mysql
>>>>>> add a clone resource for whateverdaemond
>>>>>> colocate whateverdaemond with mysql
>>>>>> colocate R1 with mysql
>>>>>>
>>>>>>>
>>>>>>> This where I get puzzled, how to translate the relation between the
>>>>>>> "if running R1"  and the agent I wrote it in heartbeat groups and
>>>>>>> constraints
>>>>>>>
>>>>>>> Hope it's clearer.
>>>>>>> Franck.
>>>>>>>
>>>>>>> Dejan Muhamedagic a ?crit :
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> On Tue, Dec 04, 2007 at 05:25:08PM +0100, Franck Ganachaud wrote:
>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>> I have a two 2 node cluster.
>>>>>>>>> group1 is active/passive set of ressource.
>>>>>>>>> group1 is currently hop-ing  from one node to the other as intended
>>>>>>>>> when one of the group ressource isn't available.
>>>>>>>>>
>>>>>>>>> Now I have a second task for heartbeat.
>>>>>>>>> On each node, I have a mysql server, if it goes wrong, I must
>>>>>>>>> shutdown a service and migrate the group1 (if it runs on this node)
>>>>>>>>> to the other node.
>>>>>>>>> And this something I don't know how to do.
>>>>>>>>>
>>>>>>>>> I was thinking about creating a group2 in an active/active
>>>>>>>>> configuration testing mysql and if goes wrong just shutdown the
>>>>>>>>> service in the stop process of the group2 but I don't know how to
>>>>>>>>> force in this case the push of the group1 to the other node of it's
>>>>>>>>> running on the group2 failing node.
>>>>>>>>>
>>>>>>>>
>>>>>>>> Sorry, but I can't follow. Can you please rephrase.
>>>>>>>>
>>>>>>>> Which configuration do you use: v1 or v2?
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>>
>>>>>>>> Dejan
>>>>>>>>
>>>>>>>>
>>>>>>>>> Any one can help me?
>>>>>>>>> Hope it's clear.
>>>>>>>>>
>>>>>>>>> Franck.
>>>>>>>>> _______________________________________________
>>>>>>>>> Linux-HA mailing list
>>>>>>>>> [email protected]
>>>>>>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>>>>>>>> See also: http://linux-ha.org/ReportingProblems
>>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Linux-HA mailing list
>>>>>>>> [email protected]
>>>>>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>>>>>>> See also: http://linux-ha.org/ReportingProblems
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> -- 
>>>>>>> GANACHAUD Franck
>>>>>>> Consultant
>>>>>>> Tel. : +33 (0)2 98 05 43 21
>>>>>>> http://www.altran.com
>>>>>>> -- 
>>>>>>> Technop?le Brest Iroise
>>>>>>> Site du Vernis - CS 23866
>>>>>>> 29238 Brest Cedex 3 - France
>>>>>>> Tel. : +33 (0)2 98 05 43 21
>>>>>>> Fax. : +33 (0)2 98 05 20 34
>>>>>>> e-mail: [EMAIL PROTECTED]
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Linux-HA mailing list
>>>>>>> [email protected]
>>>>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>>>>>> See also: http://linux-ha.org/ReportingProblems
>>>>>>
>>>>>> _______________________________________________
>>>>>> Linux-HA mailing list
>>>>>> [email protected]
>>>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>>>>> See also: http://linux-ha.org/ReportingProblems
>>>>>>
>>>>>
>>>>> -- 
>>>>> GANACHAUD Franck
>>>>> Consultant
>>>>> Tel. : +33 (0)2 98 05 43 21
>>>>> http://www.altran.com
>>>>> -- 
>>>>> Technop?le Brest Iroise
>>>>> Site du Vernis - CS 23866
>>>>> 29238 Brest Cedex 3 - France
>>>>> Tel. : +33 (0)2 98 05 43 21
>>>>> Fax. : +33 (0)2 98 05 20 34
>>>>> e-mail: [EMAIL PROTECTED]
>>>>>
>>>>> _______________________________________________
>>>>> Linux-HA mailing list
>>>>> [email protected]
>>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>>>> See also: http://linux-ha.org/ReportingProblems
>>>> _______________________________________________
>>>> Linux-HA mailing list
>>>> [email protected]
>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>>> See also: http://linux-ha.org/ReportingProblems
>>>
>>> _______________________________________________
>>> Linux-HA mailing list
>>> [email protected]
>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> See also: http://linux-ha.org/ReportingProblems
>>>
>>
>
> -- 
> GANACHAUD Franck
> Consultant
> Tel. : +33 (0)2 98 05 43 21
> http://www.altran.com
> --
> Technop?le Brest Iroise
> Site du Vernis - CS 23866
> 29238 Brest Cedex 3 - France
> Tel. : +33 (0)2 98 05 43 21
> Fax. : +33 (0)2 98 05 20 34
> e-mail: [EMAIL PROTECTED]
>
> _______________________________________________
> Linux-HA mailing list
> [email protected]
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to