Yes, I guess I have to do it manually then.

Thanks Dejan

Dejan Muhamedagic a écrit :
Hi,

On Wed, Feb 20, 2008 at 03:10:11PM +0100, Franck Ganachaud wrote:
Yes I mean by hand.

Mysql can't be managed by heartbeat for various reasons.
The fact is the agent mysql_orb is supposed to monitor mysql, if mysql fails, then another group of resources is moved away.
This is working currently.

You run mysql but it's not controlled by the cluster just
monitored. Interesting setup.

The problem is when mysql is back again, as mysql_orb monitoring doesn't occur on the "failed" node, it doesn't detect mysql status change and resource don't come back.

I know I can do it by hand using crm_resource, that's my current solution but that's a mechanism I'd like to be automated.

Hmm. That can't be done, I'm afraid. Currently, CRM doesn't check
for resources which are down. Anyway, if you already have to fix
mysql by hand, then you could also add the crm_resource
invocation, right?

Thanks,

Dejan

Thanks for answering

Dejan Muhamedagic a ?crit :
Hi,

On Wed, Feb 20, 2008 at 11:32:52AM +0100, Franck Ganachaud wrote:
Ok, I found what is the problem, when mysql_orb fails, it restarts and stays in state FAILED. I changed the monitor operation with the on_fail="stop" then it works.

But here is my problem now.
When the action is triggered due to monitoring mysql_orb fails, monitor of this cloned resource stops. we restart mysqld
I guess that you mean restart by hand?

but mysql_orb agent isn't monitored anymore and resources don't come back to this node. What do I have to change to the cloned mysql_orb resource to keep monitoring after a failure.
Clean the resource status:

crm_resource -C -r rsc

Actually, if you do this, then the cluster will try to start the
resource itself, so you could save yourself some typing.
Normally, you should let the cluster do start/stop and just take
care of whatever conditions caused a resource to fail. All you
have to do then is to signal the cluster that it should reprobe
the resource by using the above command.

Thanks,

Dejan

<clone globally_unique="false" id="MySQL_ORB" interleave="false" is_managed="true" notify="false" ordered="false">
     <instance_attributes id="MySQL_ORB_inst_attr">
       <attributes>
         <nvpair id="MySQL_ORB_attr_0" name="clone_max" value="2"/>
         <nvpair id="MySQL_ORB_attr_1" name="clone_node_max" value="1"/>
       </attributes>
     </instance_attributes>
<primitive class="ocf" id="mysql_orb1" is_managed="true" provider="heartbeat" type="mysql_orb">
       <operations>
<op id="mysql_orb_mon" interval="31s" name="monitor" on_fail="stop" timeout="30s"/>
       </operations>
     </primitive>
   </clone>

Franck Ganachaud a ?crit :
Well, On a 2 nodes cluster using heartbeat 2.1.2, I have one cloned ressource and 1 group executed preferably on nodeA but if clone if failed or off, group should be migrated to nodeB

When I stop the service, cloned ressource mysql_orb fails on nodeA but I don't know what to do to make the group_1 goes to nodeB. I don't know what to put in the configuration file to make it work, the following don't work : <rsc_colocation from="group_1" id="web_if_mysql" score="INFINITY" to="MySQL_ORB"/>

You can find cib.xml attached and log messages.

Thanks,
Franck.

------------------------------------------------------------------------

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to