Serge Dubrouski schrieb: > Lino, were you able to make it work? Well, I was... But I don't know how :-). I always received that same errormessages until I shutdown both nodes and started my secondary. Then the Clone was able to get running on my 2nd. As I powered on again the first node the clone there also was able to run and he switchted over my resources to my primary.
I will later try to shutdown both nodes again and first start my 1st node to see if the problem still exists. I suspect it to do something with the stonithd not able to contact the lrmd but i'm not sure. As for now your stonih RA works fine, thx again :-). Only thing, I wasn't able yet to configuer my crm_config with the stonith-action so that the node isn't resetted but shut off completely. When I try the value off and do a crm_verfiy it shoes the value="off" is unknown. But I think this will be a subject for another thread. Thanks again Lino > > On Mon, Mar 3, 2008 at 11:28 AM, Serge Dubrouski <[EMAIL PROTECTED]> wrote: >> On Mon, Mar 3, 2008 at 12:08 PM, Lino Moragon <[EMAIL PROTECTED]> wrote: >> > Serge Dubrouski wrote: >> > > Configuration looks right to me, I even tested it and it worked fine >> > > on my test cluster. So hints are obvious: >> > > >> > > 1. Check that you really put that script on a second node and made it >> > > executable. >> > > >> > That was my first error, but i noticed a error message in the logfile >> and corrected it. >> > So i can exclude this posibility >> > >> > >> > > 2. Nodes should be able to ping each other. That programmed in a >> > > "status" function. >> > > >> > What do you mean by "programmed in a status function?" From each node I >> can ping the other one. It also works fine with name resolution. >> >> I mean that "status" function in script pings each node from hostlist. >> If it can't ping it it fails with exit code 1. >> >> > Do you think, it could possibly be an issue with my current version >> (2.1.2-3)? >> >> No it can't because I use the same version. >> >>> With which version did you try the configurations? >> > when you start the Clone the first time, then both resources should >> start (on node1 as well as on node2) is that correct? >> >> Right. >> >> >> > Could it be another configuration error outside the clone_id section in >> the cib? Could you perhaps attach your whole CIB? >> >> Attached. >> >> >> >> > Thanks for your support so far, I'm very grateful. >> > Lino >> > >> > >> > >> > > >> > > >> > > On Mon, Mar 3, 2008 at 9:16 AM, Lino Moragon <[EMAIL PROTECTED]> wrote: >> > > >> > >> Hi, >> > >> >> > >> I'm using now the most actual xen0 stonith plugin, that Serge >> attachted >> > >> the 2008-02-28 to this thread. >> > >> I thought I configured everything correct but it seems that the >> stonith >> > >> clone cannot be started on my 2nd node. >> > >> I must admit I configured the Clone via hb_gui but I still have some >> issues. >> > >> As for these reasons the stonith plugin only works to reset my 2nd >> node >> > >> and not vice versa. >> > >> The version I am using is: heartbeat-2.1.2-3 of the Centos5.1 >> repository. >> > >> >> > >> Has anyone a clue where the failure could be? >> > >> >> > >> To this email I attach the cib.xml >> > >> Below follows the cib stonith section, crm_verfy and error msgs. >> > >> >> > >> Perhaps Serge, you could give me a hint? What did you do differently >> > >> than me? >> > >> >> > >> I would be very glad for any hints and clues. >> > >> >> > >> Thanks in advance, >> > >> Lino >> > >> >> > >> node1: mysql1 >> > >> node2: mysql2 >> > >> xen-host: simulator >> > >> >> > >> <clone id="DoFencing"> >> > >> <instance_attributes id="DoFencing_instance_attrs"> >> > >> <attributes> >> > >> <nvpair id="DoFencing_clone_max" name="clone_max" >> value="2"/> >> > >> <nvpair id="DoFencing_clone_node_max" >> name="clone_node_max" >> > >> value="1"/> >> > >> <nvpair id="DoFencing_target_role" name="target_role" >> > >> value="started"/> >> > >> </attributes> >> > >> </instance_attributes> >> > >> <primitive class="stonith" type="external/xen0" >> > >> provider="heartbeat" id="child_DoFencing"> >> > >> <instance_attributes id="child_DoFencing_instance_attrs"> >> > >> <attributes> >> > >> <nvpair name="target_role" >> > >> id="child_DoFencing_target_role" value="started"/> >> > >> <nvpair name="hostlist" >> > >> id="5525c381-5956-4564-af3d-2bc7b547812a" value="mysql1:mysql1.cfg >> > >> mysql2:mysql2.cfg"/> >> > >> <nvpair id="65feeaf5-501f-4648-a155-83b79b587fbf" >> > >> name="dom0" value="simulator"/> >> > >> </attributes> >> > >> </instance_attributes> >> > >> </primitive> >> > >> </clone> >> > >> >> > >> If I use crm_verify I get the following results: >> > >> ============================================================= >> > >> crm_verify[6169]: 2008/03/03_17:06:55 WARN: unpack_rsc_op: Processing >> > >> failed op (child_DoFencing:0_start_0) on mysql2 >> > >> crm_verify[6169]: 2008/03/03_17:06:55 WARN: unpack_rsc_op: Handling >> > >> failed start for child_DoFencing:0 on mysql2 >> > >> crm_verify[6169]: 2008/03/03_17:06:55 WARN: unpack_rsc_op: Processing >> > >> failed op (child_DoFencing:1_start_0) on mysql2 >> > >> crm_verify[6169]: 2008/03/03_17:06:55 WARN: unpack_rsc_op: Handling >> > >> failed start for child_DoFencing:1 on mysql2 >> > >> ============================================================= >> > >> >> > >> Furthermore I get the following Errors in my log: >> > >> ============================================================= >> > >> Mar 3 16:29:42 mysql2 crmd: [1478]: ERROR: process_lrm_event: LRM >> > >> operation child_DoFencing:0_start_0 (call=22, rc=1) Error unknown >> error >> > >> Mar 3 16:29:46 mysql2 crmd: [1478]: ERROR: process_lrm_event: LRM >> > >> operation child_DoFencing:1_start_0 (call=24, rc=1) Error unknown >> error >> > >> Mar 3 16:35:33 mysql2 crmd: [1478]: ERROR: process_lrm_event: LRM >> > >> operation child_DoFencing:1_start_0 (call=28, rc=1) Error unknown >> error >> > >> Mar 3 16:46:40 mysql2 crmd: [1477]: ERROR: process_lrm_event: LRM >> > >> operation child_DoFencing:0_start_0 (call=10, rc=1) Error unknown >> error >> > >> Mar 3 16:46:45 mysql2 crmd: [1477]: ERROR: process_lrm_event: LRM >> > >> operation child_DoFencing:1_start_0 (call=12, rc=1) Error unknown >> error >> > >> ============================================================== >> > >> >> > >> >> > >> >> > >> >> > >> >> > >> Serge Dubrouski schrieb: >> > >> > Attached. >> > >> > >> > >> > On Thu, Feb 28, 2008 at 3:35 AM, Dejan Muhamedagic <[EMAIL >> PROTECTED]> wrote: >> > >> >> Hi Serge, >> > >> >> >> > >> >> >> > >> >> On Tue, Feb 26, 2008 at 09:46:14AM -0700, Serge Dubrouski wrote: >> > >> >> > Dejan - >> > >> >> > >> > >> >> > I found a compromise :-) Attached is a version of that plugin >> that >> > >> >> > supports following parameters: >> > >> >> > >> > >> >> > 1. hostlist. A string that has a list of controlled nodes >> separated by >> > >> >> > space or commas. A required parameter. In a simple form its >> just a >> > >> >> > list of nodes. If one needs to use non-standard Xen >> configuration he >> > >> >> > can use extended form of this parameter: >> "node1_name:config1_file >> > >> >> > node2_name:config2_file". If config file isn't given it >> defaults to >> > >> >> > /etc/xen/node_name.cfg >> > >> >> > >> > >> >> > 2. Dom0. Name of Dom0 Xen node. A required parameter. >> > >> >> > >> > >> >> > 3. ssh_command. SSH command that is used to ssh from DomU to >> Dom0. >> > >> >> > Defaults to "/usr/bin/ssh -q -x -n -l root". If one wants to >> use SSH >> > >> >> > keys for higher security he needs to use this parameter. >> > >> >> >> > >> >> This is not necessary. One can setup everything needed in >> > >> >> ~/.ssh/config on a per host basis, i.e. key or user to connect >> > >> >> with. Since the plugin always runs as root, you might leave out >> > >> >> '-l root' as well. >> > >> >> >> > >> >> >> > >> >> > So in this form this plugin can be configured as a clone or as >> a set >> > >> >> > of resources and location constraints. >> > >> >> > >> > >> >> > >> > >> >> > I'd be very pleased if this plugin gets its way into Linux-HA >> distribution. >> > >> >> > >> > >> >> >> > >> >> Sure. Could you please just drop the ssh_command parameter. >> > >> >> >> > >> >> Many thanks for the contribution. >> > >> >> >> > >> >> Cheers, >> > >> >> >> > >> >> Dejan >> > >> >> >> > >> >> >> > >> >> >> > >> >> > >> > >> >> > On Tue, Feb 26, 2008 at 8:45 AM, Serge Dubrouski <[EMAIL >> PROTECTED]> wrote: >> > >> >> > > >> > >> >> > > On Mon, Feb 25, 2008 at 4:02 PM, Dejan Muhamedagic <[EMAIL >> PROTECTED]> wrote: >> > >> >> > > > Hi, >> > >> >> > > > >> > >> >> > > > >> > >> >> > > > On Mon, Feb 25, 2008 at 12:17:40PM -0700, Serge >> Dubrouski wrote: >> > >> >> > > > > On Mon, Feb 25, 2008 at 12:10 PM, Dejan Muhamedagic >> <[EMAIL PROTECTED]> wrote: >> > >> >> > > > > > Hi, >> > >> >> > > > > > >> > >> >> > > > > > >> > >> >> > > > > > On Mon, Feb 25, 2008 at 11:27:38AM -0700, Serge >> Dubrouski wrote: >> > >> >> > > > > > > I would love to do that and already tried it. >> Though we didn't come to >> > >> >> > > > > > > agreement on how configuration parameters should >> look like. >> > >> >> > > > > > >> > >> >> > > > > > Why? Was there a discussion on the list about it? >> The >> > >> >> > > > > > configuration is a bit unusual. Other stonith >> agents take named >> > >> >> > > > > > parameters. Though this kind of configuration also >> works, I'd >> > >> >> > > > > > prefer something similar to the others, e.g. >> > >> >> > > > > >> > >> >> > > > > Yes there was a discussion >> > >> >> > > > > >> http://lists.community.tummy.com/pipermail/linux-ha-dev/2007-February/ >> > >> >> > > > >> > >> >> > > > It's a long one and peters out inconclusively. >> > >> >> > > > >> > >> >> > > > >> > >> >> > > > > See "new stonith external plugin". The config >> parameter hostslist is >> > >> >> > > > > actually derived from original ssh plugin. I needed to >> have a full >> > >> >> > > > > list of all controlled nodes and preferred to have it >> as one >> > >> >> > > > > parameter. >> > >> >> > > > > >> > >> >> > > > > > >> > >> >> > > > > > hostname dom0 (or xenhost) config >> > >> >> > > > > > >> > >> >> > > > > >> > >> >> > > > > That would work if I needed just a dom0 host, but I >> also need a list >> > >> >> > > > > of controlled nodes and probably configuration files. >> > >> >> > > > >> > >> >> > > > That's why you can have several instances of a stonith >> resource >> > >> >> > > > (see e.g. external/ipmi). Each of them would run with >> different >> > >> >> > > > parameters. What I meant was: >> > >> >> > > > >> > >> >> > > > hostname: xen vm >> > >> >> > > > dom0: xen dom0 >> > >> >> > > > config: vm configuration file >> > >> >> > > > >> > >> >> > > >> > >> >> > > That's possible and easy to do but I'm not sure that it'll >> be better. >> > >> >> > > Current version allows to configure a clone. New version >> would require >> > >> >> > > configuring a separate resource for each node and creating >> location >> > >> >> > > constraints for each of them. Per my opinion that would be >> more >> > >> >> > > complex configuration. May be I'm missing something. >> > >> >> > > >> > >> >> > >> > >> >> > >> > >> >> > >> > >> >> > -- >> > >> >> > Serge Dubrouski. >> > >> >> >> > >> >> >> > >> >> >> > >> >>> _______________________________________________ >> > >> >> > Linux-HA mailing list >> > >> >> > [email protected] >> > >> >> > http://lists.linux-ha.org/mailman/listinfo/linux-ha >> > >> >> > See also: http://linux-ha.org/ReportingProblems >> > >> >> >> > >> >> -- >> > >> >> Dejan >> > >> >> >> > >> >> >> > >> >> _______________________________________________ >> > >> >> Linux-HA mailing list >> > >> >> [email protected] >> > >> >> http://lists.linux-ha.org/mailman/listinfo/linux-ha >> > >> >> See also: http://linux-ha.org/ReportingProblems >> > >> >> >> > >> > >> > >> > >> > >> > >> > >> > >> > >> > >> ------------------------------------------------------------------------ >> > >> > >> > >> > _______________________________________________ >> > >> > Linux-HA mailing list >> > >> > [email protected] >> > >> > http://lists.linux-ha.org/mailman/listinfo/linux-ha >> > >> > See also: http://linux-ha.org/ReportingProblems >> > >> >> > >> >> > >> >> > >> _______________________________________________ >> > >> Linux-HA mailing list >> > >> [email protected] >> > >> http://lists.linux-ha.org/mailman/listinfo/linux-ha >> > >> See also: http://linux-ha.org/ReportingProblems >> > >> >> > >> >> > > >> > > >> > > >> > > >> > >> > _______________________________________________ >> > Linux-HA mailing list >> > [email protected] >> > http://lists.linux-ha.org/mailman/listinfo/linux-ha >> > See also: http://linux-ha.org/ReportingProblems >> > >> >> >> >> -- >> Serge Dubrouski. >> > > > _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
