On Thu, Jul 3, 2008 at 09:59, Ehlers, Kolja <[EMAIL PROTECTED]> wrote: > thanks for the reply, still the problem remains.
Because you didn't follow his advice. > Failed actions: > apache_2_start_0 (node=www1test, call=6, rc=6): complete Your RA is still returning 6 (OCF_ERR_CONFIGURED) instead of 5 (OCF_ERR_INSTALLED) when the binary is missing. > If apache cannot be started/restarted it is not failed over to the second > node. I have two equal servers and I want to run the virtual ip + apache > (grouped) on either one of the nodes. To test the configuration I have > renamed httpd on the one node to httpd_ else I am not sure how to simulate a > non starting apache. But either way when heartbeat is started the apache > start is failed on www1test and nothing happens then. I have attached my CIB > and the logs > > This is what crm_mon gives me: > > Refresh in 1s... > > ============ > Last updated: Thu Jul 3 09:53:34 2008 > Current DC: www2test (5e0f97b7-6780-4487-baf9-6c36500b1276) > 2 Nodes configured. > 1 Resources configured. > ============ > > Node: www2test (5e0f97b7-6780-4487-baf9-6c36500b1276): online > Node: www1test (3a325e23-2184-46ed-9e88-42a11f28c2be): online > > Resource Group: group_1 > IPaddr_192_168_11_25 (ocf::heartbeat:IPaddr): Started > www1test > apache_2 (ocf::heartbeat:apache): Stopped > > Failed actions: > apache_2_start_0 (node=www1test, call=6, rc=6): complete > > > > www1test:~ # crm_verify -VVVVL > crm_verify[8124]: 2008/07/03_09:54:55 info: main: =#=#=#=#= Getting XML > =#=#=#=#= > crm_verify[8124]: 2008/07/03_09:54:55 info: main: Reading XML from: live > cluster > crm_verify[8124]: 2008/07/03_09:54:55 notice: main: Required feature set: 2.0 > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value 'false' for cluster option 'stonith-enabled' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value 'reboot' for cluster option 'stonith-action' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value '0' for cluster option 'default-resource-failure-stickiness' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value '60s' for cluster option 'cluster-delay' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value '30' for cluster option 'batch-limit' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value '20s' for cluster option 'default-action-timeout' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value 'true' for cluster option 'stop-orphan-resources' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value 'true' for cluster option 'stop-orphan-actions' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value 'false' for cluster option 'remove-after-stop' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value '-1' for cluster option 'pe-error-series-max' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value '-1' for cluster option 'pe-warn-series-max' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value '-1' for cluster option 'pe-input-series-max' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value 'true' for cluster option 'startup-fencing' > crm_verify[8124]: 2008/07/03_09:54:55 debug: cluster_option: Using default > value 'true' for cluster option 'start-failure-is-fatal' > crm_verify[8124]: 2008/07/03_09:54:55 debug: unpack_config: Default action > timeout: 20s > crm_verify[8124]: 2008/07/03_09:54:55 debug: unpack_config: Default > stickiness: 1000000 > crm_verify[8124]: 2008/07/03_09:54:55 debug: unpack_config: Default failure > stickiness: 0 > crm_verify[8124]: 2008/07/03_09:54:55 debug: unpack_config: STONITH of failed > nodes is disabled > crm_verify[8124]: 2008/07/03_09:54:55 debug: unpack_config: Cluster is > symmetric - resources can run anywhere by default > crm_verify[8124]: 2008/07/03_09:54:55 debug: unpack_config: On loss of CCM > Quorum: Stop ALL resources > crm_verify[8124]: 2008/07/03_09:54:55 info: determine_online_status: Node > www2test is online > crm_verify[8124]: 2008/07/03_09:54:55 info: determine_online_status: Node > www1test is online > crm_verify[8124]: 2008/07/03_09:54:55 debug: common_apply_stickiness: > fail-count-apache_2: INFINITY > crm_verify[8124]: 2008/07/03_09:54:55 ERROR: unpack_rsc_op: Hard error: > apache_2_start_0 failed with rc=6. > crm_verify[8124]: 2008/07/03_09:54:55 ERROR: unpack_rsc_op: Preventing > apache_2 from re-starting anywhere in the cluster > crm_verify[8124]: 2008/07/03_09:54:55 WARN: unpack_rsc_op: Processing failed > op apache_2_start_0 on www1test: Error > crm_verify[8124]: 2008/07/03_09:54:55 WARN: unpack_rsc_op: Compatability > handling for failed op apache_2_start_0 on www1test > crm_verify[8124]: 2008/07/03_09:54:55 notice: group_print: Resource Group: > group_1 > crm_verify[8124]: 2008/07/03_09:54:55 notice: native_print: > IPaddr_192_168_11_25 (ocf::heartbeat:IPaddr): Started www1test > crm_verify[8124]: 2008/07/03_09:54:55 notice: native_print: apache_2 > (ocf::heartbeat:apache): Stopped > crm_verify[8124]: 2008/07/03_09:54:55 debug: group_rsc_location: Processing > rsc_location pref_run_apache_group for group_1 > crm_verify[8124]: 2008/07/03_09:54:55 debug: native_merge_weights: > IPaddr_192_168_11_25: Rolling back scores from apache_2 > crm_verify[8124]: 2008/07/03_09:54:55 debug: native_assign_node: Assigning > www1test to IPaddr_192_168_11_25 > crm_verify[8124]: 2008/07/03_09:54:55 debug: native_assign_node: All nodes > for resource apache_2 are unavailable, unclean or shutting down > crm_verify[8124]: 2008/07/03_09:54:55 WARN: native_color: Resource apache_2 > cannot run anywhere > crm_verify[8124]: 2008/07/03_09:54:55 notice: NoRoleChange: Leave resource > IPaddr_192_168_11_25 (www1test) > Warnings found during check: config may not be valid > crm_verify[8124]: 2008/07/03_09:54:55 debug: cib_native_signoff: Signing out > of the CIB Service > > > > -----Ursprüngliche Nachricht----- > Von: [EMAIL PROTECTED] > [mailto:[EMAIL PROTECTED] Auftrag von Dominik Klein > Gesendet: Donnerstag, 3. Juli 2008 08:27 > An: General Linux-HA mailing list > Betreff: Re: [Linux-HA] Apache failover / renaming the binary > > > http://hg.linux-ha.org/dev/file/5072025b79b8/resources/OCF/apache > > lines 516-518 > > another example of how to use exits codes incorrectly. > > I'll commit a patch soon. > > In your script: Make line 518 look like this (on all nodes!): > exit $OCF_ERR_INSTALLED > > Then cleanup the resource or start the cluster from scratch and try > again. Should fix it. > > Regards > Dominik > > > Ehlers, Kolja wrote: >> Hello, >> >> my simple active/passive cluster seems to work but when running and I do: >> >> /opt/apache2/bin/apachectl stop && mv /opt/apache2/bin/httpd >> /opt/apache2/bin/httpd_ >> >> Heartbeat is not failing over apache to node2 (Hard error: apache_2_start_0 >> failed with rc=6.) This is really odd because the log states "All 2 cluster >> nodes are eligible to run resources." but then 4 lines further it says >> "ERROR: unpack_rsc_op: Preventing apache_2 from re-starting anywhere in >> the cluster". I am using a very simple CIB with one virtual ip and apache >> grouped. If i stop apache manually heartbeat does restart apache fine. By >> the way can I configure it so that it does failover right to the other node >> if apache is stopped or fails? When manually stopping heartbeat the failover >> does work. >> >> So I am not sure which part of my configuration or logs you need to see. I >> guess im missing something important here. >> >> This is my cib >> >> <cib admin_epoch="0" generated="true" have_quorum="true" ignore_dtd="false" >> num_peers="2" cib_feature_revision="2.0" crm_feature_set="2.0" epoch="38" >> num_updates="3" cib-last-written="Wed Jul 2 16:16:51 2008" >> ccm_transition="2" dc_uuid="5e0f97b7-6780-4487-baf9-6c36500b1276"> >> <configuration> >> <crm_config> >> <cluster_property_set id="cib-bootstrap-options"> >> <attributes> >> <nvpair id="cib-bootstrap-options-symmetric-cluster" >> name="symmetric-cluster" value="true"/> >> <nvpair id="cib-bootstrap-options-default-resource-stickiness" >> name="default-resource-stickiness" value="INFINITY"/> >> <nvpair id="cib-bootstrap-options-is-managed-default" >> name="is-managed-default" value="true"/> >> <nvpair id="cib-bootstrap-options-no-quorum-policy" >> name="no-quorum-policy" value="stop"/> >> <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" >> value="2.1.3-node: a3184d5240c6e7032aef9cce6e5b7752ded544b3"/> >> </attributes> >> </cluster_property_set> >> </crm_config> >> <nodes> >> <node id="5e0f97b7-6780-4487-baf9-6c36500b1276" uname="www2test" >> type="normal"/> >> <node id="3a325e23-2184-46ed-9e88-42a11f28c2be" uname="www1test" >> type="normal"/> >> </nodes> >> <resources> >> <group id="group_1"> >> <primitive class="ocf" id="IPaddr_192_168_11_25" >> provider="heartbeat" type="IPaddr"> >> <operations> >> <op id="IPaddr_192_168_11_25_mon" interval="5s" name="monitor" >> timeout="5s"/> >> </operations> >> <instance_attributes id="IPaddr_192_168_11_25_inst_attr"> >> <attributes> >> <nvpair id="IPaddr_192_168_11_25_attr_0" name="ip" >> value="192.168.11.25"/> >> </attributes> >> </instance_attributes> >> </primitive> >> <primitive class="ocf" id="apache_2" provider="heartbeat" >> type="apache"> >> <operations> >> <op id="apache_2_mon" interval="5s" name="monitor" >> timeout="10s"/> >> </operations> >> <instance_attributes id="apache_2_inst_attr"> >> <attributes> >> <nvpair id="apache_2_attr_0" name="configfile" >> value="/opt/apache2/conf/httpd.conf"/> >> </attributes> >> </instance_attributes> >> <instance_attributes id="apache_2"> >> <attributes> >> <nvpair id="apache_2-httpd" name="httpd" >> value="/opt/apache2/bin/httpd"/> >> </attributes> >> </instance_attributes> >> </primitive> >> </group> >> </resources> >> <constraints> >> <rsc_location id="run_group1" rsc="group_1"> >> <rule id="pref_run_apache_group" score="0"> >> <expression attribute="#uname" operation="eq" value="www1test" >> id="7667baf9-522d-40ac-a901-195bfe84a3df"/> >> </rule> >> </rsc_location> >> </constraints> >> </configuration> >> </cib> >> >> Gesch�ftsf�hrung: Dr. Michael Fischer, Reinhard Eisebitt >> Amtsgericht K�ln HRB 32356 >> Steuer-Nr.: 217/5717/0536 >> Ust.Id.-Nr.: DE 204051920 >> -- >> This email transmission and any documents, files or previous email >> messages attached to it may contain information that is confidential or >> legally privileged. If you are not the intended recipient or a person >> responsible for delivering this transmission to the intended recipient, >> you are hereby notified that any disclosure, copying, printing, >> distribution or use of this transmission is strictly prohibited. If you >> have received this transmission in error, please immediately notify the >> sender by telephone or return email and delete the original transmission >> and its attachments without reading or saving in any manner. >> >> _______________________________________________ >> Linux-HA mailing list >> [email protected] >> http://lists.linux-ha.org/mailman/listinfo/linux-ha >> See also: http://linux-ha.org/ReportingProblems >> > > > -- > > IN-telegence GmbH & Co. KG > Oskar-Jäger-Str. 125 > 50825 Köln > > Registergericht Köln - HRA 14064, USt-ID Nr. DE 194 156 373 > ph Gesellschafter: komware Unternehmensverwaltungsgesellschaft mbH, > Registergericht Köln - HRB 38396 > Geschäftsführende Gesellschafter: Christian Plätke und Holger Jansen > _______________________________________________ > Linux-HA mailing list > [email protected] > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems > > Geschäftsführung: Dr. Michael Fischer, Reinhard Eisebitt > Amtsgericht Köln HRB 32356 > Steuer-Nr.: 217/5717/0536 > Ust.Id.-Nr.: DE 204051920 > -- > This email transmission and any documents, files or previous email > messages attached to it may contain information that is confidential or > legally privileged. If you are not the intended recipient or a person > responsible for delivering this transmission to the intended recipient, > you are hereby notified that any disclosure, copying, printing, > distribution or use of this transmission is strictly prohibited. If you > have received this transmission in error, please immediately notify the > sender by telephone or return email and delete the original transmission > and its attachments without reading or saving in any manner. > > > > _______________________________________________ > Linux-HA mailing list > [email protected] > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems >
_______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
