On 5/8/07, Andrew Beekhof <[EMAIL PROTECTED]> wrote:
grep ERROR logfile try this for starters: May 7 16:31:41 qclsles01 lrmd: [5020]: info: RA output: (resource_qclvmsles02:stop:stderr) Error: the domain 'resource_qclvmsles02' does not exist. May 7 16:31:41 qclsles01 lrmd: [5020]: info: RA output: (resource_qclvmsles02:stop:stdout) Domain resource_qclvmsles02 terminated May 7 16:31:41 qclsles01 crmd: [22028]: WARN: process_lrm_event:lrm.c LRM operation (35) stop_0 on resource_qclvmsles02 Error: (4) insufficient privileges
yup I saw that.. it's weird. Heartbeat shutdown the vm, then say these errors.. and if I cleanup the ressource he restart on the correct node.. There should be something I missed lol On 5/7/07, Rene Purcell <[EMAIL PROTECTED]> wrote:
> I would like to know if someone had tried the Novell setup described in " > http://www.novell.com/linux/technical_library/has.pdf" with a x86_64 arch ? > > I've tested this setup with a classic x86 arch and everything was ok... but > I doublechecked my config and everything look good but my VM never start on > his original node when it come back online... and I can't find why! > > > here's the log when my node1 come back.. we can see the VM shutting down and > after that nothing happend in the other node.. > > May 7 16:31:25 qclsles01 cib: [22024]: info: > cib_diff_notify:notify.cUpdate (client: 6403, call:13): > 0.65.1020 -> 0.65.1021 (ok) > May 7 16:31:25 qclsles01 tengine: [22591]: info: > te_update_diff:callbacks.cProcessing diff (cib_update): > 0.65.1020 -> 0.65.1021 > May 7 16:31:25 qclsles01 tengine: [22591]: info: > extract_event:events.cAborting on transient_attributes changes > May 7 16:31:25 qclsles01 tengine: [22591]: info: update_abort_priority: > utils.c Abort priority upgraded to 1000000 > May 7 16:31:25 qclsles01 tengine: [22591]: info: update_abort_priority: > utils.c Abort action 0 superceeded by 2 > May 7 16:31:26 qclsles01 cib: [22024]: info: activateCibXml:io.c CIB size > is 161648 bytes (was 158548) > May 7 16:31:26 qclsles01 cib: [22024]: info: > cib_diff_notify:notify.cUpdate (client: 6403, call:14): > 0.65.1021 -> 0.65.1022 (ok) > May 7 16:31:26 qclsles01 haclient: on_event:evt:cib_changed > May 7 16:31:26 qclsles01 tengine: [22591]: info: > te_update_diff:callbacks.cProcessing diff (cib_update): > 0.65.1021 -> 0.65.1022 > May 7 16:31:26 qclsles01 tengine: [22591]: info: > match_graph_event:events.cAction resource_qclvmsles02_stop_0 (9) > confirmed > May 7 16:31:26 qclsles01 cib: [25889]: info: write_cib_contents:io.cWrote > version 0.65.1022 of the CIB to disk (digest: > e71c271759371d44c4bad24d50b2421d) > May 7 16:31:39 qclsles01 kernel: xenbr0: port 3(vif12.0) entering disabled > state > May 7 16:31:39 qclsles01 kernel: device vif12.0 left promiscuous mode > May 7 16:31:39 qclsles01 kernel: xenbr0: port 3(vif12.0) entering disabled > state > May 7 16:31:39 qclsles01 logger: /etc/xen/scripts/vif-bridge: offline > XENBUS_PATH=backend/vif/12/0 > May 7 16:31:40 qclsles01 logger: /etc/xen/scripts/block: remove > XENBUS_PATH=backend/vbd/12/768 > May 7 16:31:40 qclsles01 logger: /etc/xen/scripts/block: remove > XENBUS_PATH=backend/vbd/12/832 > May 7 16:31:40 qclsles01 logger: /etc/xen/scripts/block: remove > XENBUS_PATH=backend/vbd/12/5632 > May 7 16:31:40 qclsles01 logger: /etc/xen/scripts/vif-bridge: brctl delif > xenbr0 vif12.0 failed > May 7 16:31:40 qclsles01 logger: /etc/xen/scripts/vif-bridge: ifconfig > vif12.0 down failed > May 7 16:31:40 qclsles01 logger: /etc/xen/scripts/vif-bridge: Successful > vif-bridge offline for vif12.0, bridge xenbr0. > May 7 16:31:40 qclsles01 logger: /etc/xen/scripts/xen-hotplug-cleanup: > XENBUS_PATH=backend/vbd/12/5632 > May 7 16:31:40 qclsles01 logger: /etc/xen/scripts/xen-hotplug-cleanup: > XENBUS_PATH=backend/vbd/12/768 > May 7 16:31:40 qclsles01 ifdown: vif12.0 > May 7 16:31:40 qclsles01 logger: /etc/xen/scripts/xen-hotplug-cleanup: > XENBUS_PATH=backend/vif/12/0 > May 7 16:31:40 qclsles01 logger: /etc/xen/scripts/xen-hotplug-cleanup: > XENBUS_PATH=backend/vbd/12/832 > May 7 16:31:40 qclsles01 ifdown: Interface not available and no > configuration found. > May 7 16:31:41 qclsles01 lrmd: [5020]: info: RA output: > (resource_qclvmsles02:stop:stderr) Error: the domain 'resource_qclvmsles02' > does not exist. > May 7 16:31:41 qclsles01 lrmd: [5020]: info: RA output: > (resource_qclvmsles02:stop:stdout) Domain resource_qclvmsles02 terminated > May 7 16:31:41 qclsles01 crmd: [22028]: WARN: process_lrm_event:lrm.cLRM > operation (35) stop_0 on resource_qclvmsles02 Error: (4) insufficient > privileges > May 7 16:31:41 qclsles01 cib: [22024]: info: activateCibXml:io.c CIB size > is 164748 bytes (was 161648) > May 7 16:31:41 qclsles01 crmd: [22028]: info: > do_state_transition:fsa.cqclsles01: State transition > S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ > input=I_PE_CALC cause=C_IPC_MESSAGE origin=route_message ] > May 7 16:31:41 qclsles01 tengine: [22591]: info: > te_update_diff:callbacks.cProcessing diff (cib_update): > 0.65.1022 -> 0.65.1023 > May 7 16:31:41 qclsles01 cib: [22024]: info: > cib_diff_notify:notify.cUpdate (client: 22028, call:100): > 0.65.1022 -> 0.65.1023 (ok) > May 7 16:31:41 qclsles01 crmd: [22028]: info: do_state_transition:fsa.cAll > 2 cluster nodes are eligable to run resources. > May 7 16:31:41 qclsles01 tengine: [22591]: ERROR: match_graph_event: > events.c Action resource_qclvmsles02_stop_0 on qclsles01 failed (target: 0 > vs. rc: 4): Error > May 7 16:31:41 qclsles01 tengine: [22591]: info: > match_graph_event:events.cAction resource_qclvmsles02_stop_0 (10) > confirmed > May 7 16:31:41 qclsles01 tengine: [22591]: info: > run_graph:graph.c==================================================== > May 7 16:31:41 qclsles01 tengine: [22591]: notice: > run_graph:graph.cTransition 12: (Complete=3, Pending=0, Fired=0, > Skipped=2, Incomplete=0) > May 7 16:31:41 qclsles01 haclient: on_event:evt:cib_changed > May 7 16:31:41 qclsles01 cib: [26190]: info: write_cib_contents:io.cWrote > version 0.65.1023 of the CIB to disk (digest: > c80326e44b5a106fe9a384240c4a3cc9) > May 7 16:31:41 qclsles01 pengine: [22592]: info: process_pe_message: > [generation] <cib generated="true" admin_epoch="0" have_quorum="true" > num_peers="2" cib_feature_revision="1.3" ccm_transition="10" > dc_uuid="46ef9c7b-5f6e-4cc0-a0bb-94227b605170" epoch="65" > num_updates="1023"/> > May 7 16:31:41 qclsles01 pengine: [22592]: WARN: unpack_config:unpack.cNo > value specified for cluster preference: default_action_timeout > May 7 16:31:41 qclsles01 pengine: [22592]: info: > unpack_config:unpack.cDefault stickiness: 1000000 > May 7 16:31:41 qclsles01 pengine: [22592]: info: > unpack_config:unpack.cDefault failure stickiness: -500 > May 7 16:31:41 qclsles01 pengine: [22592]: info: > unpack_config:unpack.cSTONITH of failed nodes is disabled > May 7 16:31:41 qclsles01 pengine: [22592]: info: > unpack_config:unpack.cSTONITH will reboot nodes > May 7 16:31:41 qclsles01 pengine: [22592]: info: > unpack_config:unpack.cCluster is symmetric - resources can run > anywhere by default > May 7 16:31:41 qclsles01 pengine: [22592]: info: unpack_config:unpack.cOn > loss of CCM Quorum: Stop ALL resources > May 7 16:31:41 qclsles01 pengine: [22592]: info: > unpack_config:unpack.cOrphan resources are stopped > May 7 16:31:41 qclsles01 pengine: [22592]: info: > unpack_config:unpack.cOrphan resource actions are stopped > May 7 16:31:41 qclsles01 pengine: [22592]: WARN: unpack_config:unpack.cNo > value specified for cluster preference: remove_after_stop > May 7 16:31:41 qclsles01 pengine: [22592]: info: > unpack_config:unpack.cStopped resources are removed from the status > section: false > May 7 16:31:41 qclsles01 pengine: [22592]: info: unpack_config:unpack.cBy > default resources are managed > May 7 16:31:41 qclsles01 pengine: [22592]: info: determine_online_status: > unpack.c Node qclsles02 is online > May 7 16:31:41 qclsles01 pengine: [22592]: info: determine_online_status: > unpack.c Node qclsles01 is online > May 7 16:31:41 qclsles01 pengine: [22592]: WARN: > unpack_rsc_op:unpack.cProcessing failed op > (resource_qclvmsles02_stop_0) for resource_qclvmsles02 > on qclsles01 > May 7 16:31:41 qclsles01 pengine: [22592]: WARN: > unpack_rsc_op:unpack.cHandling failed stop for resource_qclvmsles02 on > qclsles01 > May 7 16:31:41 qclsles01 pengine: [22592]: info: process_orphan_resource: > Orphan resource <lrm_resource id="resource_NFS" type="nfs" class="lsb" > provider="heartbeat"> > May 7 16:31:41 qclsles01 pengine: [22592]: info: process_orphan_resource: > Orphan resource <lrm_rsc_op id="resource_NFS_monitor_0" > operation="monitor" crm-debug-origin="build_active_RAs" > transition_key="27:3a815bc6-ffaa-49b3-aac2-0ed46e85f085" > transition_magic="0:0;27:3a815bc6-ffaa-49b3-aac2-0ed46e85f085" call_id="9" > crm_feature_set="1.0.6" rc_code="0" op_status="0" interval="0" > op_digest="08b7001b97ccdaa1ca23a9f165256bc1"/> > May 7 16:31:41 qclsles01 pengine: [22592]: info: process_orphan_resource: > Orphan resource <lrm_rsc_op id="resource_NFS_stop_0" operation="stop" > crm-debug-origin="build_active_RAs" > transition_key="28:3a815bc6-ffaa-49b3-aac2-0ed46e85f085" > transition_magic="0:0;28:3a815bc6-ffaa-49b3-aac2-0ed46e85f085" call_id="10" > crm_feature_set="1.0.6" rc_code="0" op_status="0" interval="0" > op_digest="08b7001b97ccdaa1ca23a9f165256bc1"/> > May 7 16:31:41 qclsles01 pengine: [22592]: info: process_orphan_resource: > Orphan resource </lrm_resource> > May 7 16:31:41 qclsles01 pengine: [22592]: WARN: process_orphan_resource: > unpack.c Nothing known about resource resource_NFS running on qclsles01 > May 7 16:31:41 qclsles01 pengine: [22592]: info: create_fake_resource: > Orphan resource <lrm_resource id="resource_NFS" type="nfs" class="lsb" > provider="heartbeat"> > May 7 16:31:41 qclsles01 pengine: [22592]: info: create_fake_resource: > Orphan resource <lrm_rsc_op id="resource_NFS_monitor_0" > operation="monitor" crm-debug-origin="build_active_RAs" > transition_key="27:3a815bc6-ffaa-49b3-aac2-0ed46e85f085" > transition_magic="0:0;27:3a815bc6-ffaa-49b3-aac2-0ed46e85f085" call_id="9" > crm_feature_set="1.0.6" rc_code="0" op_status="0" interval="0" > op_digest="08b7001b97ccdaa1ca23a9f165256bc1"/> > May 7 16:31:41 qclsles01 pengine: [22592]: info: create_fake_resource: > Orphan resource <lrm_rsc_op id="resource_NFS_stop_0" operation="stop" > crm-debug-origin="build_active_RAs" > transition_key="28:3a815bc6-ffaa-49b3-aac2-0ed46e85f085" > transition_magic="0:0;28:3a815bc6-ffaa-49b3-aac2-0ed46e85f085" call_id="10" > crm_feature_set="1.0.6" rc_code="0" op_status="0" interval="0" > op_digest="08b7001b97ccdaa1ca23a9f165256bc1"/> > May 7 16:31:41 qclsles01 pengine: [22592]: info: create_fake_resource: > Orphan resource </lrm_resource> > May 7 16:31:41 qclsles01 pengine: [22592]: info: process_orphan_resource: > unpack.c Making sure orphan resource_NFS is stopped > May 7 16:31:41 qclsles01 pengine: [22592]: info: resource_qclvmsles01 > (heartbeat::ocf:Xen): Started qclsles01 > May 7 16:31:41 qclsles01 pengine: [22592]: info: resource_qclvmsles02 > (heartbeat::ocf:Xen): Started qclsles01 (unmanaged) FAILED > May 7 16:31:41 qclsles01 pengine: [22592]: info: resource_NFS > (lsb:nfs): Stopped > May 7 16:31:41 qclsles01 pengine: [22592]: notice: > NoRoleChange:native.cLeave resource resource_qclvmsles01 > (qclsles01) > May 7 16:31:41 qclsles01 pengine: [22592]: notice: > NoRoleChange:native.cMove resource resource_qclvmsles02 (qclsles01 > -> qclsles02) > May 7 16:31:41 qclsles01 crmd: [22028]: info: > do_state_transition:fsa.cqclsles01: State transition S_POLICY_ENGINE > -> S_TRANSITION_ENGINE [ > input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=route_message ] > May 7 16:31:41 qclsles01 pengine: [22592]: WARN: > custom_action:utils.cAction resource_qclvmsles02_stop_0 stop is for > resource_qclvmsles02 > (unmanaged) > May 7 16:31:41 qclsles01 pengine: [22592]: WARN: > custom_action:utils.cAction resource_qclvmsles02_start_0 start is for > resource_qclvmsles02 > (unmanaged) > May 7 16:31:41 qclsles01 pengine: [22592]: notice: > stage8:allocate.cCreated transition graph 13. > May 7 16:31:41 qclsles01 tengine: [22591]: info: > unpack_graph:unpack.cUnpacked transition 13: 0 actions in 0 synapses > May 7 16:31:41 qclsles01 crmd: [22028]: info: > do_state_transition:fsa.cqclsles01: State transition > S_TRANSITION_ENGINE -> S_IDLE [ > input=I_TE_SUCCESS cause=C_IPC_MESSAGE origin=route_message ] > May 7 16:31:41 qclsles01 pengine: [22592]: WARN: process_pe_message: > pengine.c No value specified for cluster preference: pe-input-series-max > May 7 16:31:41 qclsles01 tengine: [22591]: info: > run_graph:graph.cTransition 13: (Complete=0, Pending=0, Fired=0, > Skipped=0, Incomplete=0) > May 7 16:31:41 qclsles01 pengine: [22592]: info: process_pe_message: > pengine.c Transition 13: PEngine Input stored in: > /var/lib/heartbeat/pengine/pe-input-100.bz2 > May 7 16:31:41 qclsles01 tengine: [22591]: info: > notify_crmd:actions.cTransition 13 status: te_complete - (null) > > > Thanks! > > > -- > René Jr Purcell > Chargé de projet, sécurité et sytèmes > Techno Centre Logiciels Libres, http://www.tc2l.ca/ > Téléphone : (418) 681-2929 #124 > _______________________________________________ > Linux-HA mailing list > [email protected] > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems > _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
-- René Jr Purcell Chargé de projet, sécurité et sytèmes Techno Centre Logiciels Libres, http://www.tc2l.ca/ Téléphone : (418) 681-2929 #124 _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
