Hi Michael,

 i have sent all the configuration files as an attachment.

here xendomains belongs to class "heartbeat resource agent".

After starting the heartbeat, the xendomains was started successfully
by the hearbteat.

/var/log/files

Jul 30 17:31:57 node3 tengine: [16475]: info: status_from_rc: Re-mapping op
status to LRM_OP_ERROR for rc=0
Jul 30 17:31:57 node3 crmd: [16459]: info: do_state_transition: All 1
cluster nodes are eligible to run resources.
Jul 30 17:31:57 node3 tengine: [16475]: WARN: status_from_rc: Action monitor
on node3 failed (target: 7 vs. rc: 0): Error
Jul 30 17:31:57 node3 tengine: [16475]: info: match_graph_event: Action
xendomains_2_monitor_0 (4) confirmed on node3 (rc=4)
Jul 30 17:31:57 node3 tengine: [16475]: info: run_graph:
====================================================
Jul 30 17:31:57 node3 tengine: [16475]: notice: run_graph: Transition 0:
(Complete=2, Pending=0, Fired=0, Skipped=7, Incomplete=1)
Jul 30 17:31:57 node3 pengine: [16476]: info: determine_online_status: Node
node3 is online
Jul 30 17:31:57 node3 pengine: [16476]: notice: group_print: Resource Group:
group_1
Jul 30 17:31:57 node3 pengine: [16476]: notice: native_print:
IPaddr_172_25_149_246#011(heartbeat::ocf:IPaddr):#011Started node3
Jul 30 17:31:57 node3 pengine: [16476]: notice: native_print:
xendomains_2#011(heartbeat:xendomains):#011Started node3
Jul 30 17:31:57 node3 pengine: [16476]: notice: NoRoleChange: Leave resource
IPaddr_172_25_149_246#011(node3)
Jul 30 17:31:57 node3 pengine: [16476]: notice: RecurringOp: node3#011
IPaddr_172_25_149_246_monitor_5000
Jul 30 17:31:57 node3 pengine: [16476]: notice: NoRoleChange: Leave resource
xendomains_2#011(node3)
Jul 30 17:31:57 node3 pengine: [16476]: notice: RecurringOp: node3#011
xendomains_2_monitor_120000
Jul 30 17:31:57 node3 crmd: [16459]: info: do_state_transition: State
transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS
cause=C_IPC_MESSAGE origin=route_message ]
Jul 30 17:31:57 node3 tengine: [16475]: info: unpack_graph: Unpacked
transition 1: 3 actions in 3 synapses
Jul 30 17:31:57 node3 tengine: [16475]: info: send_rsc_command: Initiating
action 6: IPaddr_172_25_149_246_monitor_5000 on node3
Jul 30 17:31:57 node3 tengine: [16475]: info: send_rsc_command: Initiating
action 9: xendomains_2_monitor_120000 on node3
Jul 30 17:31:57 node3 tengine: [16475]: info: send_rsc_command: Initiating
action 2: probe_complete on node3
Jul 30 17:31:57 node3 crmd: [16459]: info: do_lrm_rsc_op: Performing
op=IPaddr_172_25_149_246_monitor_5000
key=6:1:e4241df1-add0-4969-a9a5-a9ac3c9e4139)
Jul 30 17:31:57 node3 crmd: [16459]: info: do_lrm_rsc_op: Performing
op=xendomains_2_monitor_120000 key=9:1:e4241df1-add0-4969-a9a5-a9ac3c9e4139)
Jul 30 17:31:57 node3 tengine: [16475]: info: extract_event: Aborting on
transient_attributes changes for cbec221f-d26a-478e-a2ba-f4330f46a37b
Jul 30 17:31:57 node3 tengine: [16475]: info: update_abort_priority: Abort
priority upgraded to 1000000
Jul 30 17:31:57 node3 tengine: [16475]: info: update_abort_priority: Abort
action 0 superceeded by 2
Jul 30 17:31:58 node3 pengine: [16476]: info: process_pe_message: Transition
1: PEngine Input stored in: /var/lib/heartbeat/pengine/pe-input-153.bz2
Jul 30 17:31:59 node3 haclient: on_event: from message queue:
evt:cib_changed
Jul 30 17:31:59 node3 crmd: [16459]: info: process_lrm_event: LRM operation
IPaddr_172_25_149_246_monitor_5000 (call=4, rc=0) complete
Jul 30 17:31:59 node3 tengine: [16475]: info: match_graph_event: Action
IPaddr_172_25_149_246_monitor_5000 (6) confirmed on node3 (rc=0)
Jul 30 17:32:00 node3 haclient: on_event: from message queue:
evt:cib_changed
Jul 30 17:32:02 node3 crmd: [16459]: info: process_lrm_event: LRM operation
xendomains_2_monitor_120000 (call=5, rc=0) complete
Jul 30 17:32:02 node3 haclient: on_event:evt:cib_changed
Jul 30 17:32:02 node3 crmd: [16459]: info: do_state_transition: State
transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC
cause=C_IPC_MESSAGE origin=route_message ]
Jul 30 17:32:02 node3 tengine: [16475]: info: match_graph_event: Action
xendomains_2_monitor_120000 (9) confirmed on node3 (rc=0)
Jul 30 17:32:02 node3 crmd: [16459]: info: do_state_transition: All 1
cluster nodes are eligible to run resources.
Jul 30 17:32:02 node3 tengine: [16475]: info: run_graph: Transition 1:
(Complete=3, Pending=0, Fired=0, Skipped=0, Incomplete=0)
Jul 30 17:32:02 node3 pengine: [16476]: info: determine_online_status: Node
node3 is online
Jul 30 17:32:02 node3 pengine: [16476]: notice: group_print: Resource Group:
group_1
Jul 30 17:32:02 node3 pengine: [16476]: notice: native_print:
IPaddr_172_25_149_246#011(heartbeat::ocf:IPaddr):#011Started node3
Jul 30 17:32:02 node3 pengine: [16476]: notice: native_print:
xendomains_2#011(heartbeat:xendomains):#011Started node3
Jul 30 17:32:02 node3 pengine: [16476]: notice: NoRoleChange: Leave resource
IPaddr_172_25_149_246#011(node3)
Jul 30 17:32:02 node3 pengine: [16476]: notice: NoRoleChange: Leave resource
xendomains_2#011(node3)
Jul 30 17:32:02 node3 crmd: [16459]: info: do_state_transition: State
transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS
cause=C_IPC_MESSAGE origin=route_message ]
Jul 30 17:32:02 node3 tengine: [16475]: info: unpack_graph: Unpacked
transition 2: 0 actions in 0 synapses
Jul 30 17:32:02 node3 tengine: [16475]: info: run_graph: Transition 2:
(Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0)
Jul 30 17:32:02 node3 tengine: [16475]: info: notify_crmd: Transition 2
status: te_complete - <null>
Jul 30 17:32:02 node3 crmd: [16459]: info: do_state_transition: State
transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
cause=C_IPC_MESSAGE origin=route_message ]
Jul 30 17:32:02 node3 pengine: [16476]: info: process_pe_message: Transition
2: PEngine Input stored in: /var/lib/heartbeat/pengine/pe-input-154.bz2

the crm_mon was like this

============
Last updated: Wed Jul 30 17:41:02 2008
Current DC: node3 (cbec221f-d26a-478e-a2ba-f4330f46a37b)
1 Nodes configured.
1 Resources configured.
============
Node: node3 (cbec221f-d26a-478e-a2ba-f4330f46a37b): online

Resource Group: group_1
    IPaddr_172_25_149_246       (heartbeat::ocf:IPaddr):        Started
node3
    xendomains_2        (heartbeat:xendomains): Started node3

BUT AFTER STOPPING THE RESOUCE XENDOMAIN, IT DID NOT FAILBACK

like  /etc/ha.d/resource.d/xendomains virtual1 stop

then  the /var/log/messages

Jul 30 17:48:47 node3 kernel: eth0: port 2(vif5.0) entering disabled state
Jul 30 17:48:47 node3 kernel: eth0: port 2(vif5.0) entering disabled state
Jul 30 17:48:47 node3 kernel: device vif5.0 left promiscuous mode
Jul 30 17:48:47 node3 kernel: eth0: port 2(vif5.0) entering disabled state
Jul 30 17:49:59 node3 gnome-keyring-daemon[2541]: Credentials byte was not
nul
Jul 30 17:50:22 node3 crmd: [16459]: info: process_lrm_event: LRM operation
xendomains_2_monitor_120000 (call=5, rc=7) complete
Jul 30 17:50:22 node3 tengine: [16475]: info: process_graph_event: Detected
action xendomains_2_monitor_120000 from a different transition: 1 vs. 2
Jul 30 17:50:22 node3 crmd: [16459]: info: do_state_transition: State
transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_IPC_MESSAGE
origin=route_message ]
Jul 30 17:50:22 node3 tengine: [16475]: info: update_abort_priority: Abort
priority upgraded to 1000000
Jul 30 17:50:22 node3 crmd: [16459]: info: do_state_transition: All 1
cluster nodes are eligible to run resources.
Jul 30 17:50:22 node3 tengine: [16475]: WARN: update_failcount: Updating
failcount for xendomains_2 on cbec221f-d26a-478e-a2ba-f4330f46a37b after
failed monitor: rc=7
Jul 30 17:50:22 node3 haclient: on_event:evt:cib_changed
Jul 30 17:50:23 node3 pengine: [16476]: info: determine_online_status: Node
node3 is online
Jul 30 17:50:23 node3 pengine: [16476]: WARN: unpack_rsc_op: Processing
failed op xendomains_2_monitor_120000 on node3: Error
Jul 30 17:50:23 node3 pengine: [16476]: notice: group_print: Resource Group:
group_1
Jul 30 17:50:23 node3 pengine: [16476]: notice: native_print:
IPaddr_172_25_149_246#011(heartbeat::ocf:IPaddr):#011Started node3
Jul 30 17:50:23 node3 pengine: [16476]: notice: native_print:
xendomains_2#011(heartbeat:xendomains):#011Started node3 FAILED
Jul 30 17:50:23 node3 pengine: [16476]: notice: NoRoleChange: Leave resource
IPaddr_172_25_149_246#011(node3)
Jul 30 17:50:23 node3 pengine: [16476]: notice: NoRoleChange: Recover
resource xendomains_2#011(node3)
Jul 30 17:50:23 node3 pengine: [16476]: notice: StopRsc:   node3#011Stop
xendomains_2
Jul 30 17:50:23 node3 pengine: [16476]: notice: StartRsc:  node3#011Start
xendomains_2
Jul 30 17:50:23 node3 pengine: [16476]: notice: RecurringOp: node3#011
xendomains_2_monitor_120000
Jul 30 17:50:23 node3 tengine: [16475]: info: extract_event: Aborting on
transient_attributes changes for cbec221f-d26a-478e-a2ba-f4330f46a37b
Jul 30 17:50:24 node3 pengine: [16476]: info: process_pe_message: Transition
3: PEngine Input stored in: /var/lib/heartbeat/pengine/pe-input-155.bz2
Jul 30 17:50:24 node3 pengine: [16476]: info: determine_online_status: Node
node3 is online
Jul 30 17:50:24 node3 pengine: [16476]: WARN: unpack_rsc_op: Processing
failed op xendomains_2_monitor_120000 on node3: Error
Jul 30 17:50:24 node3 pengine: [16476]: notice: group_print: Resource Group:
group_1
Jul 30 17:50:24 node3 pengine: [16476]: notice: native_print:
IPaddr_172_25_149_246#011(heartbeat::ocf:IPaddr):#011Started node3
Jul 30 17:50:24 node3 pengine: [16476]: notice: native_print:
xendomains_2#011(heartbeat:xendomains):#011Started node3 FAILED
Jul 30 17:50:24 node3 pengine: [16476]: notice: NoRoleChange: Leave resource
IPaddr_172_25_149_246#011(node3)
Jul 30 17:50:24 node3 pengine: [16476]: notice: NoRoleChange: Recover
resource xendomains_2#011(node3)
Jul 30 17:50:24 node3 pengine: [16476]: notice: StopRsc:   node3#011Stop
xendomains_2
Jul 30 17:50:24 node3 pengine: [16476]: notice: StartRsc:  node3#011Start
xendomains_2
Jul 30 17:50:24 node3 pengine: [16476]: notice: RecurringOp: node3#011
xendomains_2_monitor_120000
Jul 30 17:50:24 node3 crmd: [16459]: info: do_state_transition: State
transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS
cause=C_IPC_MESSAGE origin=route_message ]
Jul 30 17:50:24 node3 tengine: [16475]: info: unpack_graph: Unpacked
transition 4: 10 actions in 10 synapses
Jul 30 17:50:24 node3 tengine: [16475]: info: te_pseudo_action: Pseudo
action 12 fired and confirmed
Jul 30 17:50:24 node3 tengine: [16475]: info: send_rsc_command: Initiating
action 3: xendomains_2_stop_0 on node3
Jul 30 17:50:24 node3 crmd: [16459]: info: do_lrm_rsc_op: Performing
op=xendomains_2_stop_0 key=3:4:e4241df1-add0-4969-a9a5-a9ac3c9e4139)
Jul 30 17:50:24 node3 lrmd: [16456]: info: rsc:xendomains_2: stop
Jul 30 17:50:24 node3 crmd: [16459]: info: process_lrm_event: LRM operation
xendomains_2_monitor_120000 (call=5, rc=-2) Cancelled
Jul 30 17:50:24 node3 pengine: [16476]: info: process_pe_message: Transition
4: PEngine Input stored in: /var/lib/heartbeat/pengine/pe-input-156.bz2
Jul 30 17:50:28 node3 lrmd: [16456]: info: RA output:
(xendomains_2:stop:stderr) Error: Domain 'virtual1' does not exist.
Jul 30 17:50:28 node3 lrmd: [16456]: info: RA output:
(xendomains_2:stop:stdout) #033[?1034h
Jul 30 17:50:28 node3 crmd: [16459]: ERROR: process_lrm_event: LRM operation
xendomains_2_stop_0 (call=6, rc=1) Error unknown error
Jul 30 17:50:28 node3 tengine: [16475]: WARN: status_from_rc: Action stop on
node3 failed (target: <null> vs. rc: 1): Error
Jul 30 17:50:28 node3 tengine: [16475]: WARN: update_failcount: Updating
failcount for xendomains_2 on cbec221f-d26a-478e-a2ba-f4330f46a37b after
failed stop: rc=1
Jul 30 17:50:28 node3 crmd: [16459]: info: do_state_transition: State
transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC
cause=C_IPC_MESSAGE origin=route_message ]
Jul 30 17:50:28 node3 tengine: [16475]: info: update_abort_priority: Abort
priority upgraded to 1
Jul 30 17:50:28 node3 crmd: [16459]: info: do_state_transition: All 1
cluster nodes are eligible to run resources.
Jul 30 17:50:28 node3 tengine: [16475]: info: update_abort_priority: Abort
action 0 superceeded by 2
Jul 30 17:50:28 node3 tengine: [16475]: info: match_graph_event: Action
xendomains_2_stop_0 (3) confirmed on node3 (rc=4)
Jul 30 17:50:28 node3 tengine: [16475]: info: run_graph:
====================================================
Jul 30 17:50:28 node3 tengine: [16475]: notice: run_graph: Transition 4:
(Complete=2, Pending=0, Fired=0, Skipped=8, Incomplete=0)
Jul 30 17:50:28 node3 pengine: [16476]: info: determine_online_status: Node
node3 is online
Jul 30 17:50:28 node3 pengine: [16476]: WARN: unpack_rsc_op: Processing
failed op xendomains_2_stop_0 on node3: Error
Jul 30 17:50:28 node3 pengine: [16476]: WARN: unpack_rsc_op: Compatability
handling for failed op xendomains_2_stop_0 on node3
Jul 30 17:50:28 node3 pengine: [16476]: notice: group_print: Resource Group:
group_1
Jul 30 17:50:28 node3 pengine: [16476]: notice: native_print:
IPaddr_172_25_149_246#011(heartbeat::ocf:IPaddr):#011Started node3
Jul 30 17:50:28 node3 pengine: [16476]: notice: native_print:
xendomains_2#011(heartbeat:xendomains):#011Started node3 (unmanaged) FAILED
Jul 30 17:50:28 node3 pengine: [16476]: WARN: native_color: Resource
xendomains_2 cannot run anywhere
Jul 30 17:50:28 node3 pengine: [16476]: notice: NoRoleChange: Leave resource
IPaddr_172_25_149_246#011(node3)
Jul 30 17:50:28 node3 pengine: [16476]: WARN: custom_action: Action
xendomains_2_stop_0 (unmanaged)
Jul 30 17:50:28 node3 crmd: [16459]: info: do_state_transition: State
transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS
cause=C_IPC_MESSAGE origin=route_message ]
Jul 30 17:50:28 node3 tengine: [16475]: info: unpack_graph: Unpacked
transition 5: 0 actions in 0 synapses
Jul 30 17:50:28 node3 tengine: [16475]: info: run_graph: Transition 5:
(Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0)
Jul 30 17:50:28 node3 crmd: [16459]: info: do_state_transition: State
transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
cause=C_IPC_MESSAGE origin=route_message ]
Jul 30 17:50:28 node3 tengine: [16475]: info: notify_crmd: Transition 5
status: te_complete - <null>
Jul 30 17:50:28 node3 pengine: [16476]: WARN: process_pe_message: Transition
5: WARNINGs found during PE processing. PEngine Input stored in:
/var/lib/heartbeat/pengine/pe-warn-395.bz2
Jul 30 17:50:28 node3 pengine: [16476]: info: process_pe_message:
Configuration WARNINGs found during PE processing.  Please run "crm_verify
-L" to identify issues.
Jul 30 17:50:31 node3 haclient: on_event: from message queue:
evt:cib_changed
Jul 30 17:51:08 node3 haclient:last message repeated 3 times
Jul 30 17:51:08 node3 cib: [16455]: info: cib_stats: Processed 53 operations
(13962.00us average, 0% utilization) in the last 10min
Jul 30 17:51:27 node3 lrmd: [16456]: WARN: G_SIG_dispatch: Dispatch function
for SIGCHLD was delayed 1000 ms (> 100 ms) before being called (GSource:
0x8843470)
Jul 30 17:51:27 node3 lrmd: [16456]: info: G_SIG_dispatch: started at
1719774701 should have started at 1719774601


and the crm_mon

============
Last updated: Wed Jul 30 17:56:21 2008
Current DC: node3 (cbec221f-d26a-478e-a2ba-f4330f46a37b)
1 Nodes configured.
1 Resources configured.
============

Node: node3 (cbec221f-d26a-478e-a2ba-f4330f46a37b): online

Resource Group: group_1
    IPaddr_172_25_149_246       (heartbeat::ocf:IPaddr):        Started
node3
    xendomains_2        (heartbeat:xendomains): Started node3 (unmanaged)
FAILED

Failed actions:
    xendomains_2_stop_0 (node=node3, call=6, rc=1): Error

Please help me out with this, i have been working on this for almost 1 month
awaiting ur help


devi

Attachment: haresources
Description: Binary data

Attachment: ha.cf
Description: Binary data

Attachment: xendomains
Description: Binary data

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to