Hi guys,

I have a simple 2 node cluster with a VIP running on RHEL 5.3 on s390. 
Nothing else configured yet.

When I start up the cluster, all is well. The VIP starts up on the home 
node and crm_mon shows the resource and nodes as on line. No errors in 
the logs.

If I issue service heartbeat stop on the main node, the ip fails over to 
the back up node and crm_mon shows as I would expect it should, i.e. the 
ip address is on the back up node and that the other node is offline. 
However, if I do a ifconfig on the main node I see that the eth0:0 entry 
is still there so in effect the vip address is now on both servers.

If both nodes were up and running and I rebooted the main node then the 
failover works perfectly.

Would anyone know why the nodes seem unable to release the vip unless 
rebooted?

ha-log:
Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
stage6: Scheduling Node dbsuat1a.intranet.mydomain.com for shutdown
Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com pengine: [4714]: notice: 
LogActions: Move resource IPaddr_172_28_185_49    (Started 
dbsuat1a.intranet.mydomain.com -> dbsuat1b.intranet.mydomain.com)
Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_state_transition: State transition S_POLICY_ENGINE -> 
S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE 
origin=handle_response ]
Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
process_pe_message: Transition 5: PEngine Input stored in: 
/usr/var/lib/pengine/pe-input-337.bz2
Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
process_pe_message: Configuration WARNINGs found during PE processing.  
Please run "crm_verify -L" to identify issues.
Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
unpack_graph: Unpacked transition 5: 5 actions in 5 synapses
Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_te_invoke: Processing graph 5 (ref=pe_calc-dc-1269021312-26) derived 
from /usr/var/lib/pengine/pe-input-337.bz2
Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
te_rsc_command: Initiating action 6: stop IPaddr_172_28_185_49_stop_0 on 
dbsuat1a.intranet.mydomain.com (local)
Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_lrm_rsc_op: Performing key=6:5:0:888fa84e-3267-409e-966b-2ab01e579c0f 
op=IPaddr_172_28_185_49_stop_0 )
Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com lrmd: [4528]: info: 
rsc:IPaddr_172_28_185_49:5: stop
Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
process_lrm_event: LRM operation IPaddr_172_28_185_49_monitor_5000 
(call=4, status=1, cib-update=0, confirmed=true) Cancelled
Mar 19 13:55:32 DBSUAT1A.intranet.mydomain.com lrmd: [4528]: WARN: 
IPaddr_172_28_185_49:stop process (PID 5474) timed out (try 1).  Killing 
with signal SIGTERM (15).
Mar 19 13:55:32 DBSUAT1A.intranet.mydomain.com lrmd: [4528]: WARN: 
Managed IPaddr_172_28_185_49:stop process 5474 killed by signal 15 
[SIGTERM - Termination (ANSI)].
Mar 19 13:55:32 DBSUAT1A.intranet.mydomain.com lrmd: [4528]: WARN: 
operation stop[5] on ocf::IPaddr::IPaddr_172_28_185_49 for client 4531, 
its parameters: ip=[172.28.185.49] CRM_meta_timeout=[20000] 
crm_feature_set=[3.0.1] : pid [5474] timed out
Mar 19 13:55:32 DBSUAT1A.intranet.mydomain.com crmd: [4531]: ERROR: 
process_lrm_event: LRM operation IPaddr_172_28_185_49_stop_0 (5) Timed 
Out (timeout=20000ms)
Mar 19 13:55:32 DBSUAT1A.intranet.mydomain.com crmd: [4531]: WARN: 
status_from_rc: Action 6 (IPaddr_172_28_185_49_stop_0) on 
dbsuat1a.intranet.mydomain.com failed (target: 0 vs. rc: -2): Error
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: WARN: 
update_failcount: Updating failcount for IPaddr_172_28_185_49 on 
dbsuat1a.intranet.mydomain.com after failed stop: rc=-2 
(update=INFINITY, time=1269021333)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
abort_transition_graph: match_graph_event:272 - Triggered transition 
abort (complete=0, tag=lrm_rsc_op, id=IPaddr_172_28_185_49_stop_0, 
magic=2:-2;6:5:0:888fa84e-3267-409e-966b-2ab01e579c0f, cib=0.23.16) : 
Event failed
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
update_abort_priority: Abort priority upgraded from 0 to 1
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
update_abort_priority: Abort action done superceeded by restart
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
match_graph_event: Action IPaddr_172_28_185_49_stop_0 (6) confirmed on 
dbsuat1a.intranet.mydomain.com (rc=4)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
run_graph: ====================================================
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: notice: 
run_graph: Transition 5 (Complete=1, Pending=0, Fired=0, Skipped=4, 
Incomplete=0, Source=/usr/var/lib/pengine/pe-input-337.bz2): Stopped
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
te_graph_trigger: Transition 5 is now complete
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info: 
find_hash_entry: Creating hash entry for fail-count-IPaddr_172_28_185_49
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_state_transition: State transition S_TRANSITION_ENGINE -> 
S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL origin=notify_crmd ]
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_state_transition: All 2 cluster nodes are eligible to run resources.
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info: 
attrd_trigger_update: Sending flush op to all hosts for: 
fail-count-IPaddr_172_28_185_49 (INFINITY)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info: 
attrd_perform_update: Sent update 24: 
fail-count-IPaddr_172_28_185_49=INFINITY
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_pe_invoke: Query 53: Requesting the current CIB: S_POLICY_ENGINE
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info: 
find_hash_entry: Creating hash entry for last-failure-IPaddr_172_28_185_49
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
abort_transition_graph: te_update_diff:146 - Triggered transition abort 
(complete=1, tag=transient_attributes, 
id=db80324b-c9de-4995-a66a-eedf93abb42c, magic=NA, cib=0.23.17) : 
Transient attribute: update
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info: 
attrd_trigger_update: Sending flush op to all hosts for: 
last-failure-IPaddr_172_28_185_49 (1269021333)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info: 
attrd_perform_update: Sent update 27: 
last-failure-IPaddr_172_28_185_49=1269021333
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_pe_invoke_callback: Invoking the PE: ref=pe_calc-dc-1269021333-28, 
seq=2, quorate=1
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
abort_transition_graph: te_update_diff:146 - Triggered transition abort 
(complete=1, tag=transient_attributes, 
id=db80324b-c9de-4995-a66a-eedf93abb42c, magic=NA, cib=0.23.18) : 
Transient attribute: update
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: notice: 
update_validation: Upgrading transitional-0.6-style configuration to 
pacemaker-1.0 with /usr/share/pacemaker/upgrade06.xsl
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
update_validation: Transformation /usr/share/pacemaker/upgrade06.xsl 
successful
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: notice: 
update_validation: Upgraded from transitional-0.6 to pacemaker-1.0 
validation
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: WARN: 
cli_config_update: Your configuration was internally updated to the 
latest version (pacemaker-1.0)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_pe_invoke: Query 54: Requesting the current CIB: S_POLICY_ENGINE
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_pe_invoke: Query 55: Requesting the current CIB: S_POLICY_ENGINE
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_pe_invoke_callback: Invoking the PE: ref=pe_calc-dc-1269021333-29, 
seq=2, quorate=1
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
determine_online_status: Node dbsuat1a.intranet.mydomain.com is shutting 
down
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: WARN: 
unpack_rsc_op: Processing failed op IPaddr_172_28_185_49_stop_0 on 
dbsuat1a.intranet.mydomain.com: unknown exec error (-2)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
native_add_running: resource IPaddr_172_28_185_49 isnt managed
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
determine_online_status: Node dbsuat1b.intranet.mydomain.com is online
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: notice: 
native_print: IPaddr_172_28_185_49    (ocf::heartbeat:IPaddr):    
Started dbsuat1a.intranet.mydomain.com (unmanaged) FAILED
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
get_failcount: IPaddr_172_28_185_49 has failed 1000000 times on 
dbsuat1a.intranet.mydomain.com
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: WARN: 
common_apply_stickiness: Forcing IPaddr_172_28_185_49 away from 
dbsuat1a.intranet.mydomain.com after 1000000 failures (max=1000000)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
native_color: Unmanaged resource IPaddr_172_28_185_49 allocated to 
'nowhere': failed
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
stage6: Scheduling Node dbsuat1a.intranet.mydomain.com for shutdown
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: notice: 
LogActions: Leave resource IPaddr_172_28_185_49    (Started unmanaged)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
handle_response: pe_calc calculation pe_calc-dc-1269021333-28 is obsolete
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
process_pe_message: Transition 6: PEngine Input stored in: 
/usr/var/lib/pengine/pe-input-338.bz2
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
process_pe_message: Configuration WARNINGs found during PE processing.  
Please run "crm_verify -L" to identify issues.
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: notice: 
update_validation: Upgrading transitional-0.6-style configuration to 
pacemaker-1.0 with /usr/share/pacemaker/upgrade06.xsl
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
update_validation: Transformation /usr/share/pacemaker/upgrade06.xsl 
successful
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: notice: 
update_validation: Upgraded from transitional-0.6 to pacemaker-1.0 
validation
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: WARN: 
cli_config_update: Your configuration was internally updated to the 
latest version (pacemaker-1.0)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
determine_online_status: Node dbsuat1a.intranet.mydomain.com is shutting 
down
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: WARN: 
unpack_rsc_op: Processing failed op IPaddr_172_28_185_49_stop_0 on 
dbsuat1a.intranet.mydomain.com: unknown exec error (-2)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
native_add_running: resource IPaddr_172_28_185_49 isnt managed
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
determine_online_status: Node dbsuat1b.intranet.mydomain.com is online
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: notice: 
native_print: IPaddr_172_28_185_49    (ocf::heartbeat:IPaddr):    
Started dbsuat1a.intranet.mydomain.com (unmanaged) FAILED
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
get_failcount: IPaddr_172_28_185_49 has failed 1000000 times on 
dbsuat1a.intranet.mydomain.com
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: WARN: 
common_apply_stickiness: Forcing IPaddr_172_28_185_49 away from 
dbsuat1a.intranet.mydomain.com after 1000000 failures (max=1000000)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
native_color: Unmanaged resource IPaddr_172_28_185_49 allocated to 
'nowhere': failed
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
stage6: Scheduling Node dbsuat1a.intranet.mydomain.com for shutdown
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: notice: 
LogActions: Leave resource IPaddr_172_28_185_49    (Started unmanaged)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
process_pe_message: Transition 7: PEngine Input stored in: 
/usr/var/lib/pengine/pe-input-339.bz2
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_state_transition: State transition S_POLICY_ENGINE -> 
S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE 
origin=handle_response ]
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
process_pe_message: Configuration WARNINGs found during PE processing.  
Please run "crm_verify -L" to identify issues.
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
unpack_graph: Unpacked transition 7: 1 actions in 1 synapses
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_te_invoke: Processing graph 7 (ref=pe_calc-dc-1269021333-29) derived 
from /usr/var/lib/pengine/pe-input-339.bz2
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
te_crm_command: Executing crm-event (10): do_shutdown on 
dbsuat1a.intranet.mydomain.com
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
te_crm_command: crm-event (10) is a local shutdown
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
run_graph: ====================================================
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: notice: 
run_graph: Transition 7 (Complete=1, Pending=0, Fired=0, Skipped=0, 
Incomplete=0, Source=/usr/var/lib/pengine/pe-input-339.bz2): Complete
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
te_graph_trigger: Transition 7 is now complete
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_state_transition: State transition S_TRANSITION_ENGINE -> S_STOPPING 
[ input=I_STOP cause=C_FSA_INTERNAL origin=notify_crmd ]
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_dc_release: DC role released
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
stop_subsystem: Sent -TERM to pengine: [4714]
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_te_control: Transitioner is now inactive
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info: 
crm_signal_dispatch: Invoking handler for signal 15: Terminated
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_te_control: Disconnecting STONITH...
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
tengine_stonith_connection_destroy: Fencing daemon disconnected
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: notice: Not 
currently connected.
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_shutdown: Terminating the pengine
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
stop_subsystem: Sent -TERM to pengine: [4714]
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_shutdown: Waiting for subsystems to exit
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: WARN: 
register_fsa_input_adv: do_shutdown stalled the FSA with pending inputs
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_shutdown: All subsystems stopped, continuing
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: WARN: 
do_log: FSA: Input I_RELEASE_SUCCESS from do_dc_release() received in 
state S_STOPPING
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_shutdown: Terminating the pengine
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
stop_subsystem: Sent -TERM to pengine: [4714]
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_shutdown: Waiting for subsystems to exit
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_shutdown: All subsystems stopped, continuing
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
crmdManagedChildDied: Process pengine:[4714] exited (signal=0, exitcode=0)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
pe_msg_dispatch: Received HUP from pengine:[4714]
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
pe_connection_destroy: Connection to the Policy Engine released
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_shutdown: All subsystems stopped, continuing
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: ERROR: 
verify_stopped: Resource IPaddr_172_28_185_49 was active at shutdown.  
You may ignore this error if it is unmanaged.
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_lrm_control: Disconnected from the LRM
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com ccm: [4526]: info: client 
(pid=4531) removed from ccm
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_ha_control: Disconnected from Heartbeat
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_cib_control: Disconnecting CIB
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com cib: [4527]: info: 
cib_process_readwrite: We are now in R/O mode
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
crmd_cib_connection_destroy: Connection to the CIB terminated...
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_exit: Performing A_EXIT_0 - gracefully exiting the CRMd
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
free_mem: Dropping I_TERMINATE: [ state=S_STOPPING cause=C_FSA_INTERNAL 
origin=do_stop ]
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info: 
do_exit: [crmd] stopped (0)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]: info: 
killing /usr/lib64/heartbeat/attrd process group 4530 with signal 15
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info: 
crm_signal_dispatch: Invoking handler for signal 15: Terminated
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info: 
attrd_shutdown: Exiting
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info: 
main: Exiting...
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info: 
attrd_cib_connection_destroy: Connection to the CIB terminated...
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]: info: 
killing /usr/lib64/heartbeat/stonithd process group 4529 with signal 15
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com stonithd: [4529]: notice: 
/usr/lib64/heartbeat/stonithd normally quit.
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]: info: 
killing /usr/lib64/heartbeat/lrmd -r process group 4528 with signal 15
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com lrmd: [4528]: info: lrmd 
is shutting down
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com lrmd: [4528]: WARN: 
resource IPaddr_172_28_185_49 is left in UNKNOWN status.(last op stop 
finished without LRM_OP_DONE status.)
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]: info: 
killing /usr/lib64/heartbeat/cib process group 4527 with signal 15
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com cib: [4527]: info: 
crm_signal_dispatch: Invoking handler for signal 15: Terminated
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com cib: [4527]: info: 
cib_shutdown: Disconnected 0 clients
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com cib: [4527]: info: 
cib_process_disconnect: All clients disconnected...
Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com cib: [4527]: info: 
initiate_exit: Sending disconnect notification to 2 peers...
Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com cib: [4527]: info: 
cib_process_shutdown_req: Shutdown ACK from dbsuat1b.intranet.mydomain.com
Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com cib: [4527]: info: 
terminate_cib: cib_process_shutdown_req: Disconnecting heartbeat
Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com cib: [4527]: info: 
terminate_cib: Exiting...
Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com cib: [4527]: info: 
cib_process_request: Operation complete: op cib_shutdown_req for section 
'all' 
(origin=dbsuat1b.intranet.mydomain.com/dbsuat1b.intranet.mydomain.com/(null), 
version=0.0.0): ok (rc=0)
Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com cib: [4527]: info: 
ha_msg_dispatch: Lost connection to heartbeat service.
Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com cib: [4527]: info: main: 
Done
Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com ccm: [4526]: info: client 
(pid=4527) removed from ccm
Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]: info: 
killing /usr/lib64/heartbeat/ccm process group 4526 with signal 15
Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com ccm: [4526]: info: 
received SIGTERM, going to shut down
Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]: info: 
killing HBFIFO process 4522 with signal 15
Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]: info: 
killing HBWRITE process 4523 with signal 15
Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]: info: 
killing HBREAD process 4524 with signal 15
Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]: info: 
Core process 4524 exited. 3 remaining
Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]: info: 
Core process 4523 exited. 2 remaining
Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]: info: 
Core process 4522 exited. 1 remaining
Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]: info: 
dbsuat1a.intranet.mydomain.com Heartbeat shutdown complete.


my ha.cf
# Logging
debug                          1
debugfile             /var/log/ha-debug
logfile             /var/log/ha-log
logfacility             local0
#use_logd                       true
#logfacility                    daemon

# Misc Options
traditional_compression        off
compression                    bz2
coredumps                      true

# Communications
udpport                        691
bcast                          eth0
##autojoin                     any
autojoin                       none

# Thresholds (in seconds)
keepalive                      1
warntime                       6
deadtime                       10
initdead                       15

node dbsuat1a.intranet.mydomain.com
node dbsuat1b.intranet.mydomain.com
#enable pacemaker
crm yes
#enable STONITH
#crm respawn

my haresources:
DBSUAT1A.intranet.mydomain.com 172.28.185.49


_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to