Re: [ClusterLabs] Pacemaker fatal shutdown
On Thu, 2023-07-20 at 12:43 +0530, Priyanka Balotra wrote: > What I mainly want to understand is that: > - why "fatal failure" is coming The logs so far don't show that. The earliest sign is: Jul 17 14:18:20.085 FILE-6 pacemaker-fenced[19411] (remote_op_done) notice: Operation 'reboot' targeting FILE-2 by FILE- 4 for pacemaker-controld.19415@FILE-6: OK | id=4e523b34 You'd want to figure out which node was the Designated Controller (DC) at that time, and look at its logs before this time. The DC will have "Calculated transition" log messages. You want to find such messages just before the timestamp above. If you look above the "Calculated transition" message, it will show what actions the cluster wants to take, including fencing. The logs around there should say why the fencing was needed. > - why does pacemaker not start on the node after a node boots > followed by "pacemaker fatal failure" . A fatal failure is one where Pacemaker should stay down, so that's what it does. In this case, fencing completed against the node, but the node was still alive, so it shuts down and waits for manual intervention to figure out what happened. > - How can this be handled? In a situation like this, figure out (1) why fencing was needed and (2) why successful fencing did not kill the node (if you're using fabric fencing such as SCSI fencing, that could be a reason, otherwise it might be a misconfiguration). Once you know that, it should be fairly obvious what to do about it, and once it's taken care of, you can manually start Pacemaker on the node again. > > Thanks > Priyanka > > On Thu, Jul 20, 2023 at 12:41 PM Priyanka Balotra < > priyanka.14balo...@gmail.com> wrote: > > Hi, > > > > Here are FILE-6 logs: > > > > 65710:Jul 17 14:16:51.517 FILE-6 pacemaker-controld [19415] > > (throttle_mode)debug: Current load is 0.76 across 10 > > core(s) > > 65711:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (throttle_update) debug: Node FILE-2 has negligible load and > > supports at most 20 jobs; new job limit 20 > > 65712:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (handle_request) debug: The throttle changed. Trigger a graph. > > 65713:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (pcmk__set_flags_as) debug: FSA action flags 0x0002 > > (new_actions) for controller set by s_crmd_fsa:198 > > 65714:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (s_crmd_fsa) debug: Processing I_JOIN_REQUEST: [ > > state=S_INTEGRATION cause=C_HA_MESSAGE origin=route_message ] > > 65715:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (pcmk__clear_flags_as) debug: FSA action flags 0x0002 > > (an_action) for controller cleared by do_fsa_action:108 > > 65716:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (do_dc_join_filter_offer) debug: Accepting join-1 request from > > FILE-2 | ref=join_request-crmd-1689603392-8 > > 65717:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (pcmk__update_peer_expected) info: do_dc_join_filter_offer: > > Node FILE-2[2] - expected state is now member (was (null)) > > 65718:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (do_dc_join_filter_offer) debug: 2 nodes currently integrated in > > join-1 > > 65719:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (check_join_state) debug: join-1: Integration of 2 peers > > complete | state=S_INTEGRATION for=do_dc_join_filter_offer > > 65720:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (pcmk__set_flags_as) debug: FSA action flags 0x0004 > > (new_actions) for controller set by s_crmd_fsa:198 > > 65721:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (s_crmd_fsa) debug: Processing I_INTEGRATED: [ > > state=S_INTEGRATION cause=C_FSA_INTERNAL origin=check_join_state ] > > 65722:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (do_state_transition) info: State transition S_INTEGRATION -> > > S_FINALIZE_JOIN | input=I_INTEGRATED cause=C_FSA_INTERNAL > > origin=check_join_state > > 65723:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (pcmk__set_flags_as) debug: FSA action flags 0x0020 > > (A_INTEGRATE_TIMER_STOP) for controller set by > > do_state_transition:559 > > 65724:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (pcmk__set_flags_as) debug: FSA action flags 0x0040 > > (A_FINALIZE_TIMER_START) for controller set by > > do_state_transition:563 > > 65725:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (pcmk__set_flags_as) debug: FSA action flags 0x0200 > > (A_DC_TIMER_STOP) for controller set by do_state_transition:569 > > 65726:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (do_state_transition) debug: All cluster nodes (2) responded > > to join offer > > 65727:Jul 17 14:16:55.085 FILE-6 pacemaker-controld [19415] > > (pcmk__clear_flags_as) debug: FSA action flags 0x0200 >
Re: [ClusterLabs] Pacemaker fatal shutdown
On Wed, Jul 19, 2023 at 8:33 PM Priyanka Balotra wrote: > > Sure, > Here are the logs: > > > 63138:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (post_cache_update)debug: Updated cache after membership event 44. > 63139:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pcmk__set_flags_as) debug: FSA action flags 0x2 > (A_ELECTION_CHECK) for controller set by post_cache_update:81 > 63140:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pcmk__clear_flags_as) debug: FSA action flags 0x0002 (an_action) for > controller cleared by do_fsa_action:108 > 63141:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (do_started) > info: Delaying start, Config not read (0040) > 63142:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (register_fsa_input_adv) debug: Stalling the FSA pending further input: > source=do_started cause=C_FSA_INTERNAL data=(nil) queue=0 > 63143:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pcmk__set_flags_as) debug: FSA action flags 0x0002 (with_actions) > for controller set by register_fsa_input_adv:88 > 63144:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (s_crmd_fsa) > debug: Exiting the FSA: queue=0, fsa_actions=0x20002, stalled=true > 63145:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (config_query_callback)debug: Call 3 : Parsing CIB options > 63146:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (config_query_callback)debug: Shutdown escalation occurs if DC has not > responded to request in 120ms > 63147:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (config_query_callback)debug: Re-run scheduler after 90ms of > inactivity > 63148:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pe_unpack_alerts) debug: Alert pf-ha-alert: > path=/usr/lib/ocf/resource.d/pacemaker/pf_ha_alert.sh timeout=3ms > tstamp-format='%H:%M:%S.%06N' 0 vars > 63149:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pcmk__clear_flags_as) debug: FSA action flags 0x0002 (an_action) for > controller cleared by do_fsa_action:108 > 63150:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (do_started) > debug: Init server comms > 63151:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (qb_ipcs_us_publish) info: server name: crmd > 63152:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (do_started) > notice: Pacemaker controller successfully started and accepting connections > 63153:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pcmk__clear_flags_as) debug: FSA action flags 0x2 (an_action) > for controller cleared by do_fsa_action:108 > 63154:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (do_election_check)debug: Ignoring election check because we are not > in an election > 63155:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pcmk__set_flags_as) debug: FSA action flags 0x10100100 > (new_actions) for controller set by s_crmd_fsa:198 > 63156:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (s_crmd_fsa) > debug: Processing I_PENDING: [ state=S_STARTING cause=C_FSA_INTERNAL > origin=do_started ] > 63157:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pcmk__clear_flags_as) debug: FSA action flags 0x1000 > (an_action) for controller cleared by do_fsa_action:108 > 63158:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (do_log) info: > Input I_PENDING received in state S_STARTING from do_started > 63159:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (do_state_transition) notice: State transition S_STARTING -> S_PENDING | > input=I_PENDING cause=C_FSA_INTERNAL origin=do_started > 63160:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pcmk__set_flags_as) debug: FSA action flags 0x0020 > (A_INTEGRATE_TIMER_STOP) for controller set by do_state_transition:559 > 63161:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pcmk__set_flags_as) debug: FSA action flags 0x0080 > (A_FINALIZE_TIMER_STOP) for controller set by do_state_transition:565 > 63162:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pcmk__clear_flags_as) debug: FSA action flags 0x0020 (an_action) for > controller cleared by do_fsa_action:108 > 63163:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pcmk__clear_flags_as) debug: FSA action flags 0x0080 (an_action) for > controller cleared by do_fsa_action:108 > 63164:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] > (pcmk__clear_flags_as) debug: FSA action flags 0x0010 (an_action) for > controller cleared by do_fsa_action:108 > 63165:Jul 17 14:16:26.132 FILE-2 pacemaker-controld [15962] > (do_cl_join_query) debug: Querying for a DC > 63166:Jul 17 14:16:26.132 FILE-2 pacemaker-controld [15962] > (pcmk__clear_flags_as)
Re: [ClusterLabs] Pacemaker fatal shutdown
Sure, Here are the logs: 63138:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (post_cache_update)debug: Updated cache after membership event 44. 63139:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pcmk__set_flags_as) debug: FSA action flags 0x2 (A_ELECTION_CHECK) for controller set by post_cache_update:81 63140:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pcmk__clear_flags_as) debug: FSA action flags 0x0002 (an_action) for controller cleared by do_fsa_action:108 63141:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (do_started) info: Delaying start, Config not read (0040) 63142:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (register_fsa_input_adv) debug: Stalling the FSA pending further input: source=do_started cause=C_FSA_INTERNAL data=(nil) queue=0 63143:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pcmk__set_flags_as) debug: FSA action flags 0x0002 (with_actions) for controller set by register_fsa_input_adv:88 63144:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (s_crmd_fsa) debug: Exiting the FSA: queue=0, fsa_actions=0x20002, stalled=true 63145:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (config_query_callback)debug: Call 3 : Parsing CIB options 63146:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (config_query_callback)debug: Shutdown escalation occurs if DC has not responded to request in 120ms 63147:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (config_query_callback)debug: Re-run scheduler after 90ms of inactivity 63148:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pe_unpack_alerts) debug: Alert pf-ha-alert: path=/usr/lib/ocf/resource.d/pacemaker/pf_ha_alert.sh timeout=3ms tstamp-format='%H:%M:%S.%06N' 0 vars 63149:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pcmk__clear_flags_as) debug: FSA action flags 0x0002 (an_action) for controller cleared by do_fsa_action:108 63150:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (do_started) debug: Init server comms 63151:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (qb_ipcs_us_publish) info: server name: crmd 63152:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (do_started) notice: Pacemaker controller successfully started and accepting connections 63153:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pcmk__clear_flags_as) debug: FSA action flags 0x2 (an_action) for controller cleared by do_fsa_action:108 63154:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (do_election_check)debug: Ignoring election check because we are not in an election 63155:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pcmk__set_flags_as) debug: FSA action flags 0x10100100 (new_actions) for controller set by s_crmd_fsa:198 63156:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (s_crmd_fsa) debug: Processing I_PENDING: [ state=S_STARTING cause=C_FSA_INTERNAL origin=do_started ] 63157:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pcmk__clear_flags_as) debug: FSA action flags 0x1000 (an_action) for controller cleared by do_fsa_action:108 63158:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (do_log) info: Input I_PENDING received in state S_STARTING from do_started 63159:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (do_state_transition) notice: State transition S_STARTING -> S_PENDING | input=I_PENDING cause=C_FSA_INTERNAL origin=do_started 63160:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pcmk__set_flags_as) debug: FSA action flags 0x0020 (A_INTEGRATE_TIMER_STOP) for controller set by do_state_transition:559 63161:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pcmk__set_flags_as) debug: FSA action flags 0x0080 (A_FINALIZE_TIMER_STOP) for controller set by do_state_transition:565 63162:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pcmk__clear_flags_as) debug: FSA action flags 0x0020 (an_action) for controller cleared by do_fsa_action:108 63163:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pcmk__clear_flags_as) debug: FSA action flags 0x0080 (an_action) for controller cleared by do_fsa_action:108 63164:Jul 17 14:16:25.132 FILE-2 pacemaker-controld [15962] (pcmk__clear_flags_as) debug: FSA action flags 0x0010 (an_action) for controller cleared by do_fsa_action:108 63165:Jul 17 14:16:26.132 FILE-2 pacemaker-controld [15962] (do_cl_join_query) debug: Querying for a DC 63166:Jul 17 14:16:26.132 FILE-2 pacemaker-controld [15962] (pcmk__clear_flags_as) debug: FSA action flags 0x0100 (an_action) for controller cleared by do_fsa_action:108 63167:Jul 17 14:16:26.132 FILE-2 pacemaker-controld [15962] (controld_start_timer) debug: Started Election Trigger (inject I_DC_TIMEOUT if pops after 2ms, source=18) 63168:Jul 17
Re: [ClusterLabs] Pacemaker fatal shutdown
On Wed, 2023-07-19 at 23:49 +0530, Priyanka Balotra wrote: > Hi All, > I am using SLES 15 SP4. One of the nodes of the cluster is brought > down and boot up after sometime. Pacemaker service came up first but > later it faced a fatal shutdown. Due to that crm service is down. > > The logs from /var/log/pacemaker.pacemaker.log are as follows: > > Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] > (pcmk_child_exit)warning: Shutting cluster down because > pacemaker-controld[15962] had fatal failure The interesting messages will be before this. The ones with "pacemaker- controld" will be the most relevant, at least initially. > Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] > (pcmk_shutdown_worker) notice: Shutting down Pacemaker > Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] > (pcmk_shutdown_worker) debug: pacemaker-controld confirmed stopped > Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] (stop_child) > notice: Stopping pacemaker-schedulerd | sent signal 15 to process > 15961 > Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] > (crm_signal_dispatch)notice: Caught 'Terminated' signal | 15 > (invoking handler) > Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] > (qb_ipcs_us_withdraw)info: withdrawing server sockets > Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] > (qb_ipcs_unref) debug: qb_ipcs_unref() - destroying > Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] > (crm_xml_cleanup)info: Cleaning up memory from libxml2 > Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] (crm_exit) > info: Exiting pacemaker-schedulerd | with status 0 > Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957] > (qb_ipcs_event_sendv)debug: new_event_notification (/dev/shm/qb- > 15957-15962-12-RDPw6O/qb): Broken pipe (32) > Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957] > (cib_notify_send_one)warning: Could not notify client crmd: > Broken pipe | id=e29d175e-7e91-4b6a-bffb-fabfdd7a33bf > Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957] > (cib_process_request)info: Completed cib_delete operation for > section //node_state[@uname='FILE-2']/*: OK (rc=0, origin=FILE- > 6/crmd/74, version=0.24.75) > Jul 17 14:18:20.093 FILE-2 pacemaker-fenced[15958] > (xml_patch_version_check)debug: Can apply patch 0.24.75 to > 0.24.74 > Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] > (pcmk_child_exit)info: pacemaker-schedulerd[15961] exited > with status 0 (OK) > Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957] > (cib_process_request)info: Completed cib_modify operation for > section status: OK (rc=0, origin=FILE-6/crmd/75, version=0.24.75) > Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] > (pcmk_shutdown_worker) debug: pacemaker-schedulerd confirmed > stopped > Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] (stop_child) > notice: Stopping pacemaker-attrd | sent signal 15 to process 15960 > Jul 17 14:18:20.093 FILE-2 pacemaker-attrd [15960] > (crm_signal_dispatch)notice: Caught 'Terminated' signal | 15 > (invoking handler) > > Could you please help me understand the issue here. > > Regards > Priyanka > ___ > Manage your subscription: > https://lists.clusterlabs.org/mailman/listinfo/users > > ClusterLabs home: https://www.clusterlabs.org/ -- Ken Gaillot ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] Pacemaker fatal shutdown
Hi All, I am using SLES 15 SP4. One of the nodes of the cluster is brought down and boot up after sometime. Pacemaker service came up first but later it faced a fatal shutdown. Due to that crm service is down. The logs from /var/log/pacemaker.pacemaker.log are as follows: Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] (pcmk_child_exit) warning: Shutting cluster down because pacemaker-controld[15962] had fatal failure Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] (pcmk_shutdown_worker) notice: Shutting down Pacemaker Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] (pcmk_shutdown_worker) debug: pacemaker-controld confirmed stopped Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] (stop_child) notice: Stopping pacemaker-schedulerd | sent signal 15 to process 15961 Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] (crm_signal_dispatch)notice: Caught 'Terminated' signal | 15 (invoking handler) Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] (qb_ipcs_us_withdraw)info: withdrawing server sockets Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] (qb_ipcs_unref) debug: qb_ipcs_unref() - destroying Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] (crm_xml_cleanup) info: Cleaning up memory from libxml2 Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] (crm_exit) info: Exiting pacemaker-schedulerd | with status 0 Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957] (qb_ipcs_event_sendv)debug: new_event_notification (/dev/shm/qb-15957-15962-12-RDPw6O/qb): Broken pipe (32) Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957] (cib_notify_send_one)warning: Could not notify client crmd: Broken pipe | id=e29d175e-7e91-4b6a-bffb-fabfdd7a33bf Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957] (cib_process_request)info: Completed cib_delete operation for section //node_state[@uname='FILE-2']/*: OK (rc=0, origin=FILE-6/crmd/74, version=0.24.75) Jul 17 14:18:20.093 FILE-2 pacemaker-fenced[15958] (xml_patch_version_check)debug: Can apply patch 0.24.75 to 0.24.74 Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] (pcmk_child_exit) info: pacemaker-schedulerd[15961] exited with status 0 (OK) Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957] (cib_process_request)info: Completed cib_modify operation for section status: OK (rc=0, origin=FILE-6/crmd/75, version=0.24.75) Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] (pcmk_shutdown_worker) debug: pacemaker-schedulerd confirmed stopped Jul 17 14:18:20.093 FILE-2 pacemakerd [15956] (stop_child) notice: Stopping pacemaker-attrd | sent signal 15 to process 15960 Jul 17 14:18:20.093 FILE-2 pacemaker-attrd [15960] (crm_signal_dispatch)notice: Caught 'Terminated' signal | 15 (invoking handler) Could you please help me understand the issue here. Regards Priyanka ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/