Re: [ClusterLabs] Pacemaker fatal shutdown

2023-07-25 Thread Ken Gaillot
On Thu, 2023-07-20 at 12:43 +0530, Priyanka Balotra wrote:
> What I mainly want to understand is that:
> - why "fatal failure" is coming 

The logs so far don't show that. The earliest sign is:

Jul 17 14:18:20.085 FILE-6 pacemaker-fenced[19411]
(remote_op_done)   notice: Operation 'reboot' targeting FILE-2 by FILE-
4 for pacemaker-controld.19415@FILE-6: OK | id=4e523b34

You'd want to figure out which node was the Designated Controller (DC)
at that time, and look at its logs before this time. The DC will have
"Calculated transition" log messages.

You want to find such messages just before the timestamp above. If you
look above the "Calculated transition" message, it will show what
actions the cluster wants to take, including fencing. The logs around
there should say why the fencing was needed.

> - why does pacemaker not start on the node after a node boots
> followed by  "pacemaker fatal failure" .

A fatal failure is one where Pacemaker should stay down, so that's what
it does. In this case, fencing completed against the node, but the node
was still alive, so it shuts down and waits for manual intervention to
figure out what happened.

> - How can this be handled?

In a situation like this, figure out (1) why fencing was needed and (2)
why successful fencing did not kill the node (if you're using fabric
fencing such as SCSI fencing, that could be a reason, otherwise it
might be a misconfiguration).

Once you know that, it should be fairly obvious what to do about it,
and once it's taken care of, you can manually start Pacemaker on the
node again.

> 
> Thanks
> Priyanka
> 
> On Thu, Jul 20, 2023 at 12:41 PM Priyanka Balotra <
> priyanka.14balo...@gmail.com> wrote:
> > Hi, 
> > 
> > Here are FILE-6 logs: 
> > 
> > 65710:Jul 17 14:16:51.517 FILE-6 pacemaker-controld  [19415]
> > (throttle_mode)debug: Current load is 0.76 across 10
> > core(s)
> > 65711:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (throttle_update)  debug: Node FILE-2 has negligible load and
> > supports at most 20 jobs; new job limit 20
> > 65712:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (handle_request)   debug: The throttle changed. Trigger a graph.
> > 65713:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (pcmk__set_flags_as)   debug: FSA action flags 0x0002
> > (new_actions) for controller set by s_crmd_fsa:198
> > 65714:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (s_crmd_fsa)   debug: Processing I_JOIN_REQUEST: [
> > state=S_INTEGRATION cause=C_HA_MESSAGE origin=route_message ]
> > 65715:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (pcmk__clear_flags_as) debug: FSA action flags 0x0002
> > (an_action) for controller cleared by do_fsa_action:108
> > 65716:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (do_dc_join_filter_offer)  debug: Accepting join-1 request from
> > FILE-2 | ref=join_request-crmd-1689603392-8
> > 65717:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (pcmk__update_peer_expected)   info: do_dc_join_filter_offer:
> > Node FILE-2[2] - expected state is now member (was (null))
> > 65718:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (do_dc_join_filter_offer)  debug: 2 nodes currently integrated in
> > join-1
> > 65719:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (check_join_state) debug: join-1: Integration of 2 peers
> > complete | state=S_INTEGRATION for=do_dc_join_filter_offer
> > 65720:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (pcmk__set_flags_as)   debug: FSA action flags 0x0004
> > (new_actions) for controller set by s_crmd_fsa:198
> > 65721:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (s_crmd_fsa)   debug: Processing I_INTEGRATED: [
> > state=S_INTEGRATION cause=C_FSA_INTERNAL origin=check_join_state ]
> > 65722:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (do_state_transition)  info: State transition S_INTEGRATION ->
> > S_FINALIZE_JOIN | input=I_INTEGRATED cause=C_FSA_INTERNAL
> > origin=check_join_state
> > 65723:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (pcmk__set_flags_as)   debug: FSA action flags 0x0020
> > (A_INTEGRATE_TIMER_STOP) for controller set by
> > do_state_transition:559
> > 65724:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (pcmk__set_flags_as)   debug: FSA action flags 0x0040
> > (A_FINALIZE_TIMER_START) for controller set by
> > do_state_transition:563
> > 65725:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (pcmk__set_flags_as)   debug: FSA action flags 0x0200
> > (A_DC_TIMER_STOP) for controller set by do_state_transition:569
> > 65726:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (do_state_transition)  debug: All cluster nodes (2) responded
> > to join offer
> > 65727:Jul 17 14:16:55.085 FILE-6 pacemaker-controld  [19415]
> > (pcmk__clear_flags_as) debug: FSA action flags 0x0200
>

Re: [ClusterLabs] Pacemaker fatal shutdown

2023-07-19 Thread Reid Wahl
On Wed, Jul 19, 2023 at 8:33 PM Priyanka Balotra
 wrote:
>
> Sure,
> Here are the logs:
>
>
> 63138:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (post_cache_update)debug: Updated cache after membership event 44.
> 63139:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__set_flags_as)   debug: FSA action flags 0x2 
> (A_ELECTION_CHECK) for controller set by post_cache_update:81
> 63140:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__clear_flags_as) debug: FSA action flags 0x0002 (an_action) for 
> controller cleared by do_fsa_action:108
> 63141:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] (do_started) 
>   info: Delaying start, Config not read (0040)
> 63142:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (register_fsa_input_adv)   debug: Stalling the FSA pending further input: 
> source=do_started cause=C_FSA_INTERNAL data=(nil) queue=0
> 63143:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__set_flags_as)   debug: FSA action flags 0x0002 (with_actions) 
> for controller set by register_fsa_input_adv:88
> 63144:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] (s_crmd_fsa) 
>   debug: Exiting the FSA: queue=0, fsa_actions=0x20002, stalled=true
> 63145:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (config_query_callback)debug: Call 3 : Parsing CIB options
> 63146:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (config_query_callback)debug: Shutdown escalation occurs if DC has not 
> responded to request in 120ms
> 63147:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (config_query_callback)debug: Re-run scheduler after 90ms of 
> inactivity
> 63148:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pe_unpack_alerts) debug: Alert pf-ha-alert: 
> path=/usr/lib/ocf/resource.d/pacemaker/pf_ha_alert.sh timeout=3ms 
> tstamp-format='%H:%M:%S.%06N' 0 vars
> 63149:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__clear_flags_as) debug: FSA action flags 0x0002 (an_action) for 
> controller cleared by do_fsa_action:108
> 63150:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] (do_started) 
>   debug: Init server comms
> 63151:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (qb_ipcs_us_publish)   info: server name: crmd
> 63152:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] (do_started) 
>   notice: Pacemaker controller successfully started and accepting connections
> 63153:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__clear_flags_as) debug: FSA action flags 0x2 (an_action) 
> for controller cleared by do_fsa_action:108
> 63154:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (do_election_check)debug: Ignoring election check because we are not 
> in an election
> 63155:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__set_flags_as)   debug: FSA action flags 0x10100100 
> (new_actions) for controller set by s_crmd_fsa:198
> 63156:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] (s_crmd_fsa) 
>   debug: Processing I_PENDING: [ state=S_STARTING cause=C_FSA_INTERNAL 
> origin=do_started ]
> 63157:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__clear_flags_as) debug: FSA action flags 0x1000 
> (an_action) for controller cleared by do_fsa_action:108
> 63158:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] (do_log)   info: 
> Input I_PENDING received in state S_STARTING from do_started
> 63159:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (do_state_transition)  notice: State transition S_STARTING -> S_PENDING | 
> input=I_PENDING cause=C_FSA_INTERNAL origin=do_started
> 63160:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__set_flags_as)   debug: FSA action flags 0x0020 
> (A_INTEGRATE_TIMER_STOP) for controller set by do_state_transition:559
> 63161:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__set_flags_as)   debug: FSA action flags 0x0080 
> (A_FINALIZE_TIMER_STOP) for controller set by do_state_transition:565
> 63162:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__clear_flags_as) debug: FSA action flags 0x0020 (an_action) for 
> controller cleared by do_fsa_action:108
> 63163:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__clear_flags_as) debug: FSA action flags 0x0080 (an_action) for 
> controller cleared by do_fsa_action:108
> 63164:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__clear_flags_as) debug: FSA action flags 0x0010 (an_action) for 
> controller cleared by do_fsa_action:108
> 63165:Jul 17 14:16:26.132 FILE-2 pacemaker-controld  [15962] 
> (do_cl_join_query) debug: Querying for a DC
> 63166:Jul 17 14:16:26.132 FILE-2 pacemaker-controld  [15962] 
> (pcmk__clear_flags_as)  

Re: [ClusterLabs] Pacemaker fatal shutdown

2023-07-19 Thread Priyanka Balotra
Sure,
Here are the logs:


63138:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(post_cache_update)debug: Updated cache after membership event 44.
63139:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pcmk__set_flags_as)   debug: FSA action flags 0x2
(A_ELECTION_CHECK) for controller set by post_cache_update:81
63140:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pcmk__clear_flags_as) debug: FSA action flags 0x0002 (an_action)
for controller cleared by do_fsa_action:108
63141:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] (do_started)
info: Delaying start, Config not read (0040)
63142:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(register_fsa_input_adv)   debug: Stalling the FSA pending further input:
source=do_started cause=C_FSA_INTERNAL data=(nil) queue=0
63143:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pcmk__set_flags_as)   debug: FSA action flags 0x0002
(with_actions) for controller set by register_fsa_input_adv:88
63144:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] (s_crmd_fsa)
debug: Exiting the FSA: queue=0, fsa_actions=0x20002, stalled=true
63145:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(config_query_callback)debug: Call 3 : Parsing CIB options
63146:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(config_query_callback)debug: Shutdown escalation occurs if DC has not
responded to request in 120ms
63147:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(config_query_callback)debug: Re-run scheduler after 90ms of
inactivity
63148:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pe_unpack_alerts) debug: Alert pf-ha-alert:
path=/usr/lib/ocf/resource.d/pacemaker/pf_ha_alert.sh timeout=3ms
tstamp-format='%H:%M:%S.%06N' 0 vars
63149:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pcmk__clear_flags_as) debug: FSA action flags 0x0002 (an_action)
for controller cleared by do_fsa_action:108
63150:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] (do_started)
debug: Init server comms
63151:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(qb_ipcs_us_publish)   info: server name: crmd
63152:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] (do_started)
notice: Pacemaker controller successfully started and accepting
connections
63153:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pcmk__clear_flags_as) debug: FSA action flags 0x2 (an_action)
for controller cleared by do_fsa_action:108
63154:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(do_election_check)debug: Ignoring election check because we are
not in an election
63155:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pcmk__set_flags_as)   debug: FSA action flags 0x10100100
(new_actions) for controller set by s_crmd_fsa:198
63156:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] (s_crmd_fsa)
debug: Processing I_PENDING: [ state=S_STARTING cause=C_FSA_INTERNAL
origin=do_started ]
63157:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pcmk__clear_flags_as) debug: FSA action flags 0x1000
(an_action) for controller cleared by do_fsa_action:108
63158:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962] (do_log)
info: Input I_PENDING received in state S_STARTING from do_started
63159:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(do_state_transition)  notice: State transition S_STARTING -> S_PENDING
| input=I_PENDING cause=C_FSA_INTERNAL origin=do_started
63160:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pcmk__set_flags_as)   debug: FSA action flags 0x0020
(A_INTEGRATE_TIMER_STOP) for controller set by do_state_transition:559
63161:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pcmk__set_flags_as)   debug: FSA action flags 0x0080
(A_FINALIZE_TIMER_STOP) for controller set by do_state_transition:565
63162:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pcmk__clear_flags_as) debug: FSA action flags 0x0020 (an_action)
for controller cleared by do_fsa_action:108
63163:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pcmk__clear_flags_as) debug: FSA action flags 0x0080 (an_action)
for controller cleared by do_fsa_action:108
63164:Jul 17 14:16:25.132 FILE-2 pacemaker-controld  [15962]
(pcmk__clear_flags_as) debug: FSA action flags 0x0010 (an_action)
for controller cleared by do_fsa_action:108
63165:Jul 17 14:16:26.132 FILE-2 pacemaker-controld  [15962]
(do_cl_join_query) debug: Querying for a DC
63166:Jul 17 14:16:26.132 FILE-2 pacemaker-controld  [15962]
(pcmk__clear_flags_as) debug: FSA action flags 0x0100 (an_action)
for controller cleared by do_fsa_action:108
63167:Jul 17 14:16:26.132 FILE-2 pacemaker-controld  [15962]
(controld_start_timer) debug: Started Election Trigger (inject
I_DC_TIMEOUT if pops after 2ms, source=18)
63168:Jul 17 

Re: [ClusterLabs] Pacemaker fatal shutdown

2023-07-19 Thread Ken Gaillot
On Wed, 2023-07-19 at 23:49 +0530, Priyanka Balotra wrote:
> Hi All, 
> I am using SLES 15 SP4. One of the nodes of the cluster is brought
> down and boot up after sometime. Pacemaker service came up first but
> later it faced a fatal shutdown. Due to that crm service is down. 
> 
> The logs from /var/log/pacemaker.pacemaker.log are as follows:
> 
> Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956]
> (pcmk_child_exit)warning: Shutting cluster down because
> pacemaker-controld[15962] had fatal failure

The interesting messages will be before this. The ones with "pacemaker-
controld" will be the most relevant, at least initially.

> Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956]
> (pcmk_shutdown_worker)   notice: Shutting down Pacemaker
> Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956]
> (pcmk_shutdown_worker)   debug: pacemaker-controld confirmed stopped
> Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956] (stop_child)  
>   notice: Stopping pacemaker-schedulerd | sent signal 15 to process
> 15961
> Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961]
> (crm_signal_dispatch)notice: Caught 'Terminated' signal | 15
> (invoking handler)
> Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961]
> (qb_ipcs_us_withdraw)info: withdrawing server sockets
> Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961]
> (qb_ipcs_unref)  debug: qb_ipcs_unref() - destroying
> Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961]
> (crm_xml_cleanup)info: Cleaning up memory from libxml2
> Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] (crm_exit)
>   info: Exiting pacemaker-schedulerd | with status 0
> Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957]
> (qb_ipcs_event_sendv)debug: new_event_notification (/dev/shm/qb-
> 15957-15962-12-RDPw6O/qb): Broken pipe (32)
> Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957]
> (cib_notify_send_one)warning: Could not notify client crmd:
> Broken pipe | id=e29d175e-7e91-4b6a-bffb-fabfdd7a33bf
> Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957]
> (cib_process_request)info: Completed cib_delete operation for
> section //node_state[@uname='FILE-2']/*: OK (rc=0, origin=FILE-
> 6/crmd/74, version=0.24.75)
> Jul 17 14:18:20.093 FILE-2 pacemaker-fenced[15958]
> (xml_patch_version_check)debug: Can apply patch 0.24.75 to
> 0.24.74
> Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956]
> (pcmk_child_exit)info: pacemaker-schedulerd[15961] exited
> with status 0 (OK)
> Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957]
> (cib_process_request)info: Completed cib_modify operation for
> section status: OK (rc=0, origin=FILE-6/crmd/75, version=0.24.75)
> Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956]
> (pcmk_shutdown_worker)   debug: pacemaker-schedulerd confirmed
> stopped
> Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956] (stop_child)  
>   notice: Stopping pacemaker-attrd | sent signal 15 to process 15960
> Jul 17 14:18:20.093 FILE-2 pacemaker-attrd [15960]
> (crm_signal_dispatch)notice: Caught 'Terminated' signal | 15
> (invoking handler)
> 
> Could you please help me understand the issue here.
> 
> Regards
> Priyanka
> ___
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
> 
> ClusterLabs home: https://www.clusterlabs.org/
-- 
Ken Gaillot 

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] Pacemaker fatal shutdown

2023-07-19 Thread Priyanka Balotra
Hi All,
I am using SLES 15 SP4. One of the nodes of the cluster is brought down and
boot up after sometime. Pacemaker service came up first but later it faced
a fatal shutdown. Due to that crm service is down.

The logs from /var/log/pacemaker.pacemaker.log are as follows:

Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956] (pcmk_child_exit)
 warning: Shutting cluster down because pacemaker-controld[15962] had
fatal failure
Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956]
(pcmk_shutdown_worker)   notice: Shutting down Pacemaker
Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956]
(pcmk_shutdown_worker)   debug: pacemaker-controld confirmed stopped
Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956] (stop_child)
notice: Stopping pacemaker-schedulerd | sent signal 15 to process 15961
Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961]
(crm_signal_dispatch)notice: Caught 'Terminated' signal | 15 (invoking
handler)
Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961]
(qb_ipcs_us_withdraw)info: withdrawing server sockets
Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] (qb_ipcs_unref)
 debug: qb_ipcs_unref() - destroying
Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] (crm_xml_cleanup)
 info: Cleaning up memory from libxml2
Jul 17 14:18:20.093 FILE-2 pacemaker-schedulerd[15961] (crm_exit)
info: Exiting pacemaker-schedulerd | with status 0
Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957]
(qb_ipcs_event_sendv)debug: new_event_notification
(/dev/shm/qb-15957-15962-12-RDPw6O/qb): Broken pipe (32)
Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957]
(cib_notify_send_one)warning: Could not notify client crmd: Broken pipe
| id=e29d175e-7e91-4b6a-bffb-fabfdd7a33bf
Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957]
(cib_process_request)info: Completed cib_delete operation for section
//node_state[@uname='FILE-2']/*: OK (rc=0, origin=FILE-6/crmd/74,
version=0.24.75)
Jul 17 14:18:20.093 FILE-2 pacemaker-fenced[15958]
(xml_patch_version_check)debug: Can apply patch 0.24.75 to 0.24.74
Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956] (pcmk_child_exit)
 info: pacemaker-schedulerd[15961] exited with status 0 (OK)
Jul 17 14:18:20.093 FILE-2 pacemaker-based [15957]
(cib_process_request)info: Completed cib_modify operation for section
status: OK (rc=0, origin=FILE-6/crmd/75, version=0.24.75)
Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956]
(pcmk_shutdown_worker)   debug: pacemaker-schedulerd confirmed stopped
Jul 17 14:18:20.093 FILE-2 pacemakerd  [15956] (stop_child)
notice: Stopping pacemaker-attrd | sent signal 15 to process 15960
Jul 17 14:18:20.093 FILE-2 pacemaker-attrd [15960]
(crm_signal_dispatch)notice: Caught 'Terminated' signal | 15 (invoking
handler)

Could you please help me understand the issue here.

Regards
Priyanka
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/