thank you for your reply. The network is well, firewall has been closed, and selinux has been disabled. /var/log/pacemaker.log as follows: Aug 16 13:11:12 [8980] node0 crmd: info: peer_update_callback: node0 is now (null) Aug 16 13:11:12 [8980] node0 crmd: info: crm_get_peer: Node 1 has uuid 1 Aug 16 13:11:12 [8980] node0 crmd: info: crm_update_peer_proc: cluster_connect_cpg: Node node0[1] - corosync-cpg is now online Aug 16 13:11:12 [8980] node0 crmd: info: peer_update_callback: Client node0/peer now has status [online] (DC=<null>) Aug 16 13:11:12 [8980] node0 crmd: info: init_cs_connection_once: Connection to 'corosync': established Aug 16 13:11:12 [8980] node0 crmd: notice: cluster_connect_quorum: Quorum acquired Aug 16 13:11:12 [8980] node0 crmd: info: do_ha_control: Connected to the cluster Aug 16 13:11:12 [8980] node0 crmd: info: lrmd_ipc_connect: Connecting to lrmd Aug 16 13:11:12 [8977] node0 lrmd: info: crm_client_new: Connecting 0x16025a0 for uid=189 gid=189 pid=8980 id=e317bf62-fb55-46a1-af17-9284234917b8 Aug 16 13:11:12 [8975] node0 cib: info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/3, version=0.1.0) Aug 16 13:11:12 [8980] node0 crmd: info: do_lrm_control: LRM connection established Aug 16 13:11:12 [8980] node0 crmd: info: do_started: Delaying start, no membership data (0000000000100000) Aug 16 13:11:12 [8980] node0 crmd: info: do_started: Delaying start, no membership data (0000000000100000) Aug 16 13:11:12 [8975] node0 cib: info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/4, version=0.1.0) Aug 16 13:11:12 [8980] node0 crmd: info: pcmk_quorum_notification: Membership 8571476093872377196: quorum retained (1) Aug 16 13:11:12 [8980] node0 crmd: notice: crm_update_peer_state: pcmk_quorum_notification: Node node0[1] - state is now member (was (null)) Aug 16 13:11:12 [8980] node0 crmd: info: peer_update_callback: node0 is now member (was (null)) Aug 16 13:11:12 [8980] node0 crmd: notice: crm_update_peer_state: pcmk_quorum_notification: Node node0[1] - state is now lost (was member) Aug 16 13:11:12 [8980] node0 crmd: info: peer_update_callback: node0 is now lost (was member) Aug 16 13:11:12 [8980] node0 crmd: error: reap_dead_nodes: We're not part of the cluster anymore Aug 16 13:11:12 [8975] node0 cib: info: crm_client_new: Connecting 0x1cbcd80 for uid=0 gid=0 pid=8976 id=9afb683d-8390-408a-a690-fb5447fefd37 Aug 16 13:11:12 [8976] node0 stonith-ng: notice: setup_cib: Watching for stonith topology changes Aug 16 13:11:12 [8976] node0 stonith-ng: info: qb_ipcs_us_publish: server name: stonith-ng Aug 16 13:11:12 [8976] node0 stonith-ng: info: main: Starting stonith-ng mainloop Aug 16 13:11:12 [8976] node0 stonith-ng: info: pcmk_cpg_membership: Joined[0.0] stonith-ng.1 Aug 16 13:11:12 [8976] node0 stonith-ng: info: pcmk_cpg_membership: Member[0.0] stonith-ng.1 Aug 16 13:11:12 [8980] node0 crmd: error: do_log: FSA: Input I_ERROR from reap_dead_nodes() received in state S_STARTING Aug 16 13:11:12 [8980] node0 crmd: notice: do_state_transition: State transition S_STARTING -> S_RECOVERY [ input=I_ERROR cause=C_FSA_INTERNAL origin=reap_dead_nodes ] Aug 16 13:11:12 [8980] node0 crmd: warning: do_recover: Fast-tracking shutdown in response to errors Aug 16 13:11:12 [8980] node0 crmd: error: do_started: Start cancelled... S_RECOVERY Aug 16 13:11:12 [8980] node0 crmd: error: do_log: FSA: Input I_TERMINATE from do_recover() received in state S_RECOVERY Aug 16 13:11:12 [8980] node0 crmd: info: do_state_transition: State transition S_RECOVERY -> S_TERMINATE [ input=I_TERMINATE cause=C_FSA_INTERNAL origin=do_recover ] Aug 16 13:11:12 [8980] node0 crmd: notice: lrm_state_verify_stopped: Stopped 0 recurring operations at shutdown (0 ops remaining) Aug 16 13:11:12 [8980] node0 crmd: info: do_lrm_control: Disconnecting from the LRM Aug 16 13:11:12 [8980] node0 crmd: info: lrmd_api_disconnect: Disconnecting from lrmd service Aug 16 13:11:12 [8980] node0 crmd: info: lrmd_ipc_connection_destroy: IPC connection destroyed Aug 16 13:11:12 [8980] node0 crmd: info: lrm_connection_destroy: LRM Connection disconnected Aug 16 13:11:12 [8980] node0 crmd: info: lrmd_api_disconnect: Disconnecting from lrmd service Aug 16 13:11:12 [8980] node0 crmd: notice: do_lrm_control: Disconnected from the LRM Aug 16 13:11:12 [8980] node0 crmd: info: crm_cluster_disconnect: Disconnecting from cluster infrastructure: corosync Aug 16 13:11:12 [8980] node0 crmd: notice: terminate_cs_connection: Disconnecting from Corosync Aug 16 13:11:12 [8980] node0 crmd: info: crm_cluster_disconnect: Disconnected from corosync Aug 16 13:11:12 [8980] node0 crmd: info: do_ha_control: Disconnected from the cluster Aug 16 13:11:12 [8977] node0 lrmd: info: crm_client_destroy: Destroying 0 events Aug 16 13:11:12 [8980] node0 crmd: info: do_cib_control: Disconnecting CIB Aug 16 13:11:12 [8975] node0 cib: info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.1.0) Aug 16 13:11:12 [8976] node0 stonith-ng: info: init_cib_cache_cb: Updating device list from the cib: init Aug 16 13:11:12 [8980] node0 crmd: info: crmd_cib_connection_destroy: Connection to the CIB terminated... Aug 16 13:11:12 [8980] node0 crmd: info: do_exit: Performing A_EXIT_0 - gracefully exiting the CRMd Aug 16 13:11:12 [8980] node0 crmd: info: do_exit: [crmd] stopped (0) Aug 16 13:11:12 [8980] node0 crmd: info: crmd_exit: Dropping I_TERMINATE: [ state=S_TERMINATE cause=C_FSA_INTERNAL origin=do_stop ] Aug 16 13:11:12 [8980] node0 crmd: info: crmd_quorum_destroy: connection closed Aug 16 13:11:12 [8980] node0 crmd: info: crmd_cs_destroy: connection closed Aug 16 13:11:12 [8980] node0 crmd: info: crmd_init: 8980 stopped: OK (0) Aug 16 13:11:12 [8980] node0 crmd: error: crmd_fast_exit: Could not recover from internal error Aug 16 13:11:12 [8980] node0 crmd: info: crm_xml_cleanup: Cleaning up memory from libxml2 Aug 16 13:11:12 [8975] node0 cib: info: crm_client_destroy: Destroying 0 events Aug 16 13:11:12 [8974] node0 pacemakerd: error: pcmk_child_exit: Child process crmd (8980) exited: Generic Pacemaker error (201) Aug 16 13:11:12 [8974] node0 pacemakerd: error: pcmk_child_exit: Child process crmd (8980) exited: Generic Pacemaker error (201) Aug 16 13:11:12 [8974] node0 pacemakerd: notice: pcmk_process_exit: Respawning failed child process: crmd Aug 16 13:11:12 [8974] node0 pacemakerd: notice: pcmk_process_exit: Respawning failed child process: crmd Aug 16 13:11:12 [8974] node0 pacemakerd: info: start_child: Using uid=189 and group=189 for process crmd Aug 16 13:11:12 [8974] node0 pacemakerd: info: start_child: Using uid=189 and group=189 for process crmd Aug 16 13:11:12 [8974] node0 pacemakerd: info: start_child: Forked child 8982 for process crmd Aug 16 13:11:12 [8974] node0 pacemakerd: info: start_child: Forked child 8982 for process crmd Aug 16 13:11:12 [8982] node0 crmd: info: crm_log_init: Changed active directory to /var/lib/pacemaker/cores/hacluster Aug 16 13:11:12 [8982] node0 crmd: notice: main: CRM Git Version: 368c726 Aug 16 13:11:12 [8982] node0 crmd: info: do_log: FSA: Input I_STARTUP from crmd_init() received in state S_STARTING Aug 16 13:11:12 [8982] node0 crmd: info: get_cluster_type: Verifying cluster type: 'corosync' Aug 16 13:11:12 [8982] node0 crmd: info: get_cluster_type: Assuming an active 'corosync' cluster Aug 16 13:11:12 [8975] node0 cib: info: crm_client_new: Connecting 0x1a8dd80 for uid=189 gid=189 pid=8982 id=5e050be4-f112-4f32-831b-1704ce1872dd Aug 16 13:11:12 [8982] node0 crmd: info: do_cib_control: CIB connection established Aug 16 13:11:12 [8982] node0 crmd: notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync Aug 16 13:11:12 [8975] node0 cib: info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.1.0) Aug 16 13:11:12 [8982] node0 crmd: info: crm_get_peer: Created entry c01a1bd2-9f44-4061-9ea3-95c1df3de86d/0xea3c20 for node (null)/1 (1 total) Aug 16 13:11:12 [8982] node0 crmd: info: crm_get_peer: Node 1 is now known as node0 Aug 16 13:11:12 [8982] node0 crmd: info: peer_update_callback: node0 is now (null) Aug 16 13:11:12 [8982] node0 crmd: info: crm_get_peer: Node 1 has uuid 1 Aug 16 13:11:12 [8982] node0 crmd: info: crm_update_peer_proc: cluster_connect_cpg: Node node0[1] - corosync-cpg is now online Aug 16 13:11:12 [8982] node0 crmd: info: peer_update_callback: Client node0/peer now has status [online] (DC=<null>) Aug 16 13:11:12 [8982] node0 crmd: info: init_cs_connection_once: Connection to 'corosync': established Aug 16 13:11:12 [8982] node0 crmd: notice: cluster_connect_quorum: Quorum acquired Aug 16 13:11:12 [8975] node0 cib: info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/3, version=0.1.0) Aug 16 13:11:12 [8982] node0 crmd: info: do_ha_control: Connected to the cluster Aug 16 13:11:12 [8982] node0 crmd: info: lrmd_ipc_connect: Connecting to lrmd Aug 16 13:11:12 [8977] node0 lrmd: info: crm_client_new: Connecting 0x16025a0 for uid=189 gid=189 pid=8982 id=c6776327-b798-40f2-8ca4-a4b032cff2eb Aug 16 13:11:12 [8982] node0 crmd: info: do_lrm_control: LRM connection established Aug 16 13:11:12 [8982] node0 crmd: info: do_started: Delaying start, no membership data (0000000000100000) Aug 16 13:11:12 [8982] node0 crmd: info: do_started: Delaying start, no membership data (0000000000100000) Aug 16 13:11:12 [8975] node0 cib: info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/4, version=0.1.0) Aug 16 13:11:12 [8982] node0 crmd: info: pcmk_quorum_notification: Membership 8571476093872377196: quorum retained (1) Aug 16 13:11:12 [8982] node0 crmd: notice: crm_update_peer_state: pcmk_quorum_notification: Node node0[1] - state is now member (was (null)) Aug 16 13:11:12 [8982] node0 crmd: info: peer_update_callback: node0 is now member (was (null)) Aug 16 13:11:12 [8982] node0 crmd: notice: crm_update_peer_state: pcmk_quorum_notification: Node node0[1] - state is now lost (was member) Aug 16 13:11:12 [8982] node0 crmd: info: peer_update_callback: node0 is now lost (was member) Aug 16 13:11:12 [8982] node0 crmd: error: reap_dead_nodes: We're not part of the cluster anymore Aug 16 13:11:12 [8982] node0 crmd: error: do_log: FSA: Input I_ERROR from reap_dead_nodes() received in state S_STARTING Aug 16 13:11:12 [8982] node0 crmd: notice: do_state_transition: State transition S_STARTING -> S_RECOVERY [ input=I_ERROR cause=C_FSA_INTERNAL origin=reap_dead_nodes ] Aug 16 13:11:12 [8982] node0 crmd: warning: do_recover: Fast-tracking shutdown in response to errors Aug 16 13:11:12 [8982] node0 crmd: error: do_started: Start cancelled... S_RECOVERY Aug 16 13:11:12 [8982] node0 crmd: error: do_log: FSA: Input I_TERMINATE from do_recover() received in state S_RECOVERY Aug 16 13:11:12 [8982] node0 crmd: info: do_state_transition: State transition S_RECOVERY -> S_TERMINATE [ input=I_TERMINATE cause=C_FSA_INTERNAL origin=do_recover ] Aug 16 13:11:12 [8982] node0 crmd: notice: lrm_state_verify_stopped: Stopped 0 recurring operations at shutdown (0 ops remaining) Aug 16 13:11:12 [8982] node0 crmd: info: do_lrm_control: Disconnecting from the LRM Aug 16 13:11:12 [8982] node0 crmd: info: lrmd_api_disconnect: Disconnecting from lrmd service Aug 16 13:11:12 [8982] node0 crmd: info: lrmd_ipc_connection_destroy: IPC connection destroyed Aug 16 13:11:12 [8977] node0 lrmd: info: crm_client_destroy: Destroying 0 events Aug 16 13:11:12 [8982] node0 crmd: info: lrm_connection_destroy: LRM Connection disconnected Aug 16 13:11:12 [8982] node0 crmd: info: lrmd_api_disconnect: Disconnecting from lrmd service Aug 16 13:11:12 [8982] node0 crmd: notice: do_lrm_control: Disconnected from the LRM Aug 16 13:11:12 [8982] node0 crmd: info: crm_cluster_disconnect: Disconnecting from cluster infrastructure: corosync
wsl...@126.com
_______________________________________________ Users mailing list: Users@clusterlabs.org http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org