Hi,

On 08/16/2016 01:29 PM, wsl...@126.com wrote:
thank you for your reply.
The network is well,  firewall has been closed,  and selinux has been disabled.
/var/log/pacemaker.log as follows:
To be honest, this piece of logs is useless. A whole log file should be attached, which logs messages from
the first time when things go bad until now;-)

Also, lots of info is missing, like OS distributions, packages' version and cluster configuration, etc.
A "hb_report" (man hb_report) would be great if package "cluster-glue" is 
intalled.

Eric
Aug 16 13:11:12 [8980] node0       crmd:     info: peer_update_callback:        
node0 is now (null)
Aug 16 13:11:12 [8980] node0       crmd:     info: crm_get_peer:        Node 1 
has uuid 1
Aug 16 13:11:12 [8980] node0       crmd:     info: crm_update_peer_proc:        
cluster_connect_cpg: Node node0[1] - corosync-cpg is now online
Aug 16 13:11:12 [8980] node0       crmd:     info: peer_update_callback:        
Client node0/peer now has status [online] (DC=<null>)
Aug 16 13:11:12 [8980] node0       crmd:     info: init_cs_connection_once:     
Connection to 'corosync': established
Aug 16 13:11:12 [8980] node0       crmd:   notice: cluster_connect_quorum:      
Quorum acquired
Aug 16 13:11:12 [8980] node0       crmd:     info: do_ha_control:       
Connected to the cluster
Aug 16 13:11:12 [8980] node0       crmd:     info: lrmd_ipc_connect:    
Connecting to lrmd
Aug 16 13:11:12 [8977] node0       lrmd:     info: crm_client_new:      
Connecting 0x16025a0 for uid=189 gid=189 pid=8980 
id=e317bf62-fb55-46a1-af17-9284234917b8
Aug 16 13:11:12 [8975] node0        cib:     info: cib_process_request:         
Completed cib_modify operation for section nodes: OK (rc=0, 
origin=local/crmd/3, version=0.1.0)
Aug 16 13:11:12 [8980] node0       crmd:     info: do_lrm_control:      LRM 
connection established
Aug 16 13:11:12 [8980] node0       crmd:     info: do_started:  Delaying start, 
no membership data (0000000000100000)
Aug 16 13:11:12 [8980] node0       crmd:     info: do_started:  Delaying start, 
no membership data (0000000000100000)
Aug 16 13:11:12 [8975] node0        cib:     info: cib_process_request:         
Completed cib_query operation for section crm_config: OK (rc=0, 
origin=local/crmd/4, version=0.1.0)
Aug 16 13:11:12 [8980] node0       crmd:     info: pcmk_quorum_notification:    
Membership 8571476093872377196: quorum retained (1)
Aug 16 13:11:12 [8980] node0       crmd:   notice: crm_update_peer_state:       
pcmk_quorum_notification: Node node0[1] - state is now member (was (null))
Aug 16 13:11:12 [8980] node0       crmd:     info: peer_update_callback:        
node0 is now member (was (null))
Aug 16 13:11:12 [8980] node0       crmd:   notice: crm_update_peer_state:       
pcmk_quorum_notification: Node node0[1] - state is now lost (was member)
Aug 16 13:11:12 [8980] node0       crmd:     info: peer_update_callback:        
node0 is now lost (was member)
Aug 16 13:11:12 [8980] node0       crmd:    error: reap_dead_nodes:     We're 
not part of the cluster anymore
Aug 16 13:11:12 [8975] node0        cib:     info: crm_client_new:      
Connecting 0x1cbcd80 for uid=0 gid=0 pid=8976 
id=9afb683d-8390-408a-a690-fb5447fefd37
Aug 16 13:11:12 [8976] node0 stonith-ng:   notice: setup_cib:   Watching for 
stonith topology changes
Aug 16 13:11:12 [8976] node0 stonith-ng:     info: qb_ipcs_us_publish:  server 
name: stonith-ng
Aug 16 13:11:12 [8976] node0 stonith-ng:     info: main:        Starting 
stonith-ng mainloop
Aug 16 13:11:12 [8976] node0 stonith-ng:     info: pcmk_cpg_membership:         
Joined[0.0] stonith-ng.1
Aug 16 13:11:12 [8976] node0 stonith-ng:     info: pcmk_cpg_membership:         
Member[0.0] stonith-ng.1
Aug 16 13:11:12 [8980] node0       crmd:    error: do_log:      FSA: Input 
I_ERROR from reap_dead_nodes() received in state S_STARTING
Aug 16 13:11:12 [8980] node0       crmd:   notice: do_state_transition:         
State transition S_STARTING -> S_RECOVERY [ input=I_ERROR cause=C_FSA_INTERNAL 
origin=reap_dead_nodes ]
Aug 16 13:11:12 [8980] node0       crmd:  warning: do_recover:  Fast-tracking 
shutdown in response to errors
Aug 16 13:11:12 [8980] node0       crmd:    error: do_started:  Start 
cancelled... S_RECOVERY
Aug 16 13:11:12 [8980] node0       crmd:    error: do_log:      FSA: Input 
I_TERMINATE from do_recover() received in state S_RECOVERY
Aug 16 13:11:12 [8980] node0       crmd:     info: do_state_transition:         
State transition S_RECOVERY -> S_TERMINATE [ input=I_TERMINATE 
cause=C_FSA_INTERNAL origin=do_recover ]
Aug 16 13:11:12 [8980] node0       crmd:   notice: lrm_state_verify_stopped:    
Stopped 0 recurring operations at shutdown (0 ops remaining)
Aug 16 13:11:12 [8980] node0       crmd:     info: do_lrm_control:      
Disconnecting from the LRM
Aug 16 13:11:12 [8980] node0       crmd:     info: lrmd_api_disconnect:         
Disconnecting from lrmd service
Aug 16 13:11:12 [8980] node0       crmd:     info: lrmd_ipc_connection_destroy: 
        IPC connection destroyed
Aug 16 13:11:12 [8980] node0       crmd:     info: lrm_connection_destroy:      
LRM Connection disconnected
Aug 16 13:11:12 [8980] node0       crmd:     info: lrmd_api_disconnect:         
Disconnecting from lrmd service
Aug 16 13:11:12 [8980] node0       crmd:   notice: do_lrm_control:      
Disconnected from the LRM
Aug 16 13:11:12 [8980] node0       crmd:     info: crm_cluster_disconnect:      
Disconnecting from cluster infrastructure: corosync
Aug 16 13:11:12 [8980] node0       crmd:   notice: terminate_cs_connection:     
Disconnecting from Corosync
Aug 16 13:11:12 [8980] node0       crmd:     info: crm_cluster_disconnect:      
Disconnected from corosync
Aug 16 13:11:12 [8980] node0       crmd:     info: do_ha_control:       
Disconnected from the cluster
Aug 16 13:11:12 [8977] node0       lrmd:     info: crm_client_destroy:  
Destroying 0 events
Aug 16 13:11:12 [8980] node0       crmd:     info: do_cib_control:      
Disconnecting CIB
Aug 16 13:11:12 [8975] node0        cib:     info: cib_process_request:         
Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, 
version=0.1.0)
Aug 16 13:11:12 [8976] node0 stonith-ng:     info: init_cib_cache_cb:   
Updating device list from the cib: init
Aug 16 13:11:12 [8980] node0       crmd:     info: crmd_cib_connection_destroy: 
        Connection to the CIB terminated...
Aug 16 13:11:12 [8980] node0       crmd:     info: do_exit:     Performing 
A_EXIT_0 - gracefully exiting the CRMd
Aug 16 13:11:12 [8980] node0       crmd:     info: do_exit:     [crmd] stopped 
(0)
Aug 16 13:11:12 [8980] node0       crmd:     info: crmd_exit:   Dropping 
I_TERMINATE: [ state=S_TERMINATE cause=C_FSA_INTERNAL origin=do_stop ]
Aug 16 13:11:12 [8980] node0       crmd:     info: crmd_quorum_destroy:         
connection closed
Aug 16 13:11:12 [8980] node0       crmd:     info: crmd_cs_destroy:     
connection closed
Aug 16 13:11:12 [8980] node0       crmd:     info: crmd_init:   8980 stopped: 
OK (0)
Aug 16 13:11:12 [8980] node0       crmd:    error: crmd_fast_exit:      Could 
not recover from internal error
Aug 16 13:11:12 [8980] node0       crmd:     info: crm_xml_cleanup:     
Cleaning up memory from libxml2
Aug 16 13:11:12 [8975] node0        cib:     info: crm_client_destroy:  
Destroying 0 events
Aug 16 13:11:12 [8974] node0 pacemakerd:    error: pcmk_child_exit:     Child 
process crmd (8980) exited: Generic Pacemaker error (201)
Aug 16 13:11:12 [8974] node0 pacemakerd:    error: pcmk_child_exit:     Child 
process crmd (8980) exited: Generic Pacemaker error (201)
Aug 16 13:11:12 [8974] node0 pacemakerd:   notice: pcmk_process_exit:   
Respawning failed child process: crmd
Aug 16 13:11:12 [8974] node0 pacemakerd:   notice: pcmk_process_exit:   
Respawning failed child process: crmd
Aug 16 13:11:12 [8974] node0 pacemakerd:     info: start_child:         Using 
uid=189 and group=189 for process crmd
Aug 16 13:11:12 [8974] node0 pacemakerd:     info: start_child:         Using 
uid=189 and group=189 for process crmd
Aug 16 13:11:12 [8974] node0 pacemakerd:     info: start_child:         Forked 
child 8982 for process crmd
Aug 16 13:11:12 [8974] node0 pacemakerd:     info: start_child:         Forked 
child 8982 for process crmd
Aug 16 13:11:12 [8982] node0       crmd:     info: crm_log_init:        Changed 
active directory to /var/lib/pacemaker/cores/hacluster
Aug 16 13:11:12 [8982] node0       crmd:   notice: main:        CRM Git 
Version: 368c726
Aug 16 13:11:12 [8982] node0       crmd:     info: do_log:      FSA: Input 
I_STARTUP from crmd_init() received in state S_STARTING
Aug 16 13:11:12 [8982] node0       crmd:     info: get_cluster_type:    
Verifying cluster type: 'corosync'
Aug 16 13:11:12 [8982] node0       crmd:     info: get_cluster_type:    
Assuming an active 'corosync' cluster
Aug 16 13:11:12 [8975] node0        cib:     info: crm_client_new:      
Connecting 0x1a8dd80 for uid=189 gid=189 pid=8982 
id=5e050be4-f112-4f32-831b-1704ce1872dd
Aug 16 13:11:12 [8982] node0       crmd:     info: do_cib_control:      CIB 
connection established
Aug 16 13:11:12 [8982] node0       crmd:   notice: crm_cluster_connect:         
Connecting to cluster infrastructure: corosync
Aug 16 13:11:12 [8975] node0        cib:     info: cib_process_request:         
Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, 
version=0.1.0)
Aug 16 13:11:12 [8982] node0       crmd:     info: crm_get_peer:        Created 
entry c01a1bd2-9f44-4061-9ea3-95c1df3de86d/0xea3c20 for node (null)/1 (1 total)
Aug 16 13:11:12 [8982] node0       crmd:     info: crm_get_peer:        Node 1 
is now known as node0
Aug 16 13:11:12 [8982] node0       crmd:     info: peer_update_callback:        
node0 is now (null)
Aug 16 13:11:12 [8982] node0       crmd:     info: crm_get_peer:        Node 1 
has uuid 1
Aug 16 13:11:12 [8982] node0       crmd:     info: crm_update_peer_proc:        
cluster_connect_cpg: Node node0[1] - corosync-cpg is now online
Aug 16 13:11:12 [8982] node0       crmd:     info: peer_update_callback:        
Client node0/peer now has status [online] (DC=<null>)
Aug 16 13:11:12 [8982] node0       crmd:     info: init_cs_connection_once:     
Connection to 'corosync': established
Aug 16 13:11:12 [8982] node0       crmd:   notice: cluster_connect_quorum:      
Quorum acquired
Aug 16 13:11:12 [8975] node0        cib:     info: cib_process_request:         
Completed cib_modify operation for section nodes: OK (rc=0, 
origin=local/crmd/3, version=0.1.0)
Aug 16 13:11:12 [8982] node0       crmd:     info: do_ha_control:       
Connected to the cluster
Aug 16 13:11:12 [8982] node0       crmd:     info: lrmd_ipc_connect:    
Connecting to lrmd
Aug 16 13:11:12 [8977] node0       lrmd:     info: crm_client_new:      
Connecting 0x16025a0 for uid=189 gid=189 pid=8982 
id=c6776327-b798-40f2-8ca4-a4b032cff2eb
Aug 16 13:11:12 [8982] node0       crmd:     info: do_lrm_control:      LRM 
connection established
Aug 16 13:11:12 [8982] node0       crmd:     info: do_started:  Delaying start, 
no membership data (0000000000100000)
Aug 16 13:11:12 [8982] node0       crmd:     info: do_started:  Delaying start, 
no membership data (0000000000100000)
Aug 16 13:11:12 [8975] node0        cib:     info: cib_process_request:         
Completed cib_query operation for section crm_config: OK (rc=0, 
origin=local/crmd/4, version=0.1.0)
Aug 16 13:11:12 [8982] node0       crmd:     info: pcmk_quorum_notification:    
Membership 8571476093872377196: quorum retained (1)
Aug 16 13:11:12 [8982] node0       crmd:   notice: crm_update_peer_state:       
pcmk_quorum_notification: Node node0[1] - state is now member (was (null))
Aug 16 13:11:12 [8982] node0       crmd:     info: peer_update_callback:        
node0 is now member (was (null))
Aug 16 13:11:12 [8982] node0       crmd:   notice: crm_update_peer_state:       
pcmk_quorum_notification: Node node0[1] - state is now lost (was member)
Aug 16 13:11:12 [8982] node0       crmd:     info: peer_update_callback:        
node0 is now lost (was member)
Aug 16 13:11:12 [8982] node0       crmd:    error: reap_dead_nodes:     We're 
not part of the cluster anymore
Aug 16 13:11:12 [8982] node0       crmd:    error: do_log:      FSA: Input 
I_ERROR from reap_dead_nodes() received in state S_STARTING
Aug 16 13:11:12 [8982] node0       crmd:   notice: do_state_transition:         
State transition S_STARTING -> S_RECOVERY [ input=I_ERROR cause=C_FSA_INTERNAL 
origin=reap_dead_nodes ]
Aug 16 13:11:12 [8982] node0       crmd:  warning: do_recover:  Fast-tracking 
shutdown in response to errors
Aug 16 13:11:12 [8982] node0       crmd:    error: do_started:  Start 
cancelled... S_RECOVERY
Aug 16 13:11:12 [8982] node0       crmd:    error: do_log:      FSA: Input 
I_TERMINATE from do_recover() received in state S_RECOVERY
Aug 16 13:11:12 [8982] node0       crmd:     info: do_state_transition:         
State transition S_RECOVERY -> S_TERMINATE [ input=I_TERMINATE 
cause=C_FSA_INTERNAL origin=do_recover ]
Aug 16 13:11:12 [8982] node0       crmd:   notice: lrm_state_verify_stopped:    
Stopped 0 recurring operations at shutdown (0 ops remaining)
Aug 16 13:11:12 [8982] node0       crmd:     info: do_lrm_control:      
Disconnecting from the LRM
Aug 16 13:11:12 [8982] node0       crmd:     info: lrmd_api_disconnect:         
Disconnecting from lrmd service
Aug 16 13:11:12 [8982] node0       crmd:     info: lrmd_ipc_connection_destroy: 
        IPC connection destroyed
Aug 16 13:11:12 [8977] node0       lrmd:     info: crm_client_destroy:  
Destroying 0 events
Aug 16 13:11:12 [8982] node0       crmd:     info: lrm_connection_destroy:      
LRM Connection disconnected
Aug 16 13:11:12 [8982] node0       crmd:     info: lrmd_api_disconnect:         
Disconnecting from lrmd service
Aug 16 13:11:12 [8982] node0       crmd:   notice: do_lrm_control:      
Disconnected from the LRM
Aug 16 13:11:12 [8982] node0       crmd:     info: crm_cluster_disconnect:      
Disconnecting from cluster infrastructure: corosync



wsl...@126.com


_______________________________________________
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


_______________________________________________
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to