Hello,

I have tried to config a geo-replication volume , all the master nodes 
configuration are the same, When I start this volume, the status shows partial 
faulty as following:

gluster volume geo-replication filews 
glusterfs01.sh3.ctripcorp.com::filews_slave status

MASTER NODE      MASTER VOL    MASTER BRICK          SLAVE                      
                    STATUS     CHECKPOINT STATUS    CRAWL STATUS
-------------------------------------------------------------------------------------------------------------------------------------------------
SVR8048HW2285    filews        /export/sdb/filews    
glusterfs01.sh3.ctripcorp.com::filews_slave    faulty     N/A                  
N/A
SVR8050HW2285    filews        /export/sdb/filews    
glusterfs03.sh3.ctripcorp.com::filews_slave    Passive    N/A                  
N/A
SVR8047HW2285    filews        /export/sdb/filews    
glusterfs01.sh3.ctripcorp.com::filews_slave    Active     N/A                  
Hybrid Crawl
SVR8049HW2285    filews        /export/sdb/filews    
glusterfs05.sh3.ctripcorp.com::filews_slave    Active     N/A                  
Hybrid Crawl
SH02SVR5951      filews        /export/sdb/brick1    
glusterfs06.sh3.ctripcorp.com::filews_slave    Passive    N/A                  
N/A
SH02SVR5953      filews        /export/sdb/brick1    
glusterfs01.sh3.ctripcorp.com::filews_slave    faulty     N/A                  
N/A
SVR6995HW2285    filews        /export/sdb/filews    
glusterfs01.sh3.ctripcorp.com::filews_slave    faulty     N/A                  
N/A
SH02SVR5954      filews        /export/sdb/brick1    
glusterfs01.sh3.ctripcorp.com::filews_slave    faulty     N/A                  
N/A
SVR6994HW2285    filews        /export/sdb/filews    
glusterfs02.sh3.ctripcorp.com::filews_slave    Passive    N/A                  
N/A
SVR6993HW2285    filews        /export/sdb/filews    
glusterfs01.sh3.ctripcorp.com::filews_slave    faulty     N/A                  
N/A
SH02SVR5952      filews        /export/sdb/brick1    
glusterfs01.sh3.ctripcorp.com::filews_slave    faulty     N/A                  
N/A
SVR6996HW2285    filews        /export/sdb/filews    
glusterfs04.sh3.ctripcorp.com::filews_slave    Passive    N/A                  
N/A

On the faulty node, log file /var/log/glusterfs/geo-replication/filews shows 
worker(/export/sdb/filews) died before establishing connection

[2016-05-18 16:55:46.402622] I [monitor(monitor):215:monitor] Monitor: 
------------------------------------------------------------
[2016-05-18 16:55:46.402930] I [monitor(monitor):216:monitor] Monitor: starting 
gsyncd worker
[2016-05-18 16:55:46.517460] I [changelogagent(agent):72:__init__] 
ChangelogAgent: Agent listining...
[2016-05-18 16:55:46.518066] I [repce(agent):92:service_loop] RepceServer: 
terminating on reaching EOF.
[2016-05-18 16:55:46.518279] I [syncdutils(agent):214:finalize] <top>: exiting.
[2016-05-18 16:55:46.518194] I [monitor(monitor):267:monitor] Monitor: 
worker(/export/sdb/filews) died before establishing connection
[2016-05-18 16:55:56.697036] I [monitor(monitor):215:monitor] Monitor: 
------------------------------------------------------------

Any advice and suggestions will be greatly appreciated.





Best Regards
杨雨阳 Yuyang Yang


_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to