Hi,

Gluster version is 3.5.3-1.
/var/log/gluster.log (client log) gives during the rm -rf the  following logs:
[2015-02-23 14:42:50.180091] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 
0-md1-
client-2: remote operation failed: Directory not empty
[2015-02-23 14:42:50.180134] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 
0-md1-
client-3: remote operation failed: Directory not empty
[2015-02-23 14:42:50.180740] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-5: remote operation failed: File exists. Path: 
/linux/suse/12.1/KDE4.7.4/i586
[2015-02-23 14:42:50.180772] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-4: remote operation failed: File exists. Path: 
/linux/suse/12.1/KDE4.7.4/i586
[2015-02-23 14:42:50.181129] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-3: remote operation failed: File exists. Path: 
/linux/suse/12.1/KDE4.7.4/i586
[2015-02-23 14:42:50.181160] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-2: remote operation failed: File exists. Path: 
/linux/suse/12.1/KDE4.7.4/i586
[2015-02-23 14:42:50.319213] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 
0-md1-
client-3: remote operation failed: Directory not empty
[2015-02-23 14:42:50.319762] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 
0-md1-
client-2: remote operation failed: Directory not empty
[2015-02-23 14:42:50.320501] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-0: remote operation failed: File exists. Path: 
/linux/suse/12.1/src-oss/suse/src
[2015-02-23 14:42:50.320552] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-1: remote operation failed: File exists. Path: 
/linux/suse/12.1/src-oss/suse/src
[2015-02-23 14:42:50.320842] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-2: remote operation failed: File exists. Path: 
/linux/suse/12.1/src-oss/suse/src
[2015-02-23 14:42:50.320884] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-3: remote operation failed: File exists. Path: 
/linux/suse/12.1/src-oss/suse/src
[2015-02-23 14:42:50.438982] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 
0-md1-
client-3: remote operation failed: Directory not empty
[2015-02-23 14:42:50.439347] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 
0-md1-
client-2: remote operation failed: Directory not empty
[2015-02-23 14:42:50.440235] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-0: remote operation failed: File exists. Path: 
/linux/suse/12.1/oss/suse/noarch
[2015-02-23 14:42:50.440344] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-1: remote operation failed: File exists. Path: 
/linux/suse/12.1/oss/suse/noarch
[2015-02-23 14:42:50.440603] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-2: remote operation failed: File exists. Path: 
/linux/suse/12.1/oss/suse/noarch
[2015-02-23 14:42:50.440665] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-3: remote operation failed: File exists. Path: 
/linux/suse/12.1/oss/suse/noarch
[2015-02-23 14:42:50.680827] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 
0-md1-
client-2: remote operation failed: Directory not empty
[2015-02-23 14:42:50.681721] W [client-rpc-fops.c:696:client3_3_rmdir_cbk] 
0-md1-
client-3: remote operation failed: Directory not empty
[2015-02-23 14:42:50.682482] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-3: remote operation failed: File exists. Path: 
/linux/suse/12.1/oss/suse/i586
[2015-02-23 14:42:50.682517] W [client-rpc-fops.c:322:client3_3_mkdir_cbk] 
0-md1-
client-2: remote operation failed: File exists. Path: 
/linux/suse/12.1/oss/suse/i586


Thanks,


A.


On Monday 23 February 2015 20:06:17 Ravishankar N wrote:


        
On 02/23/2015 07:04 PM, Alessandro Ipe      wrote:        
                  
Hi Ravi,      
      
      
gluster volume status md1 returns      
Status of volume: md1      
Gluster process Port Online Pid      
------------------------------------------------------------------------------  
    
Brick tsunami1:/data/glusterfs/md1/brick1        49157 Y 2260      
Brick tsunami2:/data/glusterfs/md1/brick1        49152 Y 2320      
Brick tsunami3:/data/glusterfs/md1/brick1        49156 Y 20715      
Brick tsunami4:/data/glusterfs/md1/brick1        49156 Y 10544      
Brick tsunami5:/data/glusterfs/md1/brick1        49152 Y 12588      
Brick tsunami6:/data/glusterfs/md1/brick1        49152 Y 12242      
Self-heal Daemon on localhost N/A Y 2336      
Self-heal Daemon on tsunami2 N/A Y 2359      
Self-heal Daemon on tsunami5 N/A Y 27619      
Self-heal Daemon on tsunami4 N/A Y 12318      
Self-heal Daemon on tsunami3 N/A Y 19118      
Self-heal Daemon on tsunami6 N/A Y 27650      
       
Task Status of Volume md1      
------------------------------------------------------------------------------  
    
Task : Rebalance       
ID : 9dfee1a2-49ac-4766-bdb6-00de5e5883f6      
Status : completed       
so it seems that all brick server are up.      
      
gluster volume heal md1 info returns      
Brick        tsunami1.oma.be:/data/glusterfs/md1/brick1/      
Number of entries: 0      
      
Brick        tsunami2.oma.be:/data/glusterfs/md1/brick1/      
Number of entries: 0      
      
Brick        tsunami3.oma.be:/data/glusterfs/md1/brick1/      
Number of entries: 0      
      
Brick        tsunami4.oma.be:/data/glusterfs/md1/brick1/      
Number of entries: 0      
      
Brick        tsunami5.oma.be:/data/glusterfs/md1/brick1/      
Number of entries: 0      
      
Brick        tsunami6.oma.be:/data/glusterfs/md1/brick1/      
Number of entries: 0      
      
Should I run "gluster volume heal md1 full" ?      
    
    Hi Alessandro,        Looks like there is no pending-self heals, so no need 
to run the    
heal command. Can you share the output of the client (mount) log    when you 
get the 
ENOTEMPTY during the rm -rf?        What version of gluster are you using?    
Thanks,    Ravi        
      
      
Thanks,      
_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to