On 02/01/2016 07:22 PM, ML mail wrote:
I just found out I needed to run the getfattr on a mount and not on the 
glusterfs server directly. So here are the additional output you asked for:


# getfattr -n glusterfs.gfid.string  -m .  logo-login-09.svg
# file: logo-login-09.svg
glusterfs.gfid.string="1c648409-e98b-4544-a7fa-c2aef87f92ad"

# grep 1c648409-e98b-4544-a7fa-c2aef87f92ad 
/data/myvolume/brick/.glusterfs/changelogs -rn
Binary file /data/myvolume/brick/.glusterfs/changelogs/CHANGELOG.1454278219 
matches
Great! Can you share the CHANGELOG ? ( It contains various fops carried out on this gfid)
Regards
ML



On Monday, February 1, 2016 1:30 PM, Saravanakumar Arumugam 
<[email protected]> wrote:
Hi,

On 02/01/2016 02:14 PM, ML mail wrote:
Hello,

I just set up distributed geo-replication to a slave on my 2 nodes' replicated 
volume and noticed quite a few error messages (around 70 of them) in the 
slave's brick log file:

The exact log file is: /var/log/glusterfs/bricks/data-myvolume-geo-brick.log

[2016-01-31 22:19:29.524370] E [MSGID: 113020] [posix.c:1221:posix_mknod] 
0-myvolume-geo-posix: setting gfid on 
/data/myvolume-geo/brick/data/username/files/shared/logo-login-09.svg.ocTransferId1789604916.part
 failed
[2016-01-31 22:19:29.535478] W [MSGID: 113026] [posix.c:1338:posix_mkdir] 
0-myvolume-geo-posix: mkdir 
(/data/username/files_encryption/keys/files/shared/logo-login-09.svg.ocTransferId1789604916.part):
 gfid (15bbcec6-a332-4c21-81e4-c52472b1e13d) isalready associated with 
directory 
(/data/myvolume-geo/brick/.glusterfs/49/5d/495d6868-4844-4632-8ff9-ad9646a878fe/logo-login-09.svg).
 Hence,both directories will share same gfid and thiscan lead to 
inconsistencies.
Can you grep for this gfid(of the corresponding files) in changelogs and
share those files ?

{
For example:

1. Get gfid of the files like this:

# getfattr -n glusterfs.gfid.string  -m .  /mnt/slave/file456
getfattr: Removing leading '/' from absolute path names
# file: mnt/slave/file456
glusterfs.gfid.string="05b22446-de9e-42df-a63e-399c24d690c4"

2. grep for the corresponding gfid in brick back end like below:

[root@gfvm3 changelogs]# grep 05b22446-de9e-42df-a63e-399c24d690c4
/opt/volume_test/tv_2/b1/.glusterfs/changelogs/ -rn
Binary file
/opt/volume_test/tv_2/b1/.glusterfs/changelogs/CHANGELOG.1454135265 matches
Binary file
/opt/volume_test/tv_2/b1/.glusterfs/changelogs/CHANGELOG.1454135476 matches

}
This will help in understanding what operations are carried out in
master volume, which leads to this inconsistency.

Also, get the following:
gluster version
gluster volume info
gluster volume geo-replication status

This doesn't look good at all because the file mentioned in the error message (
logo-login-09.svg.ocTransferId1789604916.part) is left there with 0 kbytes and does not 
get deleted or cleaned up by glusterfs, leaving my geo-rep slave node in an inconsistent 
state which does not reflect the reality from the master nodes. The master nodes don't 
have that file anymore (which is correct). Here below is an "ls" of the 
concerned file with the correct file on top.


-rw-r--r-- 2 www-data www-data   24312 Jan  6  2014 logo-login-09.svg
-rw-r--r-- 1 root     root           0 Jan 31 23:19 
logo-login-09.svg.ocTransferId1789604916.part
Rename issues in geo-replication are fixed lately. This looks similar to

one.

Thanks,
Saravana
_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users

_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to