On 09/11/2014 12:16 AM, Ilya Ivanov wrote:
Any insight?
Was the other file's gfid d3def9e1-c6d0-4b7d-a322-b5019305182e?
Could you check if this file exists in <brick>/.glusterfs/d3/de/
When a file is deleted this file also needs to be deleted if there are
no more hardlinks to the file
Pranith
On Tue, Sep 9, 2014 at 8:35 AM, Ilya Ivanov <[email protected]
<mailto:[email protected]>> wrote:
What's a gfid split-brain and how is it different from "normal"
split-brain?
I accessed the file with stat, but "heal info" still shows "Number
of entries: 1"
[root@gluster1 gluster]# getfattr -d -m. -e hex gv01/123
# getfattr -d -m. -e hex gv01/123
# file: gv01/123
trusted.afr.gv01-client-0=0x000000000000000000000000
trusted.afr.gv01-client-1=0x000000000000000000000000
trusted.gfid=0x35f86f4561134ba0bd1b94ef70179d4d
[root@gluster1 gluster]# getfattr -d -m. -e hex gv01
# file: gv01
trusted.afr.gv01-client-0=0x000000000000000000000000
trusted.afr.gv01-client-1=0x000000000000000000000000
trusted.gfid=0x00000000000000000000000000000001
trusted.glusterfs.dht=0x000000010000000000000000ffffffff
trusted.glusterfs.volume-id=0x31a2c4c486ca4344b838d2c2e6c716c1
On Tue, Sep 9, 2014 at 8:19 AM, Pranith Kumar Karampuri
<[email protected] <mailto:[email protected]>> wrote:
On 09/09/2014 11:35 AM, Ilya Ivanov wrote:
Ahh, thank you, now I get it. I deleted it on one node and it
replicated to another one. Now I get the following output:
[root@gluster1 var]# gluster volume heal gv01 info
Brick gluster1:/home/gluster/gv01/
<gfid:d3def9e1-c6d0-4b7d-a322-b5019305182e>
Number of entries: 1
Brick gluster2:/home/gluster/gv01/
Number of entries: 0
Is it normal? Why the number of entries isn't reset to 0?
If you access the file using ls/stat etc, it will be fixed.
But before that could you please post the output of 'getfattr
-d -m. -e hex file/path/in/backend/brick' and 'getfattr -d -m.
-e hex parent/dir/to/file/path/in/backend/brick'
Pranith
And why wouldn't the file show up in split-brain before, anyway?
Gfid split-brains are not shown in heal-info-split-brain yet.
Pranith
On Tue, Sep 9, 2014 at 7:46 AM, Pranith Kumar Karampuri
<[email protected] <mailto:[email protected]>> wrote:
On 09/09/2014 01:54 AM, Ilya Ivanov wrote:
Hello.
I've Gluster 3.5.2 on Centos 6. A primitive replicated
volume, as describe here
<https://www.digitalocean.com/community/tutorials/how-to-create-a-redundant-storage-pool-using-glusterfs-on-ubuntu-servers>.
I tried to simulate split-brain by temporarily
disconnecting the nodes and creating a file with the
same name and different contents. That worked.
The question is, how do I fix it now? All the tutorials
suggest deleting the file from one of the nodes. I can't
do that, it reports "Input/output error". The file won't
even show up in "gluster volume heal gv00 info
split-brain". That shows 0 entries.
The deletion needs to happen on one of the bricks, not
from the mount point.
Pranith
I can see the file in "gluster volume heal gv00 info
heal-failed", though.
--
Ilya.
_______________________________________________
Gluster-users mailing list
[email protected] <mailto:[email protected]>
http://supercolony.gluster.org/mailman/listinfo/gluster-users
--
Ilya.
--
Ilya.
--
Ilya.
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users