Any insight? On Tue, Sep 9, 2014 at 8:35 AM, Ilya Ivanov <[email protected]> wrote:
> What's a gfid split-brain and how is it different from "normal" > split-brain? > > I accessed the file with stat, but "heal info" still shows "Number of > entries: 1" > > [root@gluster1 gluster]# getfattr -d -m. -e hex gv01/123 > # getfattr -d -m. -e hex gv01/123 > # file: gv01/123 > trusted.afr.gv01-client-0=0x000000000000000000000000 > trusted.afr.gv01-client-1=0x000000000000000000000000 > trusted.gfid=0x35f86f4561134ba0bd1b94ef70179d4d > > [root@gluster1 gluster]# getfattr -d -m. -e hex gv01 > # file: gv01 > trusted.afr.gv01-client-0=0x000000000000000000000000 > trusted.afr.gv01-client-1=0x000000000000000000000000 > trusted.gfid=0x00000000000000000000000000000001 > trusted.glusterfs.dht=0x000000010000000000000000ffffffff > trusted.glusterfs.volume-id=0x31a2c4c486ca4344b838d2c2e6c716c1 > > > > On Tue, Sep 9, 2014 at 8:19 AM, Pranith Kumar Karampuri < > [email protected]> wrote: > >> >> On 09/09/2014 11:35 AM, Ilya Ivanov wrote: >> >> Ahh, thank you, now I get it. I deleted it on one node and it >> replicated to another one. Now I get the following output: >> >> [root@gluster1 var]# gluster volume heal gv01 info >> Brick gluster1:/home/gluster/gv01/ >> <gfid:d3def9e1-c6d0-4b7d-a322-b5019305182e> >> Number of entries: 1 >> >> Brick gluster2:/home/gluster/gv01/ >> Number of entries: 0 >> >> Is it normal? Why the number of entries isn't reset to 0? >> >> If you access the file using ls/stat etc, it will be fixed. But before >> that could you please post the output of 'getfattr -d -m. -e hex >> file/path/in/backend/brick' and 'getfattr -d -m. -e hex >> parent/dir/to/file/path/in/backend/brick' >> >> Pranith >> >> >> >> And why wouldn't the file show up in split-brain before, anyway? >> >> Gfid split-brains are not shown in heal-info-split-brain yet. >> >> Pranith >> >> >> >> On Tue, Sep 9, 2014 at 7:46 AM, Pranith Kumar Karampuri < >> [email protected]> wrote: >> >>> >>> On 09/09/2014 01:54 AM, Ilya Ivanov wrote: >>> >>> Hello. >>> >>> I've Gluster 3.5.2 on Centos 6. A primitive replicated volume, as >>> describe here >>> <https://www.digitalocean.com/community/tutorials/how-to-create-a-redundant-storage-pool-using-glusterfs-on-ubuntu-servers>. >>> I tried to simulate split-brain by temporarily disconnecting the nodes and >>> creating a file with the same name and different contents. That worked. >>> >>> The question is, how do I fix it now? All the tutorials suggest >>> deleting the file from one of the nodes. I can't do that, it reports >>> "Input/output error". The file won't even show up in "gluster volume heal >>> gv00 info split-brain". That shows 0 entries. >>> >>> The deletion needs to happen on one of the bricks, not from the mount >>> point. >>> >>> Pranith >>> >>> I can see the file in "gluster volume heal gv00 info heal-failed", >>> though. >>> >>> >>> -- >>> Ilya. >>> >>> >>> _______________________________________________ >>> Gluster-users mailing >>> [email protected]http://supercolony.gluster.org/mailman/listinfo/gluster-users >>> >>> >>> >> >> >> -- >> Ilya. >> >> >> > > > -- > Ilya. > -- Ilya.
_______________________________________________ Gluster-users mailing list [email protected] http://supercolony.gluster.org/mailman/listinfo/gluster-users
