BTW: This is the output of volume info and status. u1@u1-virtual-machine:~$ sudo gluster volume info
Volume Name: mysqldata Type: Replicate Volume ID: 27e6161b-d2d0-4369-8ef0-acf18532af73 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: 192.168.53.218:/data/gv0/brick1/mysqldata Brick2: 192.168.53.221:/data/gv0/brick1/mysqldata u1@u1-virtual-machine:~$ sudo gluster volume status Status of volume: mysqldata Gluster process Port Online Pid ------------------------------------------------------------------------------ Brick 192.168.53.218:/data/gv0/brick1/mysqldata 49154 Y 2071 Brick 192.168.53.221:/data/gv0/brick1/mysqldata 49153 Y 2170 NFS Server on localhost 2049 Y 2066 Self-heal Daemon on localhost N/A Y 2076 NFS Server on 192.168.53.221 2049 Y 2175 Self-heal Daemon on 192.168.53.221 N/A Y 2180 There are no active volume tasks 2014/1/18 Yandong Yao <[email protected]> > Hi Guys, > > I am testing glusterfs and have configured replicated volume (replica=2 on > two virtual machines), after play with the volume a while, there are > un-consistent data reported by 'heal volname info': > > u1@u1-virtual-machine:~$ sudo gluster volume heal mysqldata info > Gathering Heal info on volume mysqldata has been successful > > Brick 192.168.53.218:/data/gv0/brick1/mysqldata > Number of entries: 1 > <gfid:0ff1a4e1-b14c-41d6-826b-e749a4e6ec7f> > > Brick 192.168.53.221:/data/gv0/brick1/mysqldata > Number of entries: 1 > /ibdata1 > > > *1) What does this means? Why one entry is file itself on one host, while > another entry is gfid on another host? * > > *2) After a while (maybe 2 minutes), re-run heal info, and get following > output. What happened behind the scene? Why the entry changes to file from > gfid?* > > u1@u1-virtual-machine:~$ sudo gluster volume heal mysqldata info > Gathering Heal info on volume mysqldata has been successful > > Brick 192.168.53.218:/data/gv0/brick1/mysqldata > Number of entries: 1 > /ibdata1 > > Brick 192.168.53.221:/data/gv0/brick1/mysqldata > Number of entries: 1 > /ibdata1 > u1@u1-virtual-machine:~$ sudo gluster volume heal mysqldata info > split-brain > Gathering Heal info on volume mysqldata has been successful > > Brick 192.168.53.218:/data/gv0/brick1/mysqldata > Number of entries: 0 > > Brick 192.168.53.221:/data/gv0/brick1/mysqldata > Number of entries: 0 > > *3) I tried with both heal and heal full, while heal seems not work, I > still get above output. How could I heal this case manually? Following is > getfattr output.* > > u1@u1-virtual-machine:~$ sudo getfattr -e hex -m . -d > /data/gv0/brick1/mysqldata/ibdata1 > getfattr: Removing leading '/' from absolute path names > # file: data/gv0/brick1/mysqldata/ibdata1 > trusted.afr.mysqldata-client-0=0x000000010000000000000000 > trusted.afr.mysqldata-client-1=0x000000010000000000000000 > trusted.gfid=0x0ff1a4e1b14c41d6826be749a4e6ec7f > > > Any comments are welcome, and thanks very much in advance! > > Regards, > Yandong >
_______________________________________________ Gluster-users mailing list [email protected] http://supercolony.gluster.org/mailman/listinfo/gluster-users
