On 07/04/2013 10:25 PM, HL wrote:
On 04/07/2013 07:39 μμ, Vijay Bellur wrote:
On 07/04/2013 07:36 PM, HL wrote:
Hello list,

I have a 2 node replica  clusterfs

Volume Name: GLVol
Type: Replicate
Volume ID: 2106bc3d-d184-42db-ab6e-e547ff816113
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: node01:/export/fm_glust
Brick2: node02:/export/fe_glust
Options Reconfigured:
auth.allow: 172.16.35.*
geo-replication.indexing: on


After a cable faillure I Have an unstable system,

gluster volume heal GLVol info split-brain

Gives about 1000 of these entries ..

2013-07-04 16:46:10 <gfid:ce073ea9-a95a-4c5e-b2bd-7db1e26cbad7>
2013-07-04 16:46:10 <gfid:0d7ff6b2-5ed1-4584-b0e3-9f0c723463b8>
2013-07-04 16:46:10 /vz_data/mySQLDBS/var/lib/mysql/ib_logfile0

I've found a script in the users list on how to deal with
/vz_data/mySQLDBS/var/lib/mysql/ib_logfile0
that is known path files

but don't know how to deal with the hex entries ...  they seem to me as
orphans ...

The hex entries are "gfid"s used by GlusterFS for identifying files
and are not orphan objects. A gfid that cannot be resolved to a path
by the server immediately, would appear with the hex representation.


Is ok To delete them ???

Unless the files are in split brain, it is not a good idea to delete
them. Letting the self-heal daemon heal these files would be a better
approach.


Thank's Vijay

I am not sure I can follow
as I stated
the output of
gluster volume heal GLVol info split-brain

I am sorry, I missed the split-brain part in the original post.


was

2013-07-04 16:46:10 <gfid:ce073ea9-a95a-4c5e-b2bd-7db1e26cbad7>
2013-07-04 16:46:10 <gfid:0d7ff6b2-5ed1-4584-b0e3-9f0c723463b8>
2013-07-04 16:46:10 /vz_data/mySQLDBS/var/lib/mysql/ib_logfile0

So it seems to me that they are in split-brain

Other than that ..
How can I make sure that  the self-heal daemon
is running and is healing indeed the files

gluster  volume status all
shows

Status of volume: GLVol
Gluster process                        Port    Online    Pid
------------------------------------------------------------------------------

Brick node01:/export/fm_glust                24009    Y    2614
Brick node02:/export/fe_glust                24009    Y    2746
NFS Server on localhost                    38467    Y    2752
Self-heal Daemon on localhost                N/A    Y    2758
NFS Server on node01                    38467    Y    2626
Self-heal Daemon on node01                N/A    Y    2632

Volume status is the right way to determine if self-heal daemon is running.



So
Delete them or not???

To resolve split-brain, you will need to delete/mv one copy of the replica that you don't consider correct.


If self healing is doing any progress
My bricks are about 500G in size containg about 400K file entries
the two bricks are connected with a dedicated 1G nic
how long will it take to heal ???

Resolving split-brain needs manual intervention. You can use "volume heal <volname> info" to see the activity of self-heal daemon.

Regards,
Vijay
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to