I should add this is found in logs ---------------------------------------------------------------[ gluster02.uat ] /var/log/glusterfs/cdn-uat-rebalance.log:[2014-02-26 00:06:38.550396] I [dht-common.c:1017:dht_lookup_everywhere_cbk] 0-cdn-uat-dht: deleting stale linkfile ThePinkPanther2_2009_23_HLS_layer2_642000_95.ts on cdn-uat-replicate-2
---------------------------------------------------------------[ gluster03.uat ] /var/log/glusterfs/cdn-uat-rebalance.log:[2014-02-26 00:06:38.556472] E [afr-self-heal-common.c:2212:afr_self_heal_completion_cbk] 0-cdn-uat-replicate-2: background meta-data data entry missing-entry gfid self-heal failed on ThePinkPanther2_2009_23_HLS_layer2_642000_95.ts also these files are found on the bricks, the actuall Gluster mount seems to be ok. However I've never seen this in v3.2.5 and the error above says that smth is not right volume info here: [[email protected] g34]# gluster volume info Volume Name: cdn-uat Type: Distributed-Replicate Volume ID: 3e353d61-ac78-43d4-af20-55d1672a5cd3 Status: Started Number of Bricks: 3 x 2 = 6 Transport-type: tcp Bricks: Brick1: gluster08.uat:/mnt/gluster/brick01/data Brick2: gluster07.uat:/mnt/gluster/brick01/data Brick3: gluster01.uat:/mnt/gluster/brick01/data Brick4: gluster02.uat:/mnt/gluster/brick01/data Brick5: gluster03.uat:/mnt/gluster/brick01/data Brick6: gluster04.uat:/mnt/gluster/brick01/data Options Reconfigured: diagnostics.client-log-level: ERROR [[email protected] g34]# On Wed 26 Feb 2014 13:50:40, Viktor Villafuerte wrote: > Hi all, > > I've got these packages installed > > [[email protected] g34]# rpm -qa | grep gluster > glusterfs-3.4.2-1.el6.x86_64 > glusterfs-cli-3.4.2-1.el6.x86_64 > glusterfs-libs-3.4.2-1.el6.x86_64 > glusterfs-fuse-3.4.2-1.el6.x86_64 > glusterfs-server-3.4.2-1.el6.x86_64 > [[email protected] g34]# > > > after rebalance I have number of files in 'T' > > [[email protected] g34]# ls -l > ThePinkPanther2_2009_23_HLS_layer2_642000_95.ts > ---------T 2 1000 1000 0 Feb 26 11:06 > ThePinkPanther2_2009_23_HLS_layer2_642000_95.ts > [[email protected] g34]# > > > I've tried this twice once extend 1 (1x1) => 2 (1x1) and once 2 (1x1) => > 3 (1x1) and both times I end up with about 1000 files like that one > > > [[email protected] g34]# getfattr -m trusted.* -d > ThePinkPanther2_2009_23_HLS_layer2_642000_95.ts > # file: ThePinkPanther2_2009_23_HLS_layer2_642000_95.ts > trusted.gfid="�]�U\\�H<���-��" > trusted.glusterfs.dht.linkto="cdn-uat-replicate-0" > > [[email protected] g34]# > > > which would point to the '0' replica and surely > > > [[email protected] g34]# ls -l > ThePinkPanther2_2009_23_HLS_layer2_642000_95.ts > -rw-r--r-- 2 1000 1000 997728 Jan 8 11:14 > ThePinkPanther2_2009_23_HLS_layer2_642000_95.ts > > > > Now when I remove the file from '08' > I get > > [[email protected] g34]# ls -l > ThePinkPanther2_2009_23_HLS_layer2_642000_95.ts > ls: cannot access ThePinkPanther2_2009_23_HLS_layer2_642000_95.ts: No such > file or directory > [[email protected] g34]# > > > > but > > > > [[email protected] g34]# getfattr -m trusted.* -d > ThePinkPanther2_2009_23_HLS_layer2_642000_95.ts > # file: ThePinkPanther2_2009_23_HLS_layer2_642000_95.ts > trusted.gfid="�]�U\\�H<���-��" > trusted.glusterfs.dht.linkto="cdn-uat-replicate-0" > > [[email protected] g34]# > > > > > I'm sure that this is not by design? Is there a way how to fix this? Or > what would be the recommended series of action(s) that should be taken > now to rectify this? > > > v > > > -- > Regards > > Viktor Villafuerte > Optus Internet Engineering > t: 02 808-25265 -- Regards Viktor Villafuerte Optus Internet Engineering t: 02 808-25265 _______________________________________________ Gluster-users mailing list [email protected] http://supercolony.gluster.org/mailman/listinfo/gluster-users
