hi Andreas,
The directory is in split-brain. Do you have any files/directories, that are in split-brain in the directory 'media/ga/live/a' ??

Could you give output of
"getfattr -d -m. -e hex media/ga/live/a/*" on both the bricks?

Pranith

On 01/04/2016 05:21 PM, Andreas Tsaridas wrote:
Hello,

Please see below :
-----

web01 # getfattr -d -m . -e hex media/ga/live/a
# file: media/ga/live/a
trusted.afr.dirty=0x000000000000000000000000
trusted.afr.remote1=0x000000000000000000000000
trusted.afr.remote2=0x000000000000000000000005
trusted.afr.share-client-0=0x000000000000000000000000
trusted.afr.share-client-1=0x0000000000000000000000ee
trusted.gfid=0xb13199a1464c44918464444b3f7eeee3
trusted.glusterfs.dht=0x000000010000000000000000ffffffff


------

web02 # getfattr -d -m . -e hex media/ga/live/a
# file: media/ga/live/a
trusted.afr.dirty=0x000000000000000000000000
trusted.afr.remote1=0x000000000000000000000008
trusted.afr.remote2=0x000000000000000000000000
trusted.afr.share-client-0=0x000000000000000000000000
trusted.afr.share-client-1=0x000000000000000000000000
trusted.gfid=0xb13199a1464c44918464444b3f7eeee3
trusted.glusterfs.dht=0x000000010000000000000000ffffffff

------

Regards,
AT

On Mon, Jan 4, 2016 at 12:44 PM, Krutika Dhananjay <[email protected] <mailto:[email protected]>> wrote:

    Hi,

    Could you share the output of
    # getfattr -d -m . -e hex <abs-path-to-media/ga/live/a>

    from both the bricks?

    -Krutika
    ------------------------------------------------------------------------

        *From: *"Andreas Tsaridas" <[email protected]
        <mailto:[email protected]>>
        *To: *[email protected] <mailto:[email protected]>
        *Sent: *Monday, January 4, 2016 5:10:58 PM
        *Subject: *[Gluster-users] folder not being healed


        Hello,

        I have a cluster of two replicated nodes in glusterfs 3.6.3 in
        RedHat 6.6. Problem is that a specific folder is always trying
        to be healed but never gets healed. This has been going on for
        2 weeks now.

        -----

        # gluster volume status
        Status of volume: share
        Gluster processPortOnlinePid
        
------------------------------------------------------------------------------
        Brick 172.16.4.1:/srv/share/glusterfs49152Y10416
        Brick 172.16.4.2:/srv/share/glusterfs49152Y19907
        NFS Server on localhost2049Y22664
        Self-heal Daemon on localhostN/AY22676
        NFS Server on 172.16.4.22049Y19923
        Self-heal Daemon on 172.16.4.2N/AY19937

        Task Status of Volume share
        
------------------------------------------------------------------------------
        There are no active volume tasks

        ------

        # gluster volume info

        Volume Name: share
        Type: Replicate
        Volume ID: 17224664-645c-48b7-bc3a-b8fc84c6ab30
        Status: Started
        Number of Bricks: 1 x 2 = 2
        Transport-type: tcp
        Bricks:
        Brick1: 172.16.4.1:/srv/share/glusterfs
        Brick2: 172.16.4.2:/srv/share/glusterfs
        Options Reconfigured:
        cluster.background-self-heal-count: 20
        cluster.heal-timeout: 2
        performance.normal-prio-threads: 64
        performance.high-prio-threads: 64
        performance.least-prio-threads: 64
        performance.low-prio-threads: 64
        performance.flush-behind: off
        performance.io-thread-count: 64

        ------

        # gluster volume heal share info
        Brick web01.rsdc:/srv/share/glusterfs/
        /media/ga/live/a - Possibly undergoing heal

        Number of entries: 1

        Brick web02.rsdc:/srv/share/glusterfs/
        Number of entries: 0

        -------

        # gluster volume heal share info split-brain
        Gathering list of split brain entries on volume share has been
        successful

        Brick 172.16.4.1:/srv/share/glusterfs
        Number of entries: 0

        Brick 172.16.4.2:/srv/share/glusterfs
        Number of entries: 0

        -------

        ==> /var/log/glusterfs/glustershd.log <==
        [2016-01-04 11:35:33.004831] I
        [afr-self-heal-entry.c:554:afr_selfheal_entry_do]
        0-share-replicate-0: performing entry selfheal on
        b13199a1-464c-4491-8464-444b3f7eeee3
        [2016-01-04 11:36:07.449192] W
        [client-rpc-fops.c:2772:client3_3_lookup_cbk]
        0-share-client-1: remote operation failed: No data available.
        Path: (null) (00000000-0000-0000-0000-000000000000)
        [2016-01-04 11:36:07.449706] W
        [client-rpc-fops.c:240:client3_3_mknod_cbk] 0-share-client-1:
        remote operation failed: File exists. Path: (null)

        Could you please advise ?

        Kind regards,

        AT

        _______________________________________________
        Gluster-users mailing list
        [email protected] <mailto:[email protected]>
        http://www.gluster.org/mailman/listinfo/gluster-users





_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users

_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to