OK, many thanks Rajesh. I just wanted to add that I see a lot of warnings in the logs like the following:
[2015-06-08 13:13:10.365633] W [marker-quota.c:3162:mq_initiate_quota_task]
0-atlas-data-01-marker: inode ctx get failed, aborting quota txn
I’m not sure if this is a bug (related or not to the one you mention) or if it
is normal and harmless.
Thanks,
Alessandro
> Il giorno 08/giu/2015, alle ore 10:39, Rajesh kumar Reddy Mekala
> <[email protected]> ha scritto:
>
> We have open bug 1227724 for the similar problem
>
> Thanks,
> Rajesh
>
> On 06/08/2015 12:08 PM, Vijaikumar M wrote:
>> Hi Alessandro,
>>
>> Please provide the test-case, so that we can try to re-create this problem
>> in-house?
>>
>> Thanks,
>> Vijay
>>
>> On Saturday 06 June 2015 05:59 AM, Alessandro De Salvo wrote:
>>> Hi,
>>> just to answer to myself, it really seems the temp files from rsync are the
>>> culprit, it seems that their size are summed up to the real contents of the
>>> directories I’m synchronizing, or in other terms their size is not removed
>>> from the used size after they are removed. I suppose this is someway
>>> connected to the error on removexattr I’m seeing. The temporary solution
>>> I’ve found is to use rsync with the option to write the temp files to /tmp,
>>> but it would be very interesting to understand why this is happening.
>>> Cheers,
>>>
>>> Alessandro
>>>
>>>> Il giorno 06/giu/2015, alle ore 01:19, Alessandro De Salvo
>>>> <[email protected]>
>>>> <mailto:[email protected]> ha scritto:
>>>>
>>>> Hi,
>>>> I currently have two brick with replica 2 on the same machine, pointing to
>>>> different disks of a connected SAN.
>>>> The volume itself is fine:
>>>>
>>>> # gluster volume info atlas-home-01
>>>>
>>>> Volume Name: atlas-home-01
>>>> Type: Replicate
>>>> Volume ID: 660db960-31b8-4341-b917-e8b43070148b
>>>> Status: Started
>>>> Number of Bricks: 1 x 2 = 2
>>>> Transport-type: tcp
>>>> Bricks:
>>>> Brick1: host1:/bricks/atlas/home02/data
>>>> Brick2: host2:/bricks/atlas/home01/data
>>>> Options Reconfigured:
>>>> performance.write-behind-window-size: 4MB
>>>> performance.io-thread-count: 32
>>>> performance.readdir-ahead: on
>>>> server.allow-insecure: on
>>>> nfs.disable: true
>>>> features.quota: on
>>>> features.inode-quota: on
>>>>
>>>>
>>>> However, when I set a quota on a dir of the volume the size show is twice
>>>> the physical size of the actual dir:
>>>>
>>>> # gluster volume quota atlas-home-01 list /user1
>>>> Path Hard-limit Soft-limit Used
>>>> Available Soft-limit exceeded? Hard-limit exceeded?
>>>> ---------------------------------------------------------------------------------------------------------------------------
>>>> /user1 4.0GB 80% 3.2GB
>>>> 853.4MB No No
>>>>
>>>> # du -sh /storage/atlas/home/user1
>>>> 1.6G /storage/atlas/home/user1
>>>>
>>>> If I remove one of the bricks the quota shows the correct value.
>>>> Is there any double counting in case the bricks are on the same machine?
>>>> Also, I see a lot of errors in the logs like the following:
>>>>
>>>> [2015-06-05 21:59:27.450407] E
>>>> [posix-handle.c:157:posix_make_ancestryfromgfid] 0-atlas-home-01-posix:
>>>> could not read the link from the gfid handle
>>>> /bricks/atlas/home01/data/.glusterfs/be/e5/bee5e2b8-c639-4539-a483-96c19cd889eb
>>>> (No such file or directory)
>>>>
>>>> and also
>>>>
>>>> [2015-06-05 22:52:01.112070] E [marker-quota.c:2363:mq_mark_dirty]
>>>> 0-atlas-home-01-marker: failed to get inode ctx for /user1/file1
>>>>
>>>> When running rsync I also see the following errors:
>>>>
>>>> [2015-06-05 23:06:22.203968] E [marker-quota.c:2601:mq_remove_contri]
>>>> 0-atlas-home-01-marker: removexattr
>>>> trusted.glusterfs.quota.fddf31ba-7f1d-4ba8-a5ad-2ebd6e4030f3.contri failed
>>>> for /user1/..bashrc.O4kekp: No data available
>>>>
>>>> Those files are the temp files of rsync, I’m not sure why the throw errors
>>>> in glusterfs.
>>>> Any help?
>>>> Thanks,
>>>>
>>>> Alessandro
>>>>
>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> [email protected] <mailto:[email protected]>
>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>> <http://www.gluster.org/mailman/listinfo/gluster-users>
>>>
>>>
>>> _______________________________________________
>>> Gluster-users mailing list
>>> [email protected] <mailto:[email protected]>
>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>> <http://www.gluster.org/mailman/listinfo/gluster-users>
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> [email protected] <mailto:[email protected]>
>> http://www.gluster.org/mailman/listinfo/gluster-users
>> <http://www.gluster.org/mailman/listinfo/gluster-users>
smime.p7s
Description: S/MIME cryptographic signature
_______________________________________________ Gluster-users mailing list [email protected] http://www.gluster.org/mailman/listinfo/gluster-users
