OK, many thanks Rajesh.
I just wanted to add that I see a lot of warnings in the logs like the 
following:

[2015-06-08 13:13:10.365633] W [marker-quota.c:3162:mq_initiate_quota_task] 
0-atlas-data-01-marker: inode ctx get failed, aborting quota txn

I’m not sure if this is a bug (related or not to the one you mention) or if it 
is normal and harmless.
Thanks,

        Alessandro


> Il giorno 08/giu/2015, alle ore 10:39, Rajesh kumar Reddy Mekala 
> <[email protected]> ha scritto:
> 
> We have open bug 1227724 for the similar problem 
> 
> Thanks,
> Rajesh
> 
> On 06/08/2015 12:08 PM, Vijaikumar M wrote:
>> Hi Alessandro,
>> 
>> Please provide the test-case, so that we can try to re-create this problem 
>> in-house?
>> 
>> Thanks,
>> Vijay
>> 
>> On Saturday 06 June 2015 05:59 AM, Alessandro De Salvo wrote:
>>> Hi,
>>> just to answer to myself, it really seems the temp files from rsync are the 
>>> culprit, it seems that their size are summed up to the real contents of the 
>>> directories I’m synchronizing, or in other terms their size is not removed 
>>> from the used size after they are removed. I suppose this is someway 
>>> connected to the error on removexattr I’m seeing. The temporary solution 
>>> I’ve found is to use rsync with the option to write the temp files to /tmp, 
>>> but it would be very interesting to understand why this is happening.
>>> Cheers,
>>> 
>>>     Alessandro
>>> 
>>>> Il giorno 06/giu/2015, alle ore 01:19, Alessandro De Salvo 
>>>> <[email protected]> 
>>>> <mailto:[email protected]> ha scritto:
>>>> 
>>>> Hi,
>>>> I currently have two brick with replica 2 on the same machine, pointing to 
>>>> different disks of a connected SAN.
>>>> The volume itself is fine:
>>>> 
>>>> # gluster volume info atlas-home-01
>>>> 
>>>> Volume Name: atlas-home-01
>>>> Type: Replicate
>>>> Volume ID: 660db960-31b8-4341-b917-e8b43070148b
>>>> Status: Started
>>>> Number of Bricks: 1 x 2 = 2
>>>> Transport-type: tcp
>>>> Bricks:
>>>> Brick1: host1:/bricks/atlas/home02/data
>>>> Brick2: host2:/bricks/atlas/home01/data
>>>> Options Reconfigured:
>>>> performance.write-behind-window-size: 4MB
>>>> performance.io-thread-count: 32
>>>> performance.readdir-ahead: on
>>>> server.allow-insecure: on
>>>> nfs.disable: true
>>>> features.quota: on
>>>> features.inode-quota: on
>>>> 
>>>> 
>>>> However, when I set a quota on a dir of the volume the size show is twice 
>>>> the physical size of the actual dir:
>>>> 
>>>> # gluster volume quota atlas-home-01 list /user1
>>>>                  Path                   Hard-limit Soft-limit   Used  
>>>> Available  Soft-limit exceeded? Hard-limit exceeded?
>>>> ---------------------------------------------------------------------------------------------------------------------------
>>>> /user1                                    4.0GB       80%       3.2GB 
>>>> 853.4MB              No                   No
>>>> 
>>>> # du -sh /storage/atlas/home/user1
>>>> 1.6G    /storage/atlas/home/user1
>>>> 
>>>> If I remove one of the bricks the quota shows the correct value.
>>>> Is there any double counting in case the bricks are on the same machine?
>>>> Also, I see a lot of errors in the logs like the following:
>>>> 
>>>> [2015-06-05 21:59:27.450407] E 
>>>> [posix-handle.c:157:posix_make_ancestryfromgfid] 0-atlas-home-01-posix: 
>>>> could not read the link from the gfid handle 
>>>> /bricks/atlas/home01/data/.glusterfs/be/e5/bee5e2b8-c639-4539-a483-96c19cd889eb
>>>>  (No such file or directory)
>>>> 
>>>> and also
>>>> 
>>>> [2015-06-05 22:52:01.112070] E [marker-quota.c:2363:mq_mark_dirty] 
>>>> 0-atlas-home-01-marker: failed to get inode ctx for /user1/file1
>>>> 
>>>> When running rsync I also see the following errors:
>>>> 
>>>> [2015-06-05 23:06:22.203968] E [marker-quota.c:2601:mq_remove_contri] 
>>>> 0-atlas-home-01-marker: removexattr 
>>>> trusted.glusterfs.quota.fddf31ba-7f1d-4ba8-a5ad-2ebd6e4030f3.contri failed 
>>>> for /user1/..bashrc.O4kekp: No data available
>>>> 
>>>> Those files are the temp files of rsync, I’m not sure why the throw errors 
>>>> in glusterfs.
>>>> Any help?
>>>> Thanks,
>>>> 
>>>>    Alessandro
>>>> 
>>>> 
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> [email protected] <mailto:[email protected]>
>>>> http://www.gluster.org/mailman/listinfo/gluster-users 
>>>> <http://www.gluster.org/mailman/listinfo/gluster-users>
>>> 
>>> 
>>> _______________________________________________
>>> Gluster-users mailing list
>>> [email protected] <mailto:[email protected]>
>>> http://www.gluster.org/mailman/listinfo/gluster-users 
>>> <http://www.gluster.org/mailman/listinfo/gluster-users>
>> 
>> 
>> _______________________________________________
>> Gluster-users mailing list
>> [email protected] <mailto:[email protected]>
>> http://www.gluster.org/mailman/listinfo/gluster-users 
>> <http://www.gluster.org/mailman/listinfo/gluster-users>

Attachment: smime.p7s
Description: S/MIME cryptographic signature

_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to