Hi,

just a guess and easy to test/try: inodes? df -i?

regards,
Hubert

Am Fr., 6. März 2020 um 04:42 Uhr schrieb David Cunningham
<dcunning...@voisonics.com>:
>
> Hi Aravinda,
>
> That's what was reporting 54% used, at the same time that GlusterFS was 
> giving no space left on device errors. It's a bit worrying that they're not 
> reporting the same thing.
>
> Thank you.
>
>
> On Fri, 6 Mar 2020 at 16:33, Aravinda VK <aravi...@kadalu.io> wrote:
>>
>> Hi David,
>>
>> What is it reporting for brick’s `df` output?
>>
>> ```
>> df /nodirectwritedata/gluster/gvol0
>> ```
>>
>> —
>> regards
>> Aravinda Vishwanathapura
>> https://kadalu.io
>>
>> On 06-Mar-2020, at 2:52 AM, David Cunningham <dcunning...@voisonics.com> 
>> wrote:
>>
>> Hello,
>>
>> A major concern we have is that "df" was reporting only 54% used and yet 
>> GlusterFS was giving "No space left on device" errors. We rely on "df" to 
>> report the correct result to monitor the system and ensure stability. Does 
>> anyone know what might have been going on here?
>>
>> Thanks in advance.
>>
>>
>> On Thu, 5 Mar 2020 at 21:35, David Cunningham <dcunning...@voisonics.com> 
>> wrote:
>>>
>>> Hi Aravinda,
>>>
>>> Thanks for the reply. This test server is indeed the master server for 
>>> geo-replication to a slave.
>>>
>>> I'm really surprised that geo-replication simply keeps writing logs until 
>>> all space is consumed, without cleaning them up itself. I didn't see any 
>>> warning about it in the geo-replication install documentation which is 
>>> unfortunate. We'll come up with a solution to delete log files older than 
>>> the LAST_SYNCED time in the geo-replication status. Is anyone aware of any 
>>> other potential gotchas like this?
>>>
>>> Does anyone have an idea why in my previous note some space in the 2GB 
>>> GlusterFS partition apparently went missing? We had 0.47GB of data, 1GB 
>>> reported used by .glusterfs, which even if they were separate files would 
>>> only add up to 1.47GB used, meaning 0.53GB should have been left in the 
>>> partition. If less space is actually being used because of the hard links 
>>> then it's even harder to understand where the other 1.53GB went. So why 
>>> would GlusterFS report "No space left on device"?
>>>
>>> Thanks again for any assistance.
>>>
>>>
>>> On Thu, 5 Mar 2020 at 17:31, Aravinda VK <aravi...@kadalu.io> wrote:
>>>>
>>>> Hi David,
>>>>
>>>> Is this Volume is uses Geo-replication? Geo-replication feature enables 
>>>> Changelog to identify the latest changes happening in the GlusterFS volume.
>>>>
>>>> Content of .glusterfs directory also includes hardlinks to the actual 
>>>> data, so the size shown in .glusterfs is including data. Please refer the 
>>>> comment by Xavi 
>>>> https://github.com/gluster/glusterfs/issues/833#issuecomment-594436009
>>>>
>>>> If Changelogs files are causing issue, you can use archival tool to remove 
>>>> processed changelogs.
>>>> https://github.com/aravindavk/archive_gluster_changelogs
>>>>
>>>> —
>>>> regards
>>>> Aravinda Vishwanathapura
>>>> https://kadalu.io
>>>>
>>>>
>>>> On 05-Mar-2020, at 9:02 AM, David Cunningham <dcunning...@voisonics.com> 
>>>> wrote:
>>>>
>>>> Hello,
>>>>
>>>> We are looking for some advice on disk use. This is on a single node 
>>>> GlusterFS test server.
>>>>
>>>> There's a 2GB partition for GlusterFS. Of that, 470MB is used for actual 
>>>> data, and 1GB is used by the .glusterfs directory. The .glusterfs 
>>>> directory is mostly used by the two-character directories and the 
>>>> "changelogs" directory. Why is so much used by .glusterfs, and can we 
>>>> reduce that overhead?
>>>>
>>>> We also have a problem with this test system where GlusterFS is giving "No 
>>>> space left on device" errors. That's despite "df" reporting only 54% used, 
>>>> and even if we add the 470MB to 1GB used above, that still comes out to 
>>>> less than the 2GB available, so there should be some spare.
>>>>
>>>> Would anyone be able to advise on these please? Thank you in advance.
>>>>
>>>> The GlusterFS version is 5.11 and here is the volume information:
>>>>
>>>> Volume Name: gvol0
>>>> Type: Distribute
>>>> Volume ID: 33ed309b-0e63-4f9a-8132-ab1b0fdcbc36
>>>> Status: Started
>>>> Snapshot Count: 0
>>>> Number of Bricks: 1
>>>> Transport-type: tcp
>>>> Bricks:
>>>> Brick1: myhost:/nodirectwritedata/gluster/gvol0
>>>> Options Reconfigured:
>>>> transport.address-family: inet
>>>> nfs.disable: on
>>>> geo-replication.indexing: on
>>>> geo-replication.ignore-pid-check: on
>>>> changelog.changelog: on
>>>>
>>>> --
>>>> David Cunningham, Voisonics Limited
>>>> http://voisonics.com/
>>>> USA: +1 213 221 1092
>>>> New Zealand: +64 (0)28 2558 3782
>>>> ________
>>>>
>>>>
>>>>
>>>> Community Meeting Calendar:
>>>>
>>>> Schedule -
>>>> Every Tuesday at 14:30 IST / 09:00 UTC
>>>> Bridge: https://bluejeans.com/441850968
>>>>
>>>> Gluster-users mailing list
>>>> Gluster-users@gluster.org
>>>> https://lists.gluster.org/mailman/listinfo/gluster-users
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> David Cunningham, Voisonics Limited
>>> http://voisonics.com/
>>> USA: +1 213 221 1092
>>> New Zealand: +64 (0)28 2558 3782
>>
>>
>>
>> --
>> David Cunningham, Voisonics Limited
>> http://voisonics.com/
>> USA: +1 213 221 1092
>> New Zealand: +64 (0)28 2558 3782
>>
>>
>>
>>
>>
>>
>
>
> --
> David Cunningham, Voisonics Limited
> http://voisonics.com/
> USA: +1 213 221 1092
> New Zealand: +64 (0)28 2558 3782
> ________
>
>
>
> Community Meeting Calendar:
>
> Schedule -
> Every Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://bluejeans.com/441850968
>
> Gluster-users mailing list
> Gluster-users@gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
________



Community Meeting Calendar:

Schedule -
Every Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://bluejeans.com/441850968

Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to