Hi David,

it simply checks the inode usage. There may be enough disk space, but
no more inodes available, e.g.:

Filesystem               Inodes    IUsed      IFree IUse% Mounted on
/dev/mapper/vg0-root     655360    59698     595662   10% /

If IUse% is at 100%... ;-)


Regards,
Hubert

Am Fr., 6. März 2020 um 09:20 Uhr schrieb David Cunningham
<dcunning...@voisonics.com>:
>
> Hi Hu.
>
> Just to clarify, what should we be looking for with "df -i"?
>
>
> On Fri, 6 Mar 2020 at 18:51, Hu Bert <revi...@googlemail.com> wrote:
>>
>> Hi,
>>
>> just a guess and easy to test/try: inodes? df -i?
>>
>> regards,
>> Hubert
>>
>> Am Fr., 6. März 2020 um 04:42 Uhr schrieb David Cunningham
>> <dcunning...@voisonics.com>:
>> >
>> > Hi Aravinda,
>> >
>> > That's what was reporting 54% used, at the same time that GlusterFS was 
>> > giving no space left on device errors. It's a bit worrying that they're 
>> > not reporting the same thing.
>> >
>> > Thank you.
>> >
>> >
>> > On Fri, 6 Mar 2020 at 16:33, Aravinda VK <aravi...@kadalu.io> wrote:
>> >>
>> >> Hi David,
>> >>
>> >> What is it reporting for brick’s `df` output?
>> >>
>> >> ```
>> >> df /nodirectwritedata/gluster/gvol0
>> >> ```
>> >>
>> >> —
>> >> regards
>> >> Aravinda Vishwanathapura
>> >> https://kadalu.io
>> >>
>> >> On 06-Mar-2020, at 2:52 AM, David Cunningham <dcunning...@voisonics.com> 
>> >> wrote:
>> >>
>> >> Hello,
>> >>
>> >> A major concern we have is that "df" was reporting only 54% used and yet 
>> >> GlusterFS was giving "No space left on device" errors. We rely on "df" to 
>> >> report the correct result to monitor the system and ensure stability. 
>> >> Does anyone know what might have been going on here?
>> >>
>> >> Thanks in advance.
>> >>
>> >>
>> >> On Thu, 5 Mar 2020 at 21:35, David Cunningham <dcunning...@voisonics.com> 
>> >> wrote:
>> >>>
>> >>> Hi Aravinda,
>> >>>
>> >>> Thanks for the reply. This test server is indeed the master server for 
>> >>> geo-replication to a slave.
>> >>>
>> >>> I'm really surprised that geo-replication simply keeps writing logs 
>> >>> until all space is consumed, without cleaning them up itself. I didn't 
>> >>> see any warning about it in the geo-replication install documentation 
>> >>> which is unfortunate. We'll come up with a solution to delete log files 
>> >>> older than the LAST_SYNCED time in the geo-replication status. Is anyone 
>> >>> aware of any other potential gotchas like this?
>> >>>
>> >>> Does anyone have an idea why in my previous note some space in the 2GB 
>> >>> GlusterFS partition apparently went missing? We had 0.47GB of data, 1GB 
>> >>> reported used by .glusterfs, which even if they were separate files 
>> >>> would only add up to 1.47GB used, meaning 0.53GB should have been left 
>> >>> in the partition. If less space is actually being used because of the 
>> >>> hard links then it's even harder to understand where the other 1.53GB 
>> >>> went. So why would GlusterFS report "No space left on device"?
>> >>>
>> >>> Thanks again for any assistance.
>> >>>
>> >>>
>> >>> On Thu, 5 Mar 2020 at 17:31, Aravinda VK <aravi...@kadalu.io> wrote:
>> >>>>
>> >>>> Hi David,
>> >>>>
>> >>>> Is this Volume is uses Geo-replication? Geo-replication feature enables 
>> >>>> Changelog to identify the latest changes happening in the GlusterFS 
>> >>>> volume.
>> >>>>
>> >>>> Content of .glusterfs directory also includes hardlinks to the actual 
>> >>>> data, so the size shown in .glusterfs is including data. Please refer 
>> >>>> the comment by Xavi 
>> >>>> https://github.com/gluster/glusterfs/issues/833#issuecomment-594436009
>> >>>>
>> >>>> If Changelogs files are causing issue, you can use archival tool to 
>> >>>> remove processed changelogs.
>> >>>> https://github.com/aravindavk/archive_gluster_changelogs
>> >>>>
>> >>>> —
>> >>>> regards
>> >>>> Aravinda Vishwanathapura
>> >>>> https://kadalu.io
>> >>>>
>> >>>>
>> >>>> On 05-Mar-2020, at 9:02 AM, David Cunningham 
>> >>>> <dcunning...@voisonics.com> wrote:
>> >>>>
>> >>>> Hello,
>> >>>>
>> >>>> We are looking for some advice on disk use. This is on a single node 
>> >>>> GlusterFS test server.
>> >>>>
>> >>>> There's a 2GB partition for GlusterFS. Of that, 470MB is used for 
>> >>>> actual data, and 1GB is used by the .glusterfs directory. The 
>> >>>> .glusterfs directory is mostly used by the two-character directories 
>> >>>> and the "changelogs" directory. Why is so much used by .glusterfs, and 
>> >>>> can we reduce that overhead?
>> >>>>
>> >>>> We also have a problem with this test system where GlusterFS is giving 
>> >>>> "No space left on device" errors. That's despite "df" reporting only 
>> >>>> 54% used, and even if we add the 470MB to 1GB used above, that still 
>> >>>> comes out to less than the 2GB available, so there should be some spare.
>> >>>>
>> >>>> Would anyone be able to advise on these please? Thank you in advance.
>> >>>>
>> >>>> The GlusterFS version is 5.11 and here is the volume information:
>> >>>>
>> >>>> Volume Name: gvol0
>> >>>> Type: Distribute
>> >>>> Volume ID: 33ed309b-0e63-4f9a-8132-ab1b0fdcbc36
>> >>>> Status: Started
>> >>>> Snapshot Count: 0
>> >>>> Number of Bricks: 1
>> >>>> Transport-type: tcp
>> >>>> Bricks:
>> >>>> Brick1: myhost:/nodirectwritedata/gluster/gvol0
>> >>>> Options Reconfigured:
>> >>>> transport.address-family: inet
>> >>>> nfs.disable: on
>> >>>> geo-replication.indexing: on
>> >>>> geo-replication.ignore-pid-check: on
>> >>>> changelog.changelog: on
>> >>>>
>> >>>> --
>> >>>> David Cunningham, Voisonics Limited
>> >>>> http://voisonics.com/
>> >>>> USA: +1 213 221 1092
>> >>>> New Zealand: +64 (0)28 2558 3782
>> >>>> ________
>> >>>>
>> >>>>
>> >>>>
>> >>>> Community Meeting Calendar:
>> >>>>
>> >>>> Schedule -
>> >>>> Every Tuesday at 14:30 IST / 09:00 UTC
>> >>>> Bridge: https://bluejeans.com/441850968
>> >>>>
>> >>>> Gluster-users mailing list
>> >>>> Gluster-users@gluster.org
>> >>>> https://lists.gluster.org/mailman/listinfo/gluster-users
>> >>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>
>> >>>
>> >>> --
>> >>> David Cunningham, Voisonics Limited
>> >>> http://voisonics.com/
>> >>> USA: +1 213 221 1092
>> >>> New Zealand: +64 (0)28 2558 3782
>> >>
>> >>
>> >>
>> >> --
>> >> David Cunningham, Voisonics Limited
>> >> http://voisonics.com/
>> >> USA: +1 213 221 1092
>> >> New Zealand: +64 (0)28 2558 3782
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >
>> >
>> > --
>> > David Cunningham, Voisonics Limited
>> > http://voisonics.com/
>> > USA: +1 213 221 1092
>> > New Zealand: +64 (0)28 2558 3782
>> > ________
>> >
>> >
>> >
>> > Community Meeting Calendar:
>> >
>> > Schedule -
>> > Every Tuesday at 14:30 IST / 09:00 UTC
>> > Bridge: https://bluejeans.com/441850968
>> >
>> > Gluster-users mailing list
>> > Gluster-users@gluster.org
>> > https://lists.gluster.org/mailman/listinfo/gluster-users
>
>
>
> --
> David Cunningham, Voisonics Limited
> http://voisonics.com/
> USA: +1 213 221 1092
> New Zealand: +64 (0)28 2558 3782
________



Community Meeting Calendar:

Schedule -
Every Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://bluejeans.com/441850968

Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to