Hi, just a guess and easy to test/try: inodes? df -i?
regards, Hubert Am Fr., 6. März 2020 um 04:42 Uhr schrieb David Cunningham <dcunning...@voisonics.com>: > > Hi Aravinda, > > That's what was reporting 54% used, at the same time that GlusterFS was > giving no space left on device errors. It's a bit worrying that they're not > reporting the same thing. > > Thank you. > > > On Fri, 6 Mar 2020 at 16:33, Aravinda VK <aravi...@kadalu.io> wrote: >> >> Hi David, >> >> What is it reporting for brick’s `df` output? >> >> ``` >> df /nodirectwritedata/gluster/gvol0 >> ``` >> >> — >> regards >> Aravinda Vishwanathapura >> https://kadalu.io >> >> On 06-Mar-2020, at 2:52 AM, David Cunningham <dcunning...@voisonics.com> >> wrote: >> >> Hello, >> >> A major concern we have is that "df" was reporting only 54% used and yet >> GlusterFS was giving "No space left on device" errors. We rely on "df" to >> report the correct result to monitor the system and ensure stability. Does >> anyone know what might have been going on here? >> >> Thanks in advance. >> >> >> On Thu, 5 Mar 2020 at 21:35, David Cunningham <dcunning...@voisonics.com> >> wrote: >>> >>> Hi Aravinda, >>> >>> Thanks for the reply. This test server is indeed the master server for >>> geo-replication to a slave. >>> >>> I'm really surprised that geo-replication simply keeps writing logs until >>> all space is consumed, without cleaning them up itself. I didn't see any >>> warning about it in the geo-replication install documentation which is >>> unfortunate. We'll come up with a solution to delete log files older than >>> the LAST_SYNCED time in the geo-replication status. Is anyone aware of any >>> other potential gotchas like this? >>> >>> Does anyone have an idea why in my previous note some space in the 2GB >>> GlusterFS partition apparently went missing? We had 0.47GB of data, 1GB >>> reported used by .glusterfs, which even if they were separate files would >>> only add up to 1.47GB used, meaning 0.53GB should have been left in the >>> partition. If less space is actually being used because of the hard links >>> then it's even harder to understand where the other 1.53GB went. So why >>> would GlusterFS report "No space left on device"? >>> >>> Thanks again for any assistance. >>> >>> >>> On Thu, 5 Mar 2020 at 17:31, Aravinda VK <aravi...@kadalu.io> wrote: >>>> >>>> Hi David, >>>> >>>> Is this Volume is uses Geo-replication? Geo-replication feature enables >>>> Changelog to identify the latest changes happening in the GlusterFS volume. >>>> >>>> Content of .glusterfs directory also includes hardlinks to the actual >>>> data, so the size shown in .glusterfs is including data. Please refer the >>>> comment by Xavi >>>> https://github.com/gluster/glusterfs/issues/833#issuecomment-594436009 >>>> >>>> If Changelogs files are causing issue, you can use archival tool to remove >>>> processed changelogs. >>>> https://github.com/aravindavk/archive_gluster_changelogs >>>> >>>> — >>>> regards >>>> Aravinda Vishwanathapura >>>> https://kadalu.io >>>> >>>> >>>> On 05-Mar-2020, at 9:02 AM, David Cunningham <dcunning...@voisonics.com> >>>> wrote: >>>> >>>> Hello, >>>> >>>> We are looking for some advice on disk use. This is on a single node >>>> GlusterFS test server. >>>> >>>> There's a 2GB partition for GlusterFS. Of that, 470MB is used for actual >>>> data, and 1GB is used by the .glusterfs directory. The .glusterfs >>>> directory is mostly used by the two-character directories and the >>>> "changelogs" directory. Why is so much used by .glusterfs, and can we >>>> reduce that overhead? >>>> >>>> We also have a problem with this test system where GlusterFS is giving "No >>>> space left on device" errors. That's despite "df" reporting only 54% used, >>>> and even if we add the 470MB to 1GB used above, that still comes out to >>>> less than the 2GB available, so there should be some spare. >>>> >>>> Would anyone be able to advise on these please? Thank you in advance. >>>> >>>> The GlusterFS version is 5.11 and here is the volume information: >>>> >>>> Volume Name: gvol0 >>>> Type: Distribute >>>> Volume ID: 33ed309b-0e63-4f9a-8132-ab1b0fdcbc36 >>>> Status: Started >>>> Snapshot Count: 0 >>>> Number of Bricks: 1 >>>> Transport-type: tcp >>>> Bricks: >>>> Brick1: myhost:/nodirectwritedata/gluster/gvol0 >>>> Options Reconfigured: >>>> transport.address-family: inet >>>> nfs.disable: on >>>> geo-replication.indexing: on >>>> geo-replication.ignore-pid-check: on >>>> changelog.changelog: on >>>> >>>> -- >>>> David Cunningham, Voisonics Limited >>>> http://voisonics.com/ >>>> USA: +1 213 221 1092 >>>> New Zealand: +64 (0)28 2558 3782 >>>> ________ >>>> >>>> >>>> >>>> Community Meeting Calendar: >>>> >>>> Schedule - >>>> Every Tuesday at 14:30 IST / 09:00 UTC >>>> Bridge: https://bluejeans.com/441850968 >>>> >>>> Gluster-users mailing list >>>> Gluster-users@gluster.org >>>> https://lists.gluster.org/mailman/listinfo/gluster-users >>>> >>>> >>>> >>>> >>>> >>>> >>> >>> >>> -- >>> David Cunningham, Voisonics Limited >>> http://voisonics.com/ >>> USA: +1 213 221 1092 >>> New Zealand: +64 (0)28 2558 3782 >> >> >> >> -- >> David Cunningham, Voisonics Limited >> http://voisonics.com/ >> USA: +1 213 221 1092 >> New Zealand: +64 (0)28 2558 3782 >> >> >> >> >> >> > > > -- > David Cunningham, Voisonics Limited > http://voisonics.com/ > USA: +1 213 221 1092 > New Zealand: +64 (0)28 2558 3782 > ________ > > > > Community Meeting Calendar: > > Schedule - > Every Tuesday at 14:30 IST / 09:00 UTC > Bridge: https://bluejeans.com/441850968 > > Gluster-users mailing list > Gluster-users@gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users ________ Community Meeting Calendar: Schedule - Every Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list Gluster-users@gluster.org https://lists.gluster.org/mailman/listinfo/gluster-users