On 07/02/2019 20:21, Marc Roos wrote:
I also do not exactly know how many I have. It is sort of test setup and
the bash script creates a snapshot every day. So with 100 dirs it will
a maximum of 700. But the script first checks if there is any data with
getfattr --only-values --absolute-names -d -m ceph.dir.rbytes

I don't know what 'leaking old snapshots forever' means, how do I check
  this is happening? I am quite confident that the bash script only
creates and removes the snap dirs as it should.

Is it not strange that the snaps are shown on fs data pools I am not
  using? fs_data has indeed snapshots, fs_data.ec21.ssd is empty

I think the snapshots IDs will apply to all pools in the FS regardless of whether they contain any data referenced by the snapshots.

I just tested this and it seems each CephFS snapshot consumes two snapshots in the underlying pools, one apparently created on deletion (I wasn't aware of this). So for ~700 snapshots the output you're seeing is normal. It seems that using a "rolling snapshot" pattern in CephFS inherently creates a "one present, one deleted" pattern in the underlying pools.

Hector Martin (hec...@marcansoft.com)
Public Key: https://mrcn.st/pub
ceph-users mailing list

Reply via email to