On 07/02/2019 20:21, Marc Roos wrote:
I also do not exactly know how many I have. It is sort of test setup and
the bash script creates a snapshot every day. So with 100 dirs it will
be
a maximum of 700. But the script first checks if there is any data with
getfattr --only-values --absolute-names -d -m ceph.dir.rbytes
I don't know what 'leaking old snapshots forever' means, how do I check
this is happening? I am quite confident that the bash script only
creates and removes the snap dirs as it should.
Is it not strange that the snaps are shown on fs data pools I am not
using? fs_data has indeed snapshots, fs_data.ec21.ssd is empty
I think the snapshots IDs will apply to all pools in the FS regardless
of whether they contain any data referenced by the snapshots.
I just tested this and it seems each CephFS snapshot consumes two
snapshots in the underlying pools, one apparently created on deletion (I
wasn't aware of this). So for ~700 snapshots the output you're seeing is
normal. It seems that using a "rolling snapshot" pattern in CephFS
inherently creates a "one present, one deleted" pattern in the
underlying pools.
--
Hector Martin (hec...@marcansoft.com)
Public Key: https://mrcn.st/pub
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com