Hi, I've been searching the list for my own report from last year maybe, when we finally got rid of our cache tier as well. I can't really find it though, so I'm gonna summarize again how we managed to get out of it.
After setting the cache mode to proxy and failing to flush/evict (also with "no such file or directory" messages), I gradually reduced the cache pool size from 260 GB in steps of 40 GB or so, not sure anymore .Suddenly Ceph was able to flush and evict again. I reduced the target_max_bytes to 1000000, only rbd_header objects remained in the pool, but those couldn't be flushed/evicted anymore: failed to evict /rbd_header.133ba5e7513046: (16) Device or resource busy rbd_header.8fbda012200854 Unfortunately, we didn't find a promsing way to get rid of them while the images were in use. I had played around in a test cluster and ignored the header objects when remocing the cache tier, and it seemed to work (since the rbd_header objects should be identical in the backing pool as well), but it didn't feel safe and we were really uncomfortable trying it in production. So in the end, we shut down our VMs to be able to flush/evict everything and safely got rid of the cache tier. I know this isn't really helpful, but maybe next time I find my comment on the list. :-) Regards, Eugen Am Do., 5. Feb. 2026 um 15:16 Uhr schrieb Kees Meijs | Nefos via ceph-users <[email protected]>: > Hello list, > > After running cache tiers for years and years, we decided to turn off > caching on a mostly dormant cluster. > > Given an installation using Pacific, we followed the instructions as > dictated on: > > https://docs.ceph.com/en/pacific/rados/operations/cache-tiering/#removing-a-writeback-cache > > So, we set the "hot" tier to cache-mode proxy (as the documentation > mentions) and simply waited for a bit. Since nothing really seemed to > happen, we did a manual flush using cache-flush-evict-all (also as the > documentation mentions). So far, so good. > > The end result is a cache tier with still quite some (>81k) objects > within. Running cache-flush-evict-all again renders errors such as: > > failed to flush /rbd_data.d9d54166ef438d.0000000000020899: (2) No such > file or directory > rbd_data.d9d54166ef438d.000000000001a511 > > This is not what is expected. > > What are we looking at? And, how to move forward in removing the cache? > > Thanks for your help and wisdom! > > Cheers, > Kees > > P.S. The introduction text speaks of readproxy and not proxy. This > confusing since the instruction text speaks of proxy. > _______________________________________________ > ceph-users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > _______________________________________________ ceph-users mailing list -- [email protected] To unsubscribe send an email to [email protected]
