I ran into a problem removing the cache tier. I tried everything I could to get past it, but I ended up having to re-enable it. I'm running on 12.2.2 with all bluestore OSDs.
I successfully set allow_ec_overwrites to true, I set the cache-mode to forward, I flushed/evicted the entire cache, and then went to remove-overlay on the data pool and received the error "Error EBUSY: pool 'cephfs_data' is in use by CephFS via its tier". I made sure that no client had cephfs mounted, I even stopped the MDS daemons, and the same error came up every time. I also tested setting the cache-mode to none (from forward) after I ensured that the cache was empty and it told me "set cache-mode for pool 'cephfs_cache' to none (WARNING: pool is still configured as read or write tier)" and still had the same error for removing the overlay. I ultimately had to concede defeat and set the cache-mode back to writeback to get things working again. Does anyone have any ideas for how to remove this cache tier? Having full write on reads is not something I really want to keep around if I can get rid of it. On Mon, Jan 22, 2018 at 7:25 AM David Turner <drakonst...@gmail.com> wrote: > I've already migrated all osds to bluestore and changed my pools to use a > crush rule specifying them to use an HDD class (forced about half of my > data to move). This week I'm planning to add in some new SSDs to move the > metadata pool to. > > I have experience with adding and removing cache tiers without losing data > in the underlying pool. The documentation on this in the upgrade procedure > and from the EC documentation had me very leary. Seeing the information > about EC pools from the CephFS documentation helps me to feel much more > confident. Thank you. > > On Mon, Jan 22, 2018, 5:53 AM John Spray <jsp...@redhat.com> wrote: > >> On Sat, Jan 20, 2018 at 6:26 PM, David Turner <drakonst...@gmail.com> >> wrote: >> > I am not able to find documentation for how to convert an existing >> cephfs >> > filesystem to use allow_ec_overwrites. The documentation says that the >> > metadata pool needs to be replicated, but that the data pool can be EC. >> But >> > it says, "For Cephfs, using an erasure coded pool means setting that >> pool in >> > a file layout." Is that really necessary if your metadata pool is >> replicated >> > and you have an existing EC pool for the data? Could I just enable ec >> > overwrites and start flushing/removing the cache tier and be on my way >> to >> > just using an EC pool? >> >> That snipped in the RADOS docs is a bit misleading: you only need to >> use a file layout if you're adding an EC pool as an addition pool >> rather than using it during creation of a filesystem. >> >> The CephFS version of events is here: >> >> http://docs.ceph.com/docs/master/cephfs/createfs/#using-erasure-coded-pools-with-cephfs >> >> As for migrating from a cache tiered configuration, I haven't tried >> it, but there's nothing CephFS-specific about it. If the underlying >> pool that's set as the cephfs data pool is EC and has >> allow_ec_overwrites then CephFS won't care -- but I'm personally not >> an expert on what knobs and buttons to use to migrate away from a >> cache tiered config. >> >> Do bear in mind that your OSDs need to be using bluestore (which may >> not be the case since you're talking about migrating an existing >> system?) >> >> John >> >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@lists.ceph.com >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > >> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com