I also switched the cache tier to "readproxy", to avoid using this cache. But, it's still blocked.
You could change the cache mode to "none" to disable it. Could you paste the output of:
ceph osd pool ls detail | grep cache-bkp-foo Zitat von Olivier Bonvalet <[email protected]>:
In fact, one object (only one) seem to be blocked on the cache tier (writeback). I tried to flush the cache with "rados -p cache-bkp-foo cache-flush- evict-all", but it blocks on the object "rbd_data.f66c92ae8944a.00000000000f2596". So I reduced (a lot) the cache tier to 200MB, "rados -p cache-bkp-foo ls" now show only 3 objects : rbd_directory rbd_data.f66c92ae8944a.00000000000f2596 rbd_header.f66c92ae8944a And "cache-flush-evict-all" still hangs. I also switched the cache tier to "readproxy", to avoid using this cache. But, it's still blocked. Le vendredi 21 septembre 2018 à 02:14 +0200, Olivier Bonvalet a écrit :Hello, on a Luminous cluster, I have a PG incomplete and I can't find how to fix that. It's an EC pool (4+2) : pg 37.9c is incomplete, acting [32,50,59,1,0,75] (reducing pool bkp-sb-raid6 min_size from 4 may help; search ceph.com/docs for 'incomplete') Of course, we can't reduce min_size from 4. And the full state : https://pastebin.com/zrwu5X0w So, IO are blocked, we can't access thoses damaged data. OSD blocks too : osds 32,68,69 have stuck requests > 4194.3 sec OSD 32 is the primary of this PG. And OSD 68 and 69 are for cache tiering. Any idea how can I fix that ? Thanks, Olivier _______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
