Hi,
I think that the approach with exporting and importing PGs would be
a-priori more successful than the one based on pvmove or ddrescue. The
reason is that you don't need to export/import all data that the
failed disk holds, but only the PGs that Ceph cannot recover
otherwise. The logic here
I’ve done the pg import dance a couple of times. It was very slow but did work
ultimately.
Depending on the situation, if there is one valid copy available one can enable
recovery by temporarily setting min_size on the pool to 1, reverting it once
recovery completes.
You you run with 1
Hi,
I think that the approach with exporting and importing PGs would be
a-priori more successful than the one based on pvmove or ddrescue. The
reason is that you don't need to export/import all data that the
failed disk holds, but only the PGs that Ceph cannot recover
otherwise. The logic here is
Hi,
if the OSDs are deployed as LVs (by ceph-volume) you could try to do a
pvmove to a healthy disk. There was a thread here a couple of weeks
ago explaining the steps. I don’t have it at hand right now, but it
should be easy to find.
Of course, there’s no guarantee that this will be
Hi Carl,
you might want to use ceph-objectstore-tool to export PGs from faulty
OSDs and import them back to healthy ones.
The process could be quite tricky though.
There is also pending PR (https://github.com/ceph/ceph/pull/54991) to
make the tool more tolerant to disk errors.
The patch