Thanks Alexander! That did it. -Dave
On Fri, Sep 8, 2023 at 7:55 PM Dave S <[email protected]> wrote: > > OOH Thanks! I'll certainly give that a try. > > On Fri, Sep 8, 2023 at 7:50 PM Alexander E. Patrakov <[email protected]> > wrote: > > > > Hello Dave, > > > > I think your data is still intact. Nautilus, indeed, had issues when > > recovering erasure-coded pools. You can try temporarily setting min_size to > > 4. This bug has been fixed in Octopus or later releases. From the release > > notes at https://docs.ceph.com/en/latest/releases/octopus/: > > > >> Ceph will allow recovery below min_size for Erasure coded pools, wherever > >> possible. > > > > > > > > On Sat, Sep 9, 2023 at 5:58 AM Dave S <[email protected]> wrote: > >> > >> Hi Everyone, > >> I've been fighting with a ceph cluster that we have recently > >> physically relocated and lost 2 OSDs during the ensuing power down and > >> relocation. After powering everything back on we have > >> 3 incomplete > >> 3 remapped+incomplete > >> And indeed we have 2 OSDs that died along the way. > >> The reason I'm contacting the list is that I'm surprised that these > >> PGs are incomplete. We're running Erasure coding with K=4, M=2 which > >> in my understanding we should be able to lose 2 OSDs without an issue. > >> Am I mis-understanding this or does m=2 mean you can lose m-1 OSDs? > >> > >> Also, these two OSDs happened to be in the same server (#3 of 8 total > >> servers). > >> > >> This is an older cluster running Nautilus 14.2.9. > >> > >> Any thoughts? > >> Thanks > >> -Dave > >> _______________________________________________ > >> ceph-users mailing list -- [email protected] > >> To unsubscribe send an email to [email protected] > > > > > > > > -- > > Alexander E. Patrakov _______________________________________________ ceph-users mailing list -- [email protected] To unsubscribe send an email to [email protected]
