We had this with older Ceph versions, maybe just try to restart all OSDs of affected PGs.
-- Martin Verges Managing director Mobile: +49 174 9335695 E-Mail: [email protected] Chat: https://t.me/MartinVerges croit GmbH, Freseniusstr. 31h, 81247 Munich CEO: Martin Verges - VAT-ID: DE310638492 Com. register: Amtsgericht Munich HRB 231263 Web: https://croit.io YouTube: https://goo.gl/PGE1Bx Am So., 3. Nov. 2019 um 20:13 Uhr schrieb Kári Bertilsson < [email protected]>: > pgs: 14.377% pgs not active > 3749681/537818808 objects misplaced (0.697%) > 810 active+clean > 156 down > 124 active+remapped+backfilling > 1 active+remapped+backfill_toofull > 1 down+inconsistent > > when looking at the down pg's all disks are online > > 41.3db 53775 0 0 0 401643186092 0 > 0 3044 down 6m 161222'303144 162913:4630171 > [32,96,128,115,86,129,113,124,57,109]p32 > [32,96,128,115,86,129,113,124,57,109]p32 2019-11-03 > > Any way to see why the pg is down ? > _______________________________________________ > ceph-users mailing list -- [email protected] > To unsubscribe send an email to [email protected] >
_______________________________________________ ceph-users mailing list -- [email protected] To unsubscribe send an email to [email protected]
