If you have no way to recover the drives, you can try to reboot the OSDs
with `osd_find_best_info_ignore_history_les = true` (revert it afterwards),
you'll lose data. If after this, the PGs are down, you can mark the OSDs
blocking the PGs from become active lost.

On Sat, Mar 2, 2019 at 6:08 AM Daniel K <satha...@gmail.com> wrote:

> They all just started having read errors. Bus resets. Slow reads. Which is
> one of the reasons the cluster didn't recover fast enough to compensate.
>
> I tried to be mindful of the drive type and specifically avoided the
> larger capacity Seagates that are SMR. Used 1 SM863 for every 6 drives for
> the WAL.
>
> Not sure why they failed. The data isn't critical at this point, just need
> to get the cluster back to normal.
>
> On Sat, Mar 2, 2019, 9:00 AM <jes...@krogh.cc> wrote:
>
>> Did they break, or did something went wronng trying to replace them?
>>
>> Jespe
>>
>>
>>
>> Sent from myMail for iOS
>>
>>
>> Saturday, 2 March 2019, 14.34 +0100 from Daniel K <satha...@gmail.com>:
>>
>> I bought the wrong drives trying to be cheap. They were 2TB WD Blue
>> 5400rpm 2.5 inch laptop drives.
>>
>> They've been replace now with HGST 10K 1.8TB SAS drives.
>>
>>
>>
>> On Sat, Mar 2, 2019, 12:04 AM <jes...@krogh.cc> wrote:
>>
>>
>>
>> Saturday, 2 March 2019, 04.20 +0100 from satha...@gmail.com <
>> satha...@gmail.com>:
>>
>> 56 OSD, 6-node 12.2.5 cluster on Proxmox
>>
>> We had multiple drives fail(about 30%) within a few days of each other,
>> likely faster than the cluster could recover.
>>
>>
>> Hov did so many drives break?
>>
>> Jesper
>>
>> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to