Well, the default is k + 1, so 11. Could it be that you reduced it during a recovery phase but didn't set it back to the default?

Zitat von [email protected]:

no, disks utilization is around 86%.

What is safe value for min_size in this case?

18. 10. 2021 15:46:44 Eugen Block <[email protected]>:

Hi,

min_size = k is not the safest option, it should be only used in case  of disaster recovery. But in this case it's not related to IO  interruption, it seems. Are some disks utilized around 100% (iostat)  when this happens?


Zitat von Denis Polom <[email protected]>:

Hi,

it's

min_size: 10


On 10/18/21 14:43, Eugen Block wrote:
What is your min_size for the affected pool?


Zitat von Denis Polom <[email protected]>:

Hi,

I have 18 OSD nodes in this cluster. And it does happen even if  one OSD daemon goes down or flaps.

Running

ceph version 15.2.14 (cd3bb7e87a2f62c1b862ff3fd8b1eec13391a5be)  octopus (stable)

thx!


On 10/18/21 12:12, Eugen Block wrote:
Hi,

with this EC setup your pool min_size would be 11 (k+1), so in  case one host goes down (or several OSDs fail on this host), your  clients should not be affected. But as soon as a second host  fails you’ll notice IO pause until at least one host has  recovered. Do you have more than 12 hosts in this cluster so it  could recover one host failure?

Regards,
Eugen


Zitat von Denis Polom <[email protected]>:










_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]



_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to