Hi Ben,
we are not using EC pool on that cluster.
OSD out behavior almost stopped when we solved memory issues (less memory
allocated to OSD's).
Now we are not working on that cluster anymore so we have no other info about
that problem.
Jan
On 20/07/2020 07.59, Benoît Knecht wrote:
Hi Jan,
I understand, so I expected slow requests (like X slow requests are blocked > 32 sec) but I was not expecting that heartbeats are missed or
OSD's were restarted.
Maybe this "hard recovery" was not tested enough.
Also I'm concerned, that this OSD restart caused data degradation and recovery -
This is an expensive operation. You want to slow it down, not burden the OSDs.
> On Mar 21, 2020, at 5:46 AM, Jan Pekař - Imatic wrote:
>
> Each node has 64GB RAM so it should be enough (12 OSD's = 48GB used).
>
>> On 21/03/2020 13.14, XuYun wrote:
>> Bluestore requires more than 4G memory
We had a similar problem that caused by insufficient RAM: we have 6 OSDs and
32G RAM per host, and somehow swap partition was used by OS, which lead
sporadic performance problem.
> 2020年3月21日 下午8:45,Jan Pekař - Imatic 写道:
>
> Each node has 64GB RAM so it should be enough (12 OSD's = 48GB
Each node has 64GB RAM so it should be enough (12 OSD's = 48GB used).
On 21/03/2020 13.14, XuYun wrote:
Bluestore requires more than 4G memory per OSD, do you have enough memory?
2020年3月21日 下午8:09,Jan Pekař - Imatic 写道:
Hello,
I have ceph cluster version 14.2.7
Bluestore requires more than 4G memory per OSD, do you have enough memory?
> 2020年3月21日 下午8:09,Jan Pekař - Imatic 写道:
>
> Hello,
>
> I have ceph cluster version 14.2.7 (3d58626ebeec02d8385a4cefb92c6cbc3a45bfe8)
> nautilus (stable)
>
> 4 nodes - each node 11 HDD, 1 SSD, 10Gbit network
>
>