Sorry, just tried "osd_recovery_sleep=0" (was 0.05) and the number of objects
in flight did increase dramatically:
recovery: 0 B/s, 8.64 kobjects/s
Would be nice if there was a way to set this per pool.
Best regards,
=================
Frank Schilder
AIT Risø Campus
Bygning 109, rum S14
________________________________________
From: Frank Schilder <[email protected]>
Sent: 03 December 2020 12:53:28
To: 胡 玮文
Cc: [email protected]
Subject: [ceph-users] Re: Increase number of objects in flight during recovery
Did this already. It doesn't change the number of objects in flight.
Best regards,
=================
Frank Schilder
AIT Risø Campus
Bygning 109, rum S14
________________________________________
From: 胡 玮文 <[email protected]>
Sent: 03 December 2020 12:35:03
To: Frank Schilder
Cc: [email protected]
Subject: Re: [ceph-users] Increase number of objects in flight during recovery
Hi,
There is a “OSD recovery priority” dialog box in web dashboard. Configurations
it will change includes:
osd_max_backfill
osd_recovery_max_active
osd_recovery_max_single_start
osd_recovery_sleep
Tune these config may helps. “High” priority corresponding to 4, 4, 4, 0,
respectively. Some of these also have a _ssd/_hdd variant.
> 在 2020年12月3日,17:11,Frank Schilder <[email protected]> 写道:
>
> Hi all,
>
> I have the opposite problem as discussed in "slow down keys/s in recovery". I
> need to increase the number of objects in flight during rebalance. It is
> already all remapped PGs in state backfilling, but it looks like no more than
> 8 objects/sec are transferred per PG at a time. The pools sits on
> high-performance SSDs and could easily handle a transfer of 100 or more
> objects/sec simultaneously. Is there any way to increase the number of
> transfers/sec or simultaneous transfers? Increasing the options
> osd_max_backfills and osd_recovery_max_active has no effect.
>
> Background: The pool in question (con-fs2-meta2) is the default data pool of
> a ceph fs, which stores exclusively the kind of meta data that goes into this
> pool. Storage consumption is reported as 0, but the number of objects is huge:
>
> NAME ID USED %USED MAX AVAIL OBJECTS
> con-fs2-meta1 12 216 MiB 0.02 933 GiB
> 13311115
> con-fs2-meta2 13 0 B 0 933 GiB
> 118389897
> con-fs2-data 14 698 TiB 72.15 270 TiB
> 286826739
>
> Unfortunately, there were no recommendations on dimensioning PG numbers for
> this pool, so I used the same for con-fs2-meta1, and con-fs2-meta2. In
> hindsight, this was potentially a bad idea, the meta2 pool should have a much
> higher PG count or a much more aggressive recovery policy.
>
> I now need to rebalance PGs on meta2 and it is going way too slow compared
> with the performance of the SSDs it is located on. In a way, I would like to
> keep the PG count where it is, but increase the recovery rate for this pool
> by a factor of 10. Please let me know what options I have.
>
> Best regards,
> =================
> Frank Schilder
> AIT Risø Campus
> Bygning 109, rum S14
> _______________________________________________
> ceph-users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]