WHA? Mind blown.  I hadn't noticed that you can reduce PG counts now!  Thanks 
Richard for pointing that out.  I've already reduced the pgs in that unused 
pool to half of what it was but I think the other backfill operations have 
blocked that but for the moment I think the system is ok.... at least for the 
weekend.

Thanks for pointing that out.

-Dave

On 2022-05-05 1:09 p.m., Richard Bade wrote:
[â–³EXTERNAL]


Hi David,
Something else you could try with that other pool, if it contains little or no 
data, is to reduce the PG number. This does cause some backfill operations as 
it does a pg merge but this doesn't take long if the pg is virtually empty. The 
autoscaler has a mode where it can make recommendations for you without 
actually doing anything if you want some advice on a suitable number. Then you 
can set it manually.
If the empty pg's are a factor in the balance issues then this will help.

Also, the upmap mode on the balancer is far more effective than reweight. It 
has an option where you can control the max deviation. I have this set to one 
and it achieves a 5% spread for my EC cluster.

Note, you'll need to reweight everything back to 1, which will cause backfill 
to occur.
If you have your backfill_full level set to default this should stop any osd's 
over 85% doing any backfill.

Rich
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to