[ceph-users] Re: MANY_OBJECT_PER_PG on 1 pool which is cephfs_metadata

2024-03-25 Thread e . fazenda
Dear Eugen,

Sorry i forgot to update the case.

I have upgraded to the latest pacific release 16.2.15 and i have done the 
necessary for the pg_num :) 

Thanks for the followup on this.

Topic can be closed.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: MANY_OBJECT_PER_PG on 1 pool which is cephfs_metadata

2024-03-11 Thread Eugen Block

Hi,

I assume you're still on a "low" pacific release? This was fixed by PR  
[1][2] and the warning is supressed when autoscaler is on, it was  
merged into Pacific 16.2.8 [3].


I can't answer why autoscaler doesn't increase the pg_num, but yes,  
you can increase it by yourself. The pool for cephfs_metadata should  
be on fast storage and doesn't have huge amounts of data so it should  
be relatively quick. What's your 'ceph osd df' output?


Regards,
Eugen

[1] https://tracker.ceph.com/issues/53644
[2] https://github.com/ceph/ceph/pull/45152
[3] https://docs.ceph.com/en/latest/releases/pacific/#v16-2-8-pacific

Zitat von Edouard FAZENDA :


Hello Ceph community,



I have since this morning a warning about MANY_OBJECT_PER_PG on 1 pool which
is cephfs_metadata



# ceph health detail

HEALTH_WARN 1 pools have many more objects per pg than average

[WRN] MANY_OBJECTS_PER_PG: 1 pools have many more objects per pg than
average

pool cephfs_metadata objects per pg (154151) is more than 10.0215 times
cluster average (15382)



I have the autoscaling on on all the pool :



# ceph osd pool autoscale-status

POOL  SIZE  TARGET SIZE  RATE  RAW CAPACITY   RATIO
TARGET RATIO  EFFECTIVE RATIO  BIAS  PG_NUM  NEW PG_NUM  AUTOSCALE

device_health_metrics9523k3.026827G  0.
1.0   1  on

cephfs_data  5389G2.026827G  0.4018
1.0 512  on

cephfs_metadata 19365M2.026827G  0.0014
4.0  16  on

.rgw.root1323 3.026827G  0.
1.0  32  on

default.rgw.log 23552 3.026827G  0.
1.0  32  on

default.rgw.control 0 3.026827G  0.
1.0  32  on

default.rgw.meta11911 3.026827G  0.
4.0   8  on

default.rgw.buckets.index   0 3.026827G  0.
4.0   8  on

default.rgw.buckets.data497.0G3.026827G  0.0556
1.0  32  on

kubernetes  177.2G2.026827G  0.0132
1.0  32  on

default.rgw.buckets.non-ec432 3.026827G  0.
1.0  32  on



Actually the pg_num is 16 for the cephfs_metdata pool , but it does not
define NEW_PG_NUM



Here the replicated size of all my pool



# ceph osd dump | grep  'replicated size'

pool 1 'device_health_metrics' replicated size 3 min_size 2 crush_rule 0
object_hash rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 189372
flags hashpspool stripe_width 0 pg_num_min 1 application mgr_devicehealth

pool 10 'cephfs_data' replicated size 2 min_size 1 crush_rule 1 object_hash
rjenkins pg_num 512 pgp_num 512 autoscale_mode on last_change 189346 lfor
0/0/183690 flags hashpspool,selfmanaged_snaps stripe_width 0 application
cephfs

pool 11 'cephfs_metadata' replicated size 2 min_size 1 crush_rule 1
object_hash rjenkins pg_num 16 pgp_num 16 autoscale_mode on last_change
187861 lfor 0/187861/187859 flags hashpspool stripe_width 0
pg_autoscale_bias 4 pg_num_min 16 recovery_priority 5 application cephfs

pool 18 '.rgw.root' replicated size 3 min_size 2 crush_rule 0 object_hash
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 5265 flags
hashpspool stripe_width 0 application rgw

pool 19 'default.rgw.log' replicated size 3 min_size 2 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 5267
flags hashpspool stripe_width 0 application rgw

pool 20 'default.rgw.control' replicated size 3 min_size 2 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 5269
flags hashpspool stripe_width 0 application rgw

pool 21 'default.rgw.meta' replicated size 3 min_size 2 crush_rule 0
object_hash rjenkins pg_num 8 pgp_num 8 autoscale_mode on last_change 5398
lfor 0/5398/5396 flags hashpspool stripe_width 0 pg_autoscale_bias 4
pg_num_min 8 application rgw

pool 22 'default.rgw.buckets.index' replicated size 3 min_size 2 crush_rule
0 object_hash rjenkins pg_num 8 pgp_num 8 autoscale_mode on last_change 7491
lfor 0/7491/7489 flags hashpspool stripe_width 0 pg_autoscale_bias 4
pg_num_min 8 application rgw

pool 23 'default.rgw.buckets.data' replicated size 3 min_size 2 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 7500
flags hashpspool stripe_width 0 application rgw

pool 24 'kubernetes' replicated size 2 min_size 1 crush_rule 1 object_hash
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 189363 lfor
0/0/7560 flags hashpspool,selfmanaged_snaps stripe_width 0 application rbd

pool 25 'default.rgw.buckets.non-ec' replicated size 3 min_size 2 crush_rule
0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
23983