Oh ok. Because we have two types of ssds. 1.8TB and 3.6TB.
The 1.8TB got around 90-100pgs and the 3.6TB around 150-190pgs
Here is the output:
RAW STORAGE:
CLASS SIZE AVAIL USED RAW USED %RAW USED
ssd 956 TiB 360 TiB 595 TiB 596 TiB 62.35
TOTAL 956 TiB 360 TiB 595 TiB 596 TiB 62.35
POOLS:
POOL ID STORED OBJECTS USED %USED MAX AVAIL
pool 1 1 580 GiB 149.31k 1.7 TiB 0.58 96 TiB
pool 3 3 208 TiB 61.66M 593 TiB 67.21 96 TiB
pool 4 4 7.9 MiB 2.01k 68 MiB 0 96 TiB
pool 5 5 19 B 2 36 KiB 0 96 TiB
Am Di., 16. Juni 2020 um 14:13 Uhr schrieb Dan van der Ster
<[email protected]>:
>
> On Tue, Jun 16, 2020 at 2:00 PM Boris Behrens <[email protected]> wrote:
> >
> > See inline comments
> > Dan van der Ster <[email protected]> 于2020年6月16日周二 下午7:07写道:
> > >
> > > Could you share the output of
> > >
> > > ceph osd pool ls detail
> >
> > pool 1 'pool 1' replicated size 3 min_size 2 crush_rule 0 object_hash
> > rjenkins pg_num 256 pgp_num 256 autoscale_mode warn last_change
> > 2318859 flags hashpspool min_write_recency_for_promote 1 stripe_width
> > 0 application rbd
> > pool 3 'pool 3' replicated size 3 min_size 2 crush_rule 0 object_hash
> > rjenkins pg_num 16384 pgp_num 16384 autoscale_mode warn last_change
> > 2544040 lfor 0/0/1952329 flags hashpspool,selfmanaged_snaps
> > min_write_recency_for_promote 1 stripe_width 0 application rbd
> > pool 4 'pool 4' replicated size 3 min_size 2 crush_rule 0 object_hash
> > rjenkins pg_num 8 pgp_num 8 autoscale_mode warn last_change 2318859
> > flags hashpspool min_write_recency_for_promote 1 stripe_width 0
> > application rbd
> > pool 5 'pool 5' replicated size 3 min_size 2 crush_rule 0 object_hash
> > rjenkins pg_num 256 pgp_num 256 autoscale_mode warn last_change
> > 2318859 flags hashpspool,selfmanaged_snaps
> > min_write_recency_for_promote 1 stripe_width 0 application rbd
> >
>
> OK now maybe share the output of `ceph df` so we can see how much data
> is in each pool?
>
> Assuming that the majority of your data is in 'pool 3' with 16384 PGs,
> then your current PG values are just fine. (You should have around 110
> PGs per OSD).
> The pg_autoscaler aims for 100 per OSD and doesn't make changes unless
> a pool has 4x too few or too many PGs.
>
> Unless you are planning to put a large proportion of data into the
> other pools, I'd leave pg_autoscaler disabled and move on to the next
> task.
>
> -- Dan
>
> > the mgr module is not enabled yet.
> >
> > >
> > > ?
> > >
> > > This way we can see how the pools are configured and help recommend if
> > > pg_autoscaler is worth enabling.
> > >
> >
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]