On 5/26/22 14:09, Wesley Dillingham wrote:
What does "ceph osd pool ls detail" say?

$ ceph osd pool ls detail
pool 0 'rbd' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 64 pgp_num 64 autoscale_mode on last_change 44740 flags hashpspool,selfmanaged_snaps stripe_width 0 application rbd pool 1 '.rgw.root' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 44740 lfor 0/0/31483 owner 18446744073709551615 flags hashpspool stripe_width 0 application rgw pool 2 'default.rgw.control' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 44740 lfor 0/0/31469 owner 18446744073709551615 flags hashpspool stripe_width 0 application rgw pool 3 'default.rgw.data.root' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 44740 lfor 0/0/31471 owner 18446744073709551615 flags hashpspool stripe_width 0 application rgw pool 4 'default.rgw.gc' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 44740 lfor 0/0/31471 owner 18446744073709551615 flags hashpspool stripe_width 0 application rgw pool 5 'default.rgw.log' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 44740 lfor 0/0/31387 owner 18446744073709551615 flags hashpspool stripe_width 0 application rgw pool 6 'default.rgw.users.uid' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 44740 lfor 0/0/31387 flags hashpspool stripe_width 0 application rgw pool 12 'mathfs_data' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 44740 lfor 0/31370/31368 flags hashpspool stripe_width 0 application cephfs pool 13 'mathfs_metadata' replicated size 2 min_size 2 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 44740 lfor 0/27164/27162 flags hashpspool stripe_width 0 application cephfs pool 15 'default.rgw.lc' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 44740 lfor 0/0/31374 flags hashpspool stripe_width 0 application rgw pool 21 'libvirt' replicated size 3 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 56244 lfor 0/33144/33142 flags hashpspool,selfmanaged_snaps stripe_width 0 application rbd pool 36 'monthly_archive_metadata' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 45338 lfor 0/27845/27843 flags hashpspool stripe_width 0 application cephfs pool 37 'monthly_archive_data' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 45334 lfor 0/44535/44533 flags hashpspool stripe_width 0 application cephfs pool 38 'device_health_metrics' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 56507 flags hashpspool stripe_width 0 pg_num_min 1 application mgr_devicehealth pool 41 'lensfun_metadata' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 54066 flags hashpspool stripe_width 0 pg_autoscale_bias 4 pg_num_min 16 recovery_priority 5 application cephfs pool 42 'lensfun_data' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 54066 flags hashpspool stripe_width 0 application cephfs


_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to