I tried to do $ ceph tell osd.* config set osd_op_queue wpq but got Error EPERM: error setting 'osd_op_queue' to 'wpq': (1) Operation not permitted - is there some other way how to enable wpq instead of mclock?
I'd recommend in general for ~98% config changes to go through the `ceph config set _` because these ones get stored in the mon. ` ceph tell osd.*` config changes don't persist after a daemon reboots. daemons read their config from ceph.config and the config dump. ceph config set osd osd_op_queue wpq just a heads up that you need to restart all your osds for this command to take into effect. On another note, maybe check as well if you have writeback caches disabled since you are on hdds. write through mode is generally recommended. https://docs.ceph.com/en/latest/start/hardware-recommendations/#write-caches On Tue, Oct 14, 2025 at 7:16 AM Jan Kasprzak <[email protected]> wrote: > [replying to multiple messages at once] > > Anthony D'Atri wrote: > > `ceph osd df` > > `ceph osd dump | grep pool` > > Attached below. > > > >> Note that you have just 36 OSDs, and each of your EC PGs needs to > reserve 6 of them, so you have a certain gridlock factor. > > Yes, but I guess I should see at least some %iowait or something, > when HDDs are the bottleneck. Currently I see this: > > $ for i in $hosts; do ssh root@$i 'top -n 1 -b | grep ^%Cpu'; done > %Cpu(s): 0.3 us, 0.3 sy, 0.0 ni, 99.1 id, 0.3 wa, 0.0 hi, 0.1 si, > 0.0 st > %Cpu(s): 0.3 us, 0.5 sy, 0.0 ni, 98.4 id, 0.6 wa, 0.1 hi, 0.2 si, > 0.0 st > %Cpu(s): 0.2 us, 0.4 sy, 0.0 ni, 98.3 id, 0.9 wa, 0.1 hi, 0.2 si, > 0.0 st > %Cpu(s): 0.1 us, 0.3 sy, 0.0 ni, 99.6 id, 0.1 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu(s): 0.1 us, 0.3 sy, 0.0 ni, 99.1 id, 0.5 wa, 0.0 hi, 0.1 si, > 0.0 st > %Cpu(s): 0.1 us, 0.2 sy, 0.0 ni, 99.5 id, 0.2 wa, 0.0 hi, 0.1 si, > 0.0 st > %Cpu(s): 0.1 us, 0.3 sy, 0.0 ni, 99.6 id, 0.1 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu(s): 0.3 us, 0.3 sy, 0.0 ni, 98.5 id, 0.7 wa, 0.1 hi, 0.2 si, > 0.0 st > %Cpu(s): 0.1 us, 0.3 sy, 0.0 ni, 99.0 id, 0.4 wa, 0.1 hi, 0.1 si, > 0.0 st > (the hosts have 96 CPUs as seen in /proc/cpuinfo, so a single busy CPU > is slightly more than 1 % of total CPU time). > > > >> Plus the usual wpq vs mclock. > > > > > > Could you elaborate? > > > > Google ceph wpq mclock. In the last couple of releases the mclock > > scheduler is default, but on hdds especially and especially with EC it > > can cause backfill to be super slow. There are some improvements in the > > pipe, but a lot of people have just reverted to the wpq scheduler for > now. > > You can find instructions for that. > > I tried to do > > $ ceph tell osd.* config set osd_op_queue wpq > > but got > > Error EPERM: error setting 'osd_op_queue' to 'wpq': (1) Operation not > permitted > > - is there some other way how to enable wpq instead of mclock? > > Kirby Haze wrote: > > I will assume you are running the default mclock scheduler and not wpq. > I'm > > not too familiar with tuning mclock settings but this is the docs to look > > at > > > https://docs.ceph.com/en/latest/rados/configuration/mclock-config-ref/#recovery-backfill-options > > > > osd_max_backfills is set to 1 by default and this is the first thing I > > would tune if you want faster backfilling. > > Yes, but I tried to increase it to 30, but still got only one pg in > the backfilling state and the rest in backfill_wait. > > Stephan Hohn wrote: > > Using mclock scheduler you can do the following: > > > > set osd_mclock_override_recovery_settings to true e.g. > > And this is what I missed. By itself, setting osd_max_backfills > does nothing, unless the above option is set to true. With this, > I got all backfill_wait'ing PGs to backfilling state. This got me to > ~150-200 MB/s of recovery I/O: > > ceph tell osd.* config set osd_mclock_override_recovery_settings true > ceph tell osd.* config set osd_max_backfills 30 > > Setting the high_recovery_ops profile increased the recovery speed > further to 300-400 MB/s: > > ceph tell osd.* config set osd_mclock_profile high_recovery_ops > > But still my cluster is mostly idle (see above the %Cpu stats), > and the recovery speed can probably reach 800-900 MB/s, as I have > 10GbaseT network. Maybe the balacer can schedule more than 11 PGs > for resharding? > > Wannes Smet wrote: > > Perhaps not the root cause, but might be worth looking at if you haven't > already: > > > > * > > How is the host power profile configured? Eg, I'm running HPe: > > System Configuration > BIOS/Platform Configuration (RBSU) > Power > Management > Power Profile > Maximum Performance > > * > > In which C-state are your cores running? Use linux-cpupower or similar > tool to verify. From my experience if I don't configure anything 99% of the > time it's in C6 and I want it to be in C0/C1. > > According to "cpupower idle-info" my CPUs support POLL C1 C2 idle states, > and spend most of the time in C2. I tried to do > > # cpupower idle-set --disable 2 > > on two of my hosts (thus disabling the C2 state), but ping over a minute > between them did not change in a significant way. > > > What about CPU wait states? Do you see any? To > > visualize and correlate with HDDs, I personally like nmon > > (http://kb.ictbanking.net/article.php?id=550&oid=1) then press lower > > case 'L' to get a long term graph of CPU usage. From my experience, if > > you see blue blocks ('W' if color isn't enabled), that's wait states > > and ideally you only want to see none at all. A very occasional blue > > (W) block might be ~acceptable but if it's more than that, there's > > very likely hardware (HDDs would be my main suspect) noticably dragging > > down performance. > > > > Pressing 'c' in nmon will toggle an overview per core. That'll give > > a bit more "visual" insight into how much time cores are spending in > > user/system/wait . > > In nmon, you can also press 'd' to toggle disk stats ('h' to show help). > > > > To correlate with disk activity: press 'd' to toggle a graph showing > > R/W activity on each disk. > > Interesting. I'll try it when I get time. > > So, thanks to all the replies! > > -Yenya > > $ ceph osd df > ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META > AVAIL %USE VAR PGS STATUS > 4 hdd 20.10739 1.00000 20 TiB 7.0 TiB 6.9 TiB 1.1 GiB 10 > GiB 13 TiB 34.94 0.84 182 up > 9 hdd 20.10739 1.00000 20 TiB 11 TiB 10 TiB 800 MiB 15 > GiB 9.5 TiB 52.66 1.26 209 up > 20 hdd 20.10739 1.00000 20 TiB 11 TiB 11 TiB 519 MiB 16 > GiB 8.8 TiB 56.09 1.34 201 up > 28 hdd 20.10739 1.00000 20 TiB 7.0 TiB 6.9 TiB 934 MiB 10 > GiB 13 TiB 34.76 0.83 180 up > 2 hdd 20.10739 1.00000 20 TiB 8.5 TiB 8.4 TiB 1.1 GiB 13 > GiB 12 TiB 42.05 1.01 189 up > 10 hdd 20.10739 1.00000 20 TiB 6.5 TiB 6.4 TiB 704 MiB 9.5 > GiB 14 TiB 32.43 0.78 169 up > 18 hdd 20.10739 1.00000 20 TiB 7.8 TiB 7.7 TiB 1.0 GiB 11 > GiB 12 TiB 38.68 0.92 198 up > 29 hdd 20.10739 1.00000 20 TiB 11 TiB 11 TiB 1.2 GiB 16 > GiB 8.8 TiB 56.17 1.34 220 up > 1 hdd 20.10739 1.00000 20 TiB 8.8 TiB 8.7 TiB 1.2 GiB 13 > GiB 11 TiB 43.73 1.05 209 up > 11 hdd 20.10739 1.00000 20 TiB 8.0 TiB 7.9 TiB 861 MiB 11 > GiB 12 TiB 39.89 0.95 193 up > 19 hdd 20.10739 1.00000 20 TiB 7.0 TiB 6.9 TiB 607 MiB 10 > GiB 13 TiB 34.92 0.83 190 up > 27 hdd 20.10739 1.00000 20 TiB 8.5 TiB 8.4 TiB 1.2 GiB 13 > GiB 12 TiB 42.21 1.01 187 up > 3 hdd 20.10739 1.00000 20 TiB 8.3 TiB 8.2 TiB 1.1 GiB 12 > GiB 12 TiB 41.41 0.99 190 up > 13 hdd 20.10739 1.00000 20 TiB 11 TiB 11 TiB 774 MiB 15 > GiB 9.4 TiB 53.03 1.27 208 up > 22 hdd 20.10739 1.00000 20 TiB 5.1 TiB 5.0 TiB 1.3 GiB 7.6 > GiB 15 TiB 25.47 0.61 184 up > 31 hdd 20.10739 1.00000 20 TiB 6.3 TiB 6.2 TiB 663 MiB 9.1 > GiB 14 TiB 31.17 0.74 175 up > 5 hdd 20.10739 1.00000 20 TiB 5.8 TiB 5.7 TiB 840 MiB 8.4 > GiB 14 TiB 29.00 0.69 184 up > 12 hdd 20.10739 1.00000 20 TiB 7.5 TiB 7.4 TiB 907 MiB 11 > GiB 13 TiB 37.53 0.90 177 up > 21 hdd 20.10739 1.00000 20 TiB 9.0 TiB 8.9 TiB 976 MiB 13 > GiB 11 TiB 44.62 1.07 200 up > 30 hdd 20.10739 1.00000 20 TiB 8.4 TiB 8.3 TiB 1.0 GiB 12 > GiB 12 TiB 41.76 1.00 193 up > 6 hdd 20.10739 1.00000 20 TiB 8.0 TiB 7.9 TiB 1.1 GiB 12 > GiB 12 TiB 39.87 0.95 195 up > 14 hdd 20.10739 1.00000 20 TiB 9.0 TiB 8.9 TiB 1.1 GiB 13 > GiB 11 TiB 44.75 1.07 209 up > 23 hdd 20.10739 1.00000 20 TiB 9.0 TiB 8.9 TiB 236 MiB 13 > GiB 11 TiB 44.81 1.07 188 up > 33 hdd 20.10739 1.00000 20 TiB 7.2 TiB 7.1 TiB 355 MiB 11 > GiB 13 TiB 35.94 0.86 161 up > 8 hdd 20.10739 1.00000 20 TiB 9.7 TiB 9.6 TiB 1.0 GiB 14 > GiB 10 TiB 48.29 1.15 191 up > 16 hdd 20.10739 1.00000 20 TiB 6.9 TiB 6.8 TiB 533 MiB 10 > GiB 13 TiB 34.18 0.82 192 up > 25 hdd 20.10739 1.00000 20 TiB 8.3 TiB 8.2 TiB 1.0 GiB 12 > GiB 12 TiB 41.45 0.99 187 up > 32 hdd 20.10739 1.00000 20 TiB 8.6 TiB 8.5 TiB 1.6 GiB 12 > GiB 12 TiB 42.55 1.02 200 up > 7 hdd 20.10739 1.00000 20 TiB 10 TiB 10 TiB 832 MiB 15 > GiB 9.9 TiB 50.52 1.21 192 up > 17 hdd 20.10739 1.00000 20 TiB 9.8 TiB 9.7 TiB 832 MiB 14 > GiB 10 TiB 48.85 1.17 205 up > 24 hdd 20.10739 1.00000 20 TiB 8.5 TiB 8.4 TiB 918 MiB 13 > GiB 12 TiB 42.44 1.01 189 up > 34 hdd 20.10739 1.00000 20 TiB 7.9 TiB 7.8 TiB 720 MiB 12 > GiB 12 TiB 39.52 0.94 187 up > 0 hdd 20.10739 1.00000 20 TiB 9.6 TiB 9.5 TiB 793 MiB 14 > GiB 10 TiB 47.91 1.15 207 up > 15 hdd 20.10739 1.00000 20 TiB 9.3 TiB 9.2 TiB 325 MiB 13 > GiB 11 TiB 46.21 1.10 198 up > 26 hdd 20.10739 1.00000 20 TiB 8.6 TiB 8.5 TiB 702 MiB 12 > GiB 12 TiB 42.59 1.02 197 up > 35 hdd 20.10739 1.00000 20 TiB 8.8 TiB 8.7 TiB 1.6 GiB 13 > GiB 11 TiB 43.83 1.05 179 up > TOTAL 724 TiB 303 TiB 299 TiB 32 GiB 442 > GiB 421 TiB 41.84 > MIN/MAX VAR: 0.61/1.34 STDDEV: 7.12 > > > $ ceph osd dump | grep pool > pool 1 '.mgr' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 46 flags > hashpspool stripe_width 0 pg_num_max 32 pg_num_min 1 application mgr > read_balance_score 37.50 > pool 2 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 68 lfor 0/0/66 > flags hashpspool stripe_width 0 application mystorage read_balance_score > 4.52 > pool 4 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 1 > object_hash rjenkins pg_num 128 pgp_num 79 pgp_num_target 128 > autoscale_mode on last_change 1370 lfor 0/0/423 flags > hashpspool,ec_overwrites stripe_width 16384 application mystorage > pool 5 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 157 lfor > 0/0/155 flags hashpspool stripe_width 0 application mystorage > read_balance_score 3.39 > pool 6 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 2 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 157 > lfor 0/0/155 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 9 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor > 0/0/249 flags hashpspool stripe_width 0 application mystorage > read_balance_score 3.36 > pool 10 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 4 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 > lfor 0/0/251 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 11 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor > 0/0/251 flags hashpspool stripe_width 0 application mystorage > read_balance_score 4.52 > pool 12 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 5 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 > lfor 0/0/253 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 13 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor > 0/0/253 flags hashpspool stripe_width 0 application mystorage > read_balance_score 3.36 > pool 14 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 6 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 > lfor 0/0/255 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 15 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor > 0/0/255 flags hashpspool stripe_width 0 application mystorage > read_balance_score 2.27 > pool 16 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 7 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 > lfor 0/0/257 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 17 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor > 0/0/257 flags hashpspool stripe_width 0 application mystorage > read_balance_score 3.39 > pool 18 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 8 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 > lfor 0/0/259 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 19 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor > 0/0/259 flags hashpspool stripe_width 0 application mystorage > read_balance_score 4.52 > pool 20 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 9 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 > lfor 0/0/261 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 21 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor > 0/0/261 flags hashpspool stripe_width 0 application mystorage > read_balance_score 4.50 > pool 22 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 10 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 > lfor 0/0/263 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 23 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor > 0/0/263 flags hashpspool stripe_width 0 application mystorage > read_balance_score 4.51 > pool 24 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 11 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 > lfor 0/0/265 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 25 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor > 0/0/322 flags hashpspool stripe_width 0 application mystorage > read_balance_score 3.39 > pool 26 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 12 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 > lfor 0/0/324 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 27 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor > 0/0/324 flags hashpspool stripe_width 0 application mystorage > read_balance_score 4.49 > pool 28 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 13 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 > lfor 0/0/326 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 29 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor > 0/0/326 flags hashpspool stripe_width 0 application mystorage > read_balance_score 3.37 > pool 30 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 14 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 > lfor 0/0/328 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 31 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor > 0/0/328 flags hashpspool stripe_width 0 application mystorage > read_balance_score 3.38 > pool 32 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 15 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 > lfor 0/0/330 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 33 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor > 0/0/330 flags hashpspool stripe_width 0 application mystorage > read_balance_score 4.52 > pool 34 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 16 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 > lfor 0/0/332 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 35 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor > 0/0/332 flags hashpspool stripe_width 0 application mystorage > read_balance_score 3.38 > pool 36 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 17 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 > lfor 0/0/334 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 37 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor > 0/0/334 flags hashpspool stripe_width 0 application mystorage > read_balance_score 3.38 > pool 38 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 18 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 > lfor 0/0/336 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 39 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor > 0/0/336 flags hashpspool stripe_width 0 application mystorage > read_balance_score 2.27 > pool 40 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 19 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 > lfor 0/0/338 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 41 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor > 0/0/338 flags hashpspool stripe_width 0 application mystorage > read_balance_score 4.52 > pool 42 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 20 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 > lfor 0/0/340 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 43 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor > 0/0/340 flags hashpspool stripe_width 0 application mystorage > read_balance_score 3.36 > pool 44 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 21 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 > lfor 0/0/342 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 45 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor > 0/0/342 flags hashpspool stripe_width 0 application mystorage > read_balance_score 3.37 > pool 46 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 22 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 > lfor 0/0/344 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > pool 47 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor > 0/0/344 flags hashpspool stripe_width 0 application mystorage > read_balance_score 3.37 > pool 48 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 23 > object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 > lfor 0/0/346 flags hashpspool,ec_overwrites stripe_width 16384 application > mystorage > > -- > | Jan "Yenya" Kasprzak <kas at {fi.muni.cz - work | yenya.net - private}> > | > | https://www.fi.muni.cz/~kas/ GPG: 4096R/A45477D5 > | > We all agree on the necessity of compromise. We just can't agree on > when it's necessary to compromise. --Larry Wall > _______________________________________________ > ceph-users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > _______________________________________________ ceph-users mailing list -- [email protected] To unsubscribe send an email to [email protected]
