[replying to multiple messages at once]

Anthony D'Atri wrote:
> `ceph osd df`
> `ceph osd dump | grep pool`

Attached below.

> >> Note that you have just 36 OSDs, and each of your EC PGs needs to reserve 
> >> 6 of them, so you have a certain gridlock factor.

Yes, but I guess I should see at least some %iowait or something,
when HDDs are the bottleneck. Currently I see this:

$ for i in $hosts; do ssh root@$i 'top -n 1 -b | grep ^%Cpu'; done
%Cpu(s):  0.3 us,  0.3 sy,  0.0 ni, 99.1 id,  0.3 wa,  0.0 hi,  0.1 si,  0.0 st 
%Cpu(s):  0.3 us,  0.5 sy,  0.0 ni, 98.4 id,  0.6 wa,  0.1 hi,  0.2 si,  0.0 st 
%Cpu(s):  0.2 us,  0.4 sy,  0.0 ni, 98.3 id,  0.9 wa,  0.1 hi,  0.2 si,  0.0 st 
%Cpu(s):  0.1 us,  0.3 sy,  0.0 ni, 99.6 id,  0.1 wa,  0.0 hi,  0.0 si,  0.0 st 
%Cpu(s):  0.1 us,  0.3 sy,  0.0 ni, 99.1 id,  0.5 wa,  0.0 hi,  0.1 si,  0.0 st 
%Cpu(s):  0.1 us,  0.2 sy,  0.0 ni, 99.5 id,  0.2 wa,  0.0 hi,  0.1 si,  0.0 st 
%Cpu(s):  0.1 us,  0.3 sy,  0.0 ni, 99.6 id,  0.1 wa,  0.0 hi,  0.0 si,  0.0 st 
%Cpu(s):  0.3 us,  0.3 sy,  0.0 ni, 98.5 id,  0.7 wa,  0.1 hi,  0.2 si,  0.0 st 
%Cpu(s):  0.1 us,  0.3 sy,  0.0 ni, 99.0 id,  0.4 wa,  0.1 hi,  0.1 si,  0.0 st 
(the hosts have 96 CPUs as seen in /proc/cpuinfo, so a single busy CPU
is slightly more than 1 % of total CPU time).

> >> Plus the usual wpq vs mclock.
> > 
> > Could you elaborate?
> 
> Google ceph wpq mclock.  In the last couple of releases the mclock
> scheduler is default, but on hdds especially and especially with EC it
> can cause backfill to be super slow.  There are some improvements in the
> pipe, but a lot of people have just reverted to the wpq scheduler for now.
> You can find instructions for that.

I tried to do

$ ceph tell osd.* config set osd_op_queue wpq

but got

Error EPERM: error setting 'osd_op_queue' to 'wpq': (1) Operation not permitted

- is there some other way how to enable wpq instead of mclock?

Kirby Haze wrote:
> I will assume you are running the default mclock scheduler and not wpq. I'm
> not too familiar with tuning mclock settings but this is the docs to look
> at
> https://docs.ceph.com/en/latest/rados/configuration/mclock-config-ref/#recovery-backfill-options
> 
> osd_max_backfills is set to 1 by default and this is the first thing I
> would tune if you want faster backfilling.

Yes, but I tried to increase it to 30, but still got only one pg in
the backfilling state and the rest in backfill_wait.

Stephan Hohn wrote:
> Using mclock scheduler you can do the following:
> 
> set osd_mclock_override_recovery_settings to true e.g.

And this is what I missed. By itself, setting osd_max_backfills
does nothing, unless the above option is set to true. With this,
I got all backfill_wait'ing PGs to backfilling state. This got me to
~150-200 MB/s of recovery I/O:

ceph tell osd.* config set osd_mclock_override_recovery_settings true
ceph tell osd.* config set osd_max_backfills 30

Setting the high_recovery_ops profile increased the recovery speed
further to 300-400 MB/s:

ceph tell osd.* config set osd_mclock_profile high_recovery_ops

But still my cluster is mostly idle (see above the %Cpu stats),
and the recovery speed can probably reach 800-900 MB/s, as I have
10GbaseT network. Maybe the balacer can schedule more than 11 PGs
for resharding?

Wannes Smet wrote:
> Perhaps not the root cause, but might be worth looking at if you haven't 
> already:
> 
>   *
> How is the host power profile configured? Eg, I'm running HPe:
>     System Configuration > BIOS/Platform Configuration (RBSU) > Power 
> Management > Power Profile > Maximum Performance
>   *
> In which C-state are your cores running?  Use linux-cpupower or similar tool 
> to verify. From my experience if I don't configure anything 99% of the time 
> it's in C6 and I want it to be in C0/C1.

According to "cpupower idle-info" my CPUs support POLL C1 C2 idle states,
and spend most of the time in C2. I tried to do

# cpupower idle-set --disable 2

on two of my hosts (thus disabling the C2 state), but ping over a minute
between them did not change in a significant way.

> What about CPU wait states? Do you see any? To
> visualize and correlate with HDDs, I personally like nmon
> (http://kb.ictbanking.net/article.php?id=550&oid=1) then press lower
> case 'L' to get a long term graph of CPU usage. From my experience, if
> you see blue blocks ('W' if color isn't enabled), that's wait states
> and ideally you only want to see none at all. A very occasional blue
> (W) block might be ~acceptable but if it's more than that, there's
> very likely hardware (HDDs would be my main suspect) noticably dragging
> down performance.
>
> Pressing 'c' in nmon will toggle an overview per core. That'll give
> a bit more "visual" insight into how much time cores are spending in
> user/system/wait .
> In nmon, you can also press 'd' to toggle disk stats ('h' to show help).
>
> To correlate with disk activity: press 'd' to toggle a graph showing
> R/W activity on each disk.

Interesting. I'll try it when I get time.

So, thanks to all the replies!

-Yenya

$ ceph osd df
ID  CLASS  WEIGHT    REWEIGHT  SIZE     RAW USE  DATA     OMAP     META     
AVAIL    %USE   VAR   PGS  STATUS
 4    hdd  20.10739   1.00000   20 TiB  7.0 TiB  6.9 TiB  1.1 GiB   10 GiB   13 
TiB  34.94  0.84  182      up
 9    hdd  20.10739   1.00000   20 TiB   11 TiB   10 TiB  800 MiB   15 GiB  9.5 
TiB  52.66  1.26  209      up
20    hdd  20.10739   1.00000   20 TiB   11 TiB   11 TiB  519 MiB   16 GiB  8.8 
TiB  56.09  1.34  201      up
28    hdd  20.10739   1.00000   20 TiB  7.0 TiB  6.9 TiB  934 MiB   10 GiB   13 
TiB  34.76  0.83  180      up
 2    hdd  20.10739   1.00000   20 TiB  8.5 TiB  8.4 TiB  1.1 GiB   13 GiB   12 
TiB  42.05  1.01  189      up
10    hdd  20.10739   1.00000   20 TiB  6.5 TiB  6.4 TiB  704 MiB  9.5 GiB   14 
TiB  32.43  0.78  169      up
18    hdd  20.10739   1.00000   20 TiB  7.8 TiB  7.7 TiB  1.0 GiB   11 GiB   12 
TiB  38.68  0.92  198      up
29    hdd  20.10739   1.00000   20 TiB   11 TiB   11 TiB  1.2 GiB   16 GiB  8.8 
TiB  56.17  1.34  220      up
 1    hdd  20.10739   1.00000   20 TiB  8.8 TiB  8.7 TiB  1.2 GiB   13 GiB   11 
TiB  43.73  1.05  209      up
11    hdd  20.10739   1.00000   20 TiB  8.0 TiB  7.9 TiB  861 MiB   11 GiB   12 
TiB  39.89  0.95  193      up
19    hdd  20.10739   1.00000   20 TiB  7.0 TiB  6.9 TiB  607 MiB   10 GiB   13 
TiB  34.92  0.83  190      up
27    hdd  20.10739   1.00000   20 TiB  8.5 TiB  8.4 TiB  1.2 GiB   13 GiB   12 
TiB  42.21  1.01  187      up
 3    hdd  20.10739   1.00000   20 TiB  8.3 TiB  8.2 TiB  1.1 GiB   12 GiB   12 
TiB  41.41  0.99  190      up
13    hdd  20.10739   1.00000   20 TiB   11 TiB   11 TiB  774 MiB   15 GiB  9.4 
TiB  53.03  1.27  208      up
22    hdd  20.10739   1.00000   20 TiB  5.1 TiB  5.0 TiB  1.3 GiB  7.6 GiB   15 
TiB  25.47  0.61  184      up
31    hdd  20.10739   1.00000   20 TiB  6.3 TiB  6.2 TiB  663 MiB  9.1 GiB   14 
TiB  31.17  0.74  175      up
 5    hdd  20.10739   1.00000   20 TiB  5.8 TiB  5.7 TiB  840 MiB  8.4 GiB   14 
TiB  29.00  0.69  184      up
12    hdd  20.10739   1.00000   20 TiB  7.5 TiB  7.4 TiB  907 MiB   11 GiB   13 
TiB  37.53  0.90  177      up
21    hdd  20.10739   1.00000   20 TiB  9.0 TiB  8.9 TiB  976 MiB   13 GiB   11 
TiB  44.62  1.07  200      up
30    hdd  20.10739   1.00000   20 TiB  8.4 TiB  8.3 TiB  1.0 GiB   12 GiB   12 
TiB  41.76  1.00  193      up
 6    hdd  20.10739   1.00000   20 TiB  8.0 TiB  7.9 TiB  1.1 GiB   12 GiB   12 
TiB  39.87  0.95  195      up
14    hdd  20.10739   1.00000   20 TiB  9.0 TiB  8.9 TiB  1.1 GiB   13 GiB   11 
TiB  44.75  1.07  209      up
23    hdd  20.10739   1.00000   20 TiB  9.0 TiB  8.9 TiB  236 MiB   13 GiB   11 
TiB  44.81  1.07  188      up
33    hdd  20.10739   1.00000   20 TiB  7.2 TiB  7.1 TiB  355 MiB   11 GiB   13 
TiB  35.94  0.86  161      up
 8    hdd  20.10739   1.00000   20 TiB  9.7 TiB  9.6 TiB  1.0 GiB   14 GiB   10 
TiB  48.29  1.15  191      up
16    hdd  20.10739   1.00000   20 TiB  6.9 TiB  6.8 TiB  533 MiB   10 GiB   13 
TiB  34.18  0.82  192      up
25    hdd  20.10739   1.00000   20 TiB  8.3 TiB  8.2 TiB  1.0 GiB   12 GiB   12 
TiB  41.45  0.99  187      up
32    hdd  20.10739   1.00000   20 TiB  8.6 TiB  8.5 TiB  1.6 GiB   12 GiB   12 
TiB  42.55  1.02  200      up
 7    hdd  20.10739   1.00000   20 TiB   10 TiB   10 TiB  832 MiB   15 GiB  9.9 
TiB  50.52  1.21  192      up
17    hdd  20.10739   1.00000   20 TiB  9.8 TiB  9.7 TiB  832 MiB   14 GiB   10 
TiB  48.85  1.17  205      up
24    hdd  20.10739   1.00000   20 TiB  8.5 TiB  8.4 TiB  918 MiB   13 GiB   12 
TiB  42.44  1.01  189      up
34    hdd  20.10739   1.00000   20 TiB  7.9 TiB  7.8 TiB  720 MiB   12 GiB   12 
TiB  39.52  0.94  187      up
 0    hdd  20.10739   1.00000   20 TiB  9.6 TiB  9.5 TiB  793 MiB   14 GiB   10 
TiB  47.91  1.15  207      up
15    hdd  20.10739   1.00000   20 TiB  9.3 TiB  9.2 TiB  325 MiB   13 GiB   11 
TiB  46.21  1.10  198      up
26    hdd  20.10739   1.00000   20 TiB  8.6 TiB  8.5 TiB  702 MiB   12 GiB   12 
TiB  42.59  1.02  197      up
35    hdd  20.10739   1.00000   20 TiB  8.8 TiB  8.7 TiB  1.6 GiB   13 GiB   11 
TiB  43.83  1.05  179      up
                        TOTAL  724 TiB  303 TiB  299 TiB   32 GiB  442 GiB  421 
TiB  41.84                   
MIN/MAX VAR: 0.61/1.34  STDDEV: 7.12


$ ceph osd dump | grep pool
pool 1 '.mgr' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins 
pg_num 1 pgp_num 1 autoscale_mode on last_change 46 flags hashpspool 
stripe_width 0 pg_num_max 32 pg_num_min 1 application mgr read_balance_score 
37.50
pool 2 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 68 lfor 0/0/66 
flags hashpspool stripe_width 0 application mystorage read_balance_score 4.52
pool 4 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 1 
object_hash rjenkins pg_num 128 pgp_num 79 pgp_num_target 128 autoscale_mode on 
last_change 1370 lfor 0/0/423 flags hashpspool,ec_overwrites stripe_width 16384 
application mystorage
pool 5 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 157 lfor 0/0/155 
flags hashpspool stripe_width 0 application mystorage read_balance_score 3.39
pool 6 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 2 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 157 
lfor 0/0/155 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 9 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor 0/0/249 
flags hashpspool stripe_width 0 application mystorage read_balance_score 3.36
pool 10 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 4 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 
lfor 0/0/251 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 11 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor 0/0/251 
flags hashpspool stripe_width 0 application mystorage read_balance_score 4.52
pool 12 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 5 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 
lfor 0/0/253 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 13 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor 0/0/253 
flags hashpspool stripe_width 0 application mystorage read_balance_score 3.36
pool 14 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 6 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 
lfor 0/0/255 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 15 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor 0/0/255 
flags hashpspool stripe_width 0 application mystorage read_balance_score 2.27
pool 16 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 7 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 
lfor 0/0/257 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 17 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor 0/0/257 
flags hashpspool stripe_width 0 application mystorage read_balance_score 3.39
pool 18 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 8 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 
lfor 0/0/259 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 19 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor 0/0/259 
flags hashpspool stripe_width 0 application mystorage read_balance_score 4.52
pool 20 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 9 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 
lfor 0/0/261 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 21 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor 0/0/261 
flags hashpspool stripe_width 0 application mystorage read_balance_score 4.50
pool 22 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 10 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 
lfor 0/0/263 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 23 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 lfor 0/0/263 
flags hashpspool stripe_width 0 application mystorage read_balance_score 4.51
pool 24 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 11 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 309 
lfor 0/0/265 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 25 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor 0/0/322 
flags hashpspool stripe_width 0 application mystorage read_balance_score 3.39
pool 26 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 12 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 
lfor 0/0/324 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 27 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor 0/0/324 
flags hashpspool stripe_width 0 application mystorage read_balance_score 4.49
pool 28 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 13 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 
lfor 0/0/326 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 29 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor 0/0/326 
flags hashpspool stripe_width 0 application mystorage read_balance_score 3.37
pool 30 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 14 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 
lfor 0/0/328 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 31 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor 0/0/328 
flags hashpspool stripe_width 0 application mystorage read_balance_score 3.38
pool 32 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 15 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 
lfor 0/0/330 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 33 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor 0/0/330 
flags hashpspool stripe_width 0 application mystorage read_balance_score 4.52
pool 34 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 16 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 
lfor 0/0/332 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 35 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor 0/0/332 
flags hashpspool stripe_width 0 application mystorage read_balance_score 3.38
pool 36 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 17 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 
lfor 0/0/334 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 37 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor 0/0/334 
flags hashpspool stripe_width 0 application mystorage read_balance_score 3.38
pool 38 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 18 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 
lfor 0/0/336 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 39 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor 0/0/336 
flags hashpspool stripe_width 0 application mystorage read_balance_score 2.27
pool 40 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 19 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 
lfor 0/0/338 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 41 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor 0/0/338 
flags hashpspool stripe_width 0 application mystorage read_balance_score 4.52
pool 42 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 20 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 
lfor 0/0/340 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 43 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor 0/0/340 
flags hashpspool stripe_width 0 application mystorage read_balance_score 3.36
pool 44 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 21 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 
lfor 0/0/342 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 45 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor 0/0/342 
flags hashpspool stripe_width 0 application mystorage read_balance_score 3.37
pool 46 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 22 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 
lfor 0/0/344 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage
pool 47 'xxx.meta' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 lfor 0/0/344 
flags hashpspool stripe_width 0 application mystorage read_balance_score 3.37
pool 48 'xxx.data' erasure profile k4m2 size 6 min_size 5 crush_rule 23 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 348 
lfor 0/0/346 flags hashpspool,ec_overwrites stripe_width 16384 application 
mystorage

-- 
| Jan "Yenya" Kasprzak <kas at {fi.muni.cz - work | yenya.net - private}> |
| https://www.fi.muni.cz/~kas/                        GPG: 4096R/A45477D5 |
    We all agree on the necessity of compromise. We just can't agree on
    when it's necessary to compromise.                     --Larry Wall
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to