Hi,
I have the following configuration of OSDs:
ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS
0 hdd 5.45599 1.00000 5587G 2259G 3327G 40.45 1.10 234
1 hdd 5.45599 1.00000 5587G 2295G 3291G 41.08 1.11 231
2 hdd 5.45599 1.00000 5587G 2321G 3265G 41.56 1.13 232
3 hdd 5.45599 1.00000 5587G 2313G 3273G 41.42 1.12 234
4 hdd 5.45599 1.00000 5587G 2105G 3481G 37.68 1.02 212
5 hdd 5.45599 1.00000 5587G 2231G 3355G 39.94 1.08 218
41 ssd 0.87299 1.00000 894G 9637M 884G 1.05 0.03 31
42 ssd 0.87299 1.00000 894G 13361M 881G 1.46 0.04 41
6 hdd 5.45599 1.00000 5587G 2404G 3182G 43.03 1.17 239
7 hdd 5.45599 1.00000 5587G 2226G 3360G 39.85 1.08 222
8 hdd 5.45599 1.00000 5587G 2668G 2918G 47.76 1.30 256
9 hdd 5.45599 1.00000 5587G 2366G 3220G 42.36 1.15 236
10 hdd 5.45599 1.00000 5587G 2454G 3132G 43.92 1.19 253
11 hdd 5.45599 1.00000 5587G 2405G 3181G 43.06 1.17 245
43 ssd 0.87299 1.00000 894G 15498M 879G 1.69 0.05 47
44 ssd 0.87299 1.00000 894G 10104M 884G 1.10 0.03 27
12 hdd 5.45599 1.00000 5587G 2242G 3344G 40.14 1.09 229
13 hdd 5.45599 1.00000 5587G 2551G 3035G 45.67 1.24 247
14 hdd 5.45599 1.00000 5587G 2513G 3074G 44.98 1.22 245
15 hdd 5.45599 1.00000 5587G 2014G 3572G 36.06 0.98 209
16 hdd 5.45599 1.00000 5587G 2586G 3000G 46.29 1.26 249
17 hdd 5.45599 1.00000 5587G 2459G 3127G 44.02 1.19 243
45 ssd 0.87299 1.00000 894G 9697M 884G 1.06 0.03 35
46 ssd 0.87299 1.00000 894G 12975M 881G 1.42 0.04 37
18 hdd 3.63699 1.00000 3724G 1595G 2128G 42.84 1.16 156
19 hdd 3.63699 1.00000 3724G 1387G 2336G 37.25 1.01 147
20 hdd 3.63699 1.00000 3724G 1551G 2172G 41.67 1.13 157
21 hdd 3.63699 1.00000 3724G 1535G 2189G 41.22 1.12 155
22 hdd 3.63699 1.00000 3724G 1459G 2264G 39.20 1.06 155
23 hdd 3.63699 1.00000 3724G 1395G 2329G 37.46 1.02 147
24 hdd 3.63699 1.00000 3724G 1489G 2234G 40.00 1.08 160
25 hdd 3.63699 1.00000 3724G 1634G 2090G 43.88 1.19 159
26 hdd 3.63699 1.00000 3724G 1566G 2157G 42.06 1.14 154
49 ssd 0.87299 1.00000 894G 9385M 884G 1.03 0.03 32
50 ssd 0.87299 1.00000 894G 12757M 881G 1.39 0.04 36
27 hdd 5.45599 1.00000 5587G 2462G 3124G 44.08 1.20 244
28 hdd 5.45599 1.00000 5587G 2314G 3272G 41.43 1.12 237
29 hdd 5.45599 1.00000 5587G 2166G 3420G 38.79 1.05 221
30 hdd 5.45599 1.00000 5587G 2484G 3102G 44.47 1.21 242
31 hdd 5.45599 1.00000 5587G 2292G 3294G 41.03 1.11 225
32 hdd 5.45599 1.00000 5587G 1982G 3604G 35.49 0.96 208
47 ssd 0.87299 1.00000 894G 12015M 882G 1.31 0.04 39
48 ssd 0.87299 1.00000 894G 14820M 879G 1.62 0.04 46
33 hdd 5.45599 1.00000 5587G 2002G 3584G 35.85 0.97 205
34 hdd 5.45599 1.00000 5587G 2069G 3517G 37.04 1.00 209
35 hdd 5.45599 1.00000 5587G 2187G 3399G 39.16 1.06 226
36 hdd 5.45599 1.00000 5587G 1821G 3765G 32.60 0.88 185
37 hdd 5.45599 1.00000 5587G 2123G 3463G 38.01 1.03 205
38 hdd 5.45599 1.00000 5587G 2197G 3390G 39.32 1.07 228
39 hdd 5.45599 1.00000 5587G 2180G 3406G 39.02 1.06 217
40 hdd 5.45599 1.00000 5587G 2232G 3354G 39.97 1.08 228
51 ssd 0.87320 1.00000 894G 14747M 879G 1.61 0.04 38
52 ssd 0.87320 1.00000 894G 7716M 886G 0.84 0.02 18
53 ssd 0.87320 1.00000 894G 12660M 881G 1.38 0.04 33
54 ssd 0.87320 1.00000 894G 11155M 883G 1.22 0.03 31
55 ssd 0.87320 1.00000 894G 9350M 885G 1.02 0.03 24
56 ssd 0.87320 1.00000 894G 13816M 880G 1.51 0.04 38
57 ssd 0.87320 1.00000 894G 10122M 884G 1.11 0.03 28
58 ssd 0.87320 1.00000 894G 10096M 884G 1.10 0.03 32
59 ssd 0.87320 1.00000 894G 13750M 880G 1.50 0.04 36
60 ssd 0.87320 1.00000 894G 16168M 878G 1.77 0.05 35
61 ssd 0.87320 1.00000 894G 11401M 883G 1.25 0.03 31
62 ssd 0.87320 1.00000 894G 12105M 882G 1.32 0.04 32
63 ssd 0.87320 1.00000 894G 13998M 880G 1.53 0.04 40
64 ssd 0.87320 1.00000 894G 17468M 877G 1.91 0.05 42
65 ssd 0.87320 1.00000 894G 9540M 884G 1.04 0.03 25
66 ssd 0.87320 1.00000 894G 15109M 879G 1.65 0.04 42
TOTAL 230T 86866G 145T 36.88
MIN/MAX VAR: 0.02/1.30 STDDEV: 22.48
With the following crush ruleset:
# begin crush map
tunable choose_local_tries 0
tunable choose_local_fallback_tries 0
tunable choose_total_tries 50
tunable chooseleaf_descend_once 1
tunable chooseleaf_vary_r 1
tunable chooseleaf_stable 1
tunable straw_calc_version 1
tunable allowed_bucket_algs 54
# devices
device 0 osd.0 class hdd
device 1 osd.1 class hdd
device 2 osd.2 class hdd
device 3 osd.3 class hdd
device 4 osd.4 class hdd
device 5 osd.5 class hdd
device 6 osd.6 class hdd
device 7 osd.7 class hdd
device 8 osd.8 class hdd
device 9 osd.9 class hdd
device 10 osd.10 class hdd
device 11 osd.11 class hdd
device 12 osd.12 class hdd
device 13 osd.13 class hdd
device 14 osd.14 class hdd
device 15 osd.15 class hdd
device 16 osd.16 class hdd
device 17 osd.17 class hdd
device 18 osd.18 class hdd
device 19 osd.19 class hdd
device 20 osd.20 class hdd
device 21 osd.21 class hdd
device 22 osd.22 class hdd
device 23 osd.23 class hdd
device 24 osd.24 class hdd
device 25 osd.25 class hdd
device 26 osd.26 class hdd
device 27 osd.27 class hdd
device 28 osd.28 class hdd
device 29 osd.29 class hdd
device 30 osd.30 class hdd
device 31 osd.31 class hdd
device 32 osd.32 class hdd
device 33 osd.33 class hdd
device 34 osd.34 class hdd
device 35 osd.35 class hdd
device 36 osd.36 class hdd
device 37 osd.37 class hdd
device 38 osd.38 class hdd
device 39 osd.39 class hdd
device 40 osd.40 class hdd
device 41 osd.41 class ssd
device 42 osd.42 class ssd
device 43 osd.43 class ssd
device 44 osd.44 class ssd
device 45 osd.45 class ssd
device 46 osd.46 class ssd
device 47 osd.47 class ssd
device 48 osd.48 class ssd
device 49 osd.49 class ssd
device 50 osd.50 class ssd
device 51 osd.51 class ssd
device 52 osd.52 class ssd
device 53 osd.53 class ssd
device 54 osd.54 class ssd
device 55 osd.55 class ssd
device 56 osd.56 class ssd
device 57 osd.57 class ssd
device 58 osd.58 class ssd
device 59 osd.59 class ssd
device 60 osd.60 class ssd
device 61 osd.61 class ssd
device 62 osd.62 class ssd
device 63 osd.63 class ssd
device 64 osd.64 class ssd
device 65 osd.65 class ssd
device 66 osd.66 class ssd
# types
type 0 osd
type 1 host
type 2 chassis
type 3 rack
type 4 row
type 5 pdu
type 6 pod
type 7 room
type 8 datacenter
type 9 region
type 10 root
# buckets
host node1 {
id -2 # do not change unnecessarily
id -8 class hdd # do not change unnecessarily
id -15 class ssd # do not change unnecessarily
# weight 34.482
alg straw2
hash 0 # rjenkins1
item osd.0 weight 5.456
item osd.1 weight 5.456
item osd.2 weight 5.456
item osd.3 weight 5.456
item osd.4 weight 5.456
item osd.5 weight 5.456
item osd.41 weight 0.873
item osd.42 weight 0.873
}
host node2 {
id -3 # do not change unnecessarily
id -9 class hdd # do not change unnecessarily
id -16 class ssd # do not change unnecessarily
# weight 34.482
alg straw2
hash 0 # rjenkins1
item osd.6 weight 5.456
item osd.7 weight 5.456
item osd.8 weight 5.456
item osd.9 weight 5.456
item osd.10 weight 5.456
item osd.11 weight 5.456
item osd.43 weight 0.873
item osd.44 weight 0.873
}
host node3 {
id -4 # do not change unnecessarily
id -10 class hdd # do not change unnecessarily
id -17 class ssd # do not change unnecessarily
# weight 34.482
alg straw2
hash 0 # rjenkins1
item osd.12 weight 5.456
item osd.13 weight 5.456
item osd.14 weight 5.456
item osd.15 weight 5.456
item osd.16 weight 5.456
item osd.17 weight 5.456
item osd.45 weight 0.873
item osd.46 weight 0.873
}
host node4 {
id -5 # do not change unnecessarily
id -11 class hdd # do not change unnecessarily
id -18 class ssd # do not change unnecessarily
# weight 34.479
alg straw2
hash 0 # rjenkins1
item osd.18 weight 3.637
item osd.21 weight 3.637
item osd.24 weight 3.637
item osd.25 weight 3.637
item osd.22 weight 3.637
item osd.26 weight 3.637
item osd.19 weight 3.637
item osd.23 weight 3.637
item osd.20 weight 3.637
item osd.49 weight 0.873
item osd.50 weight 0.873
}
host node5 {
id -6 # do not change unnecessarily
id -12 class hdd # do not change unnecessarily
id -19 class ssd # do not change unnecessarily
# weight 34.482
alg straw2
hash 0 # rjenkins1
item osd.27 weight 5.456
item osd.28 weight 5.456
item osd.29 weight 5.456
item osd.30 weight 5.456
item osd.31 weight 5.456
item osd.32 weight 5.456
item osd.47 weight 0.873
item osd.48 weight 0.873
}
host node6 {
id -7 # do not change unnecessarily
id -13 class hdd # do not change unnecessarily
id -20 class ssd # do not change unnecessarily
# weight 43.648
alg straw2
hash 0 # rjenkins1
item osd.33 weight 5.456
item osd.34 weight 5.456
item osd.35 weight 5.456
item osd.36 weight 5.456
item osd.37 weight 5.456
item osd.38 weight 5.456
item osd.39 weight 5.456
item osd.40 weight 5.456
}
host node7 {
id -22 # do not change unnecessarily
id -23 class hdd # do not change unnecessarily
id -24 class ssd # do not change unnecessarily
# weight 6.986
alg straw2
hash 0 # rjenkins1
item osd.51 weight 0.873
item osd.52 weight 0.873
item osd.53 weight 0.873
item osd.54 weight 0.873
item osd.55 weight 0.873
item osd.56 weight 0.873
item osd.57 weight 0.873
item osd.58 weight 0.873
}
host node8 {
id -25 # do not change unnecessarily
id -26 class hdd # do not change unnecessarily
id -27 class ssd # do not change unnecessarily
# weight 6.986
alg straw2
hash 0 # rjenkins1
item osd.59 weight 0.873
item osd.60 weight 0.873
item osd.61 weight 0.873
item osd.62 weight 0.873
item osd.63 weight 0.873
item osd.64 weight 0.873
item osd.65 weight 0.873
item osd.66 weight 0.873
}
root default {
id -1 # do not change unnecessarily
id -14 class hdd # do not change unnecessarily
id -21 class ssd # do not change unnecessarily
# weight 230.027
alg straw2
hash 0 # rjenkins1
item node1 weight 34.482
item node2 weight 34.482
item node3 weight 34.482
item node4 weight 34.479
item node5 weight 34.482
item node6 weight 43.649
item node7 weight 6.986
item node8 weight 6.986
}
# rules
rule replicated_ruleset {
id 0
type replicated
min_size 1
max_size 10
step take default class hdd
step chooseleaf firstn 0 type host
step emit
}
rule hybrid {
id 1
type replicated
min_size 1
max_size 10
step take default class ssd
step chooseleaf firstn 1 type host
step emit
step take default class hdd
step chooseleaf firstn -1 type host
step emit
}
rule flash {
id 2
type replicated
min_size 1
max_size 10
step take default class ssd
step chooseleaf firstn 0 type host
step emit
}
# end crush map
And this is shown by ceph df, with the used CRUSH Rule added, all pools using a
replication of 3:
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
230T 145T 86866G 36.88
POOLS:
NAME ID USED %USED MAX AVAIL OBJECTS
pool1 7 10454G 23.82 33430G 2871159 ->
replicated_ruleset
pool2 8 13333G 28.51 33430G 3660994 ->
replicated_ruleset
rbd 18 3729G 10.04 33430G 1021785 ->
replicated_ruleset
cephfs_data 24 270G 3.62 7213G 77568 ->
hybrid
cephfs_metadata 25 39781k 0 7213G 44 ->
flash
I have 23244G in SSDs in total.
As you can see, the "replicated_ruleset" chooses only HDDs, the "flash" rule
only SSDs and the "hybrid" rule uses one SSD (as primary, since HDDs primary
affinity is set to 0.5, while the SSDs is on 1.0) and the rest as HDDs.
ceph df shows MAX AVAIL for the "replicated_ruleset" and "flash" rule correctly
for my pools with replication 3. However the "hybrid" rule shows the available
space as if it was using 3 SSDs, even though it's only using one, even with a
replication of 3.
Is this a display bug or is something wrong with my configuration? And if it's
a bug, will it break anything if MAX AVAIL or %USED fills up to over 100%?
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com