Hello,
I currently have 5 servers within my CEPH Cluster
2 x (10 * 8TB Disks)
3 x (10 * 4TB Disks)
Currently seeing a larger difference in OSD use across the two separate server
types, as well as within the server itself.
For example on one 4TB server I have an OSD at 64% and one at 84%, where on the
8TB servers the OSD range from 49% to 64%, where the highest used OSD's are on
the 4TB.
Each drive has a weight set correctly for the drive size and each server has
the correct weight set, below is my crush map. Apart from running the command
to adjust the re-weight is there anything I am doing wrong or should change for
better spread of data, not looking for near perfect but where the 8TB drives
are sitting at 64% max and 4TB are sitting at 80%'s causes a big inbalance.
# begin crush map
tunable choose_local_tries 0
tunable choose_local_fallback_tries 0
tunable choose_total_tries 50
tunable chooseleaf_descend_once 1
tunable chooseleaf_vary_r 1
tunable straw_calc_version 1
tunable allowed_bucket_algs 54
# buckets
host sn1 {
id -2 # do not change unnecessarily
# weight 72.800
alg straw2
hash 0 # rjenkins1
item osd.0 weight 7.280
item osd.1 weight 7.280
item osd.3 weight 7.280
item osd.4 weight 7.280
item osd.2 weight 7.280
item osd.5 weight 7.280
item osd.6 weight 7.280
item osd.7 weight 7.280
item osd.8 weight 7.280
item osd.9 weight 7.280
}
host sn3 {
id -6 # do not change unnecessarily
# weight 72.800
alg straw2
hash 0 # rjenkins1
item osd.10 weight 7.280
item osd.11 weight 7.280
item osd.12 weight 7.280
item osd.13 weight 7.280
item osd.14 weight 7.280
item osd.15 weight 7.280
item osd.16 weight 7.280
item osd.17 weight 7.280
item osd.18 weight 7.280
item osd.19 weight 7.280
}
host sn4 {
id -7 # do not change unnecessarily
# weight 36.060
alg straw2
hash 0 # rjenkins1
item osd.20 weight 3.640
item osd.21 weight 3.640
item osd.22 weight 3.640
item osd.23 weight 3.640
item osd.24 weight 3.640
item osd.25 weight 3.640
item osd.26 weight 3.640
item osd.27 weight 3.640
item osd.28 weight 3.640
item osd.29 weight 3.300
}
host sn5 {
id -8 # do not change unnecessarily
# weight 36.060
alg straw2
hash 0 # rjenkins1
item osd.30 weight 3.640
item osd.31 weight 3.640
item osd.32 weight 3.640
item osd.33 weight 3.640
item osd.34 weight 3.640
item osd.35 weight 3.640
item osd.36 weight 3.640
item osd.37 weight 3.640
item osd.38 weight 3.640
item osd.39 weight 3.640
}
host sn6 {
id -9 # do not change unnecessarily
# weight 36.060
alg straw2
hash 0 # rjenkins1
item osd.40 weight 3.640
item osd.41 weight 3.640
item osd.42 weight 3.640
item osd.43 weight 3.640
item osd.44 weight 3.640
item osd.45 weight 3.640
item osd.46 weight 3.640
item osd.47 weight 3.640
item osd.48 weight 3.640
item osd.49 weight 3.640
}
root default {
id -1 # do not change unnecessarily
# weight 253.780
alg straw2
hash 0 # rjenkins1
item sn1 weight 72.800
item sn3 weight 72.800
item sn4 weight 36.060
item sn5 weight 36.060
item sn6 weight 36.060
}
# rules
rule replicated_ruleset {
ruleset 0
type replicated
min_size 1
max_size 10
step take default
step chooseleaf firstn 0 type host
step emit
}
Thanks,
Ashley
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com