here yiu are
The osd that have been added is osd.40 ssd, and it's an nautilus cluster
Thanks for helping
------------- Crush Map
ID CLASS WEIGHT (compat) TYPE NAME
-26 33.37372 root HDD10
-33 6.67474 6.67474 host HDD10-ceph01
1 hdd10 1.66869 1.66869 osd.1
5 hdd10 1.66869 1.66869 osd.5
10 hdd10 1.66869 1.66869 osd.10
14 hdd10 1.66869 1.66869 osd.14
-34 6.67474 6.67474 host HDD10-ceph02
2 hdd10 1.66869 1.66869 osd.2
7 hdd10 1.66869 1.66869 osd.7
11 hdd10 1.66869 1.66869 osd.11
15 hdd10 1.66869 1.66869 osd.15
-35 6.67474 6.67474 host HDD10-ceph03
0 hdd10 1.66869 1.66869 osd.0
6 hdd10 1.66869 1.66869 osd.6
12 hdd10 1.66869 1.66869 osd.12
16 hdd10 1.66869 1.66869 osd.16
-36 6.67474 6.67474 host HDD10-ceph04
3 hdd10 1.66869 1.66869 osd.3
8 hdd10 1.66869 1.66869 osd.8
13 hdd10 1.66869 1.66869 osd.13
17 hdd10 1.66869 1.66869 osd.17
-37 6.67474 6.67474 host HDD10-ceph05
4 hdd10 1.66869 1.66869 osd.4
34 hdd10 1.66869 1.66869 osd.34
35 hdd10 1.66869 1.66869 osd.35
36 hdd10 1.66869 1.66869 osd.36
-25 10.53297 root SSD
-42 3.54659 3.54659 host SSD-ceph01
31 ssd 1.74660 1.74660 osd.31
40 ssd 1.79999 1.79999 osd.40
-41 1.74660 1.74660 host SSD-ceph02
30 ssd 1.74660 1.74660 osd.30
-40 1.74660 1.74660 host SSD-ceph03
33 ssd 1.74660 1.74660 osd.33
-39 1.74660 1.74660 host SSD-ceph04
32 ssd 1.74660 1.74660 osd.32
-38 1.74660 1.74660 host SSD-ceph05
39 ssd 1.74660 1.74660 osd.39
-1 166.64085 root default
-3 33.32817 33.32817 host ceph01
18 hdd 11.10939 11.10939 osd.18
22 hdd 11.10939 11.10939 osd.22
26 hdd 11.10939 11.10939 osd.26
-5 33.32817 33.32817 host ceph02
19 hdd 11.10939 11.10939 osd.19
23 hdd 11.10939 11.10939 osd.23
27 hdd 11.10939 11.10939 osd.27
-9 33.32817 33.32817 host ceph03
21 hdd 11.10939 11.10939 osd.21
25 hdd 11.10939 11.10939 osd.25
29 hdd 11.10939 11.10939 osd.29
-7 33.32817 33.32817 host ceph04
20 hdd 11.10939 11.10939 osd.20
24 hdd 11.10939 11.10939 osd.24
28 hdd 11.10939 11.10939 osd.28
-11 33.32817 33.32817 host ceph05
9 hdd 11.10939 11.10939 osd.9
37 hdd 11.10939 11.10939 osd.37
38 hdd 11.10939 11.10939 osd.38
------------- Crush rules
[
{
"rule_id": 0,
"rule_name": "replicated_rule",
"ruleset": 0,
"type": 1,
"min_size": 1,
"max_size": 10,
"steps": [
{
"op": "take",
"item": -1,
"item_name": "default"
},
{
"op": "chooseleaf_firstn",
"num": 0,
"type": "host"
},
{
"op": "emit"
}
]
},
{
"rule_id": 1,
"rule_name": "replicated-ssd",
"ruleset": 1,
"type": 1,
"min_size": 1,
"max_size": 10,
"steps": [
{
"op": "take",
"item": -32,
"item_name": "SSD~ssd"
},
{
"op": "chooseleaf_firstn",
"num": 0,
"type": "host"
},
{
"op": "emit"
}
]
},
{
"rule_id": 2,
"rule_name": "replicated-hdd10",
"ruleset": 2,
"type": 1,
"min_size": 1,
"max_size": 10,
"steps": [
{
"op": "take",
"item": -27,
"item_name": "HDD10~hdd10"
},
{
"op": "chooseleaf_firstn",
"num": 0,
"type": "host"
},
{
"op": "emit"
}
]
},
{
"rule_id": 3,
"rule_name": "replicated-hdd72",
"ruleset": 3,
"type": 1,
"min_size": 1,
"max_size": 10,
"steps": [
{
"op": "take",
"item": -18,
"item_name": "default~hdd"
},
{
"op": "chooseleaf_firstn",
"num": 0,
"type": "host"
},
{
"op": "emit"
}
]
},
{
"rule_id": 4,
"rule_name": "ec-ssd",
"ruleset": 4,
"type": 3,
"min_size": 3,
"max_size": 5,
"steps": [
{
"op": "set_chooseleaf_tries",
"num": 5
},
{
"op": "set_choose_tries",
"num": 100
},
{
"op": "take",
"item": -32,
"item_name": "SSD~ssd"
},
{
"op": "chooseleaf_indep",
"num": 0,
"type": "host"
},
{
"op": "emit"
}
]
},
{
"rule_id": 5,
"rule_name": "ec-hdd10",
"ruleset": 5,
"type": 3,
"min_size": 3,
"max_size": 5,
"steps": [
{
"op": "set_chooseleaf_tries",
"num": 5
},
{
"op": "set_choose_tries",
"num": 100
},
{
"op": "take",
"item": -27,
"item_name": "HDD10~hdd10"
},
{
"op": "chooseleaf_indep",
"num": 0,
"type": "host"
},
{
"op": "emit"
}
]
},
{
"rule_id": 6,
"rule_name": "ec-hdd72",
"ruleset": 6,
"type": 3,
"min_size": 3,
"max_size": 5,
"steps": [
{
"op": "set_chooseleaf_tries",
"num": 5
},
{
"op": "set_choose_tries",
"num": 100
},
{
"op": "take",
"item": -18,
"item_name": "default~hdd"
},
{
"op": "chooseleaf_indep",
"num": 0,
"type": "host"
},
{
"op": "emit"
}
]
}
{
"rule_id": 6,
"rule_name": "ec-hdd72",
"ruleset": 6,
"type": 3,
"min_size": 3,
"max_size": 5,
"steps": [
{
"op": "set_chooseleaf_tries",
"num": 5
},
{
"op": "set_choose_tries",
"num": 100
},
{
"op": "take",
"item": -18,
"item_name": "default~hdd"
},
{
"op": "chooseleaf_indep",
"num": 0,
"type": "host"
},
{
"op": "emit"
}
]
}
]
------------- Crush Map
cluster:
id: 8700e000-d2ac-4393-8380-1cf4779166b5
health: HEALTH_WARN
nodeep-scrub flag(s) set
71 pgs not deep-scrubbed in time
services:
mon: 5 daemons, quorum ceph01,ceph02,ceph03,ceph05,ceph04 (age 26h)
mgr: ceph04(active, since 2d), standbys: ceph03, ceph05, ceph02, ceph01
mds: cephfs:1 {0=ceph02=up:active} 4 up:standby
osd: 41 osds: 41 up (since 2h), 41 in (since 2h)
flags nodeep-scrub
rgw: 5 daemons active (ceph01.rgw0, ceph02.rgw0, ceph03.rgw0, ceph04.rgw0,
ceph05.rgw0)
data:
pools: 23 pools, 1576 pgs
objects: 11.55M objects, 44 TiB
usage: 84 TiB used, 126 TiB / 210 TiB avail
pgs: 1576 active+clean
io:
client: 154 MiB/s rd, 51 MiB/s wr, 2.02k op/s rd, 1.88k op/s wr
cache: 23 MiB/s flush, 144 MiB/s evict, 44 op/s promote, 1 PGs flushing
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]