root@adminnode:~# ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 30.82903 root default
-16 30.82903 datacenter dc01
-19 30.82903 pod dc01-agg01
-10 17.43365 rack dc01-rack02
-4 7.20665 host node1001
0 hdd 0.90999 osd.0 up 1.00000 1.00000
1 hdd 0.90999 osd.1 up 1.00000 1.00000
5 hdd 0.90999 osd.5 up 1.00000 1.00000
29 hdd 0.90970 osd.29 up 1.00000 1.00000
32 hdd 0.90970 osd.32 down 0 1.00000
33 hdd 0.90970 osd.33 up 1.00000 1.00000
2 ssd 0.43700 osd.2 up 1.00000 1.00000
3 ssd 0.43700 osd.3 up 1.00000 1.00000
4 ssd 0.43700 osd.4 up 1.00000 1.00000
30 ssd 0.43660 osd.30 up 1.00000 1.00000
-7 6.29724 host node1002
9 hdd 0.90999 osd.9 up 1.00000 1.00000
10 hdd 0.90999 osd.10 up 1.00000 1.00000
11 hdd 0.90999 osd.11 up 1.00000 1.00000
12 hdd 0.90999 osd.12 up 1.00000 1.00000
35 hdd 0.90970 osd.35 up 1.00000 1.00000
6 ssd 0.43700 osd.6 up 1.00000 1.00000
7 ssd 0.43700 osd.7 up 1.00000 1.00000
8 ssd 0.43700 osd.8 up 1.00000 1.00000
31 ssd 0.43660 osd.31 up 1.00000 1.00000
-28 2.18318 host node1005
34 ssd 0.43660 osd.34 up 1.00000 1.00000
36 ssd 0.87329 osd.36 up 1.00000 1.00000
37 ssd 0.87329 osd.37 up 1.00000 1.00000
-29 1.74658 host node1006
42 ssd 0.87329 osd.42 up 1.00000 1.00000
43 ssd 0.87329 osd.43 up 1.00000 1.00000
-11 13.39537 rack dc01-rack03
-22 5.38794 host node1003
17 hdd 0.90999 osd.17 up 1.00000 1.00000
18 hdd 0.90999 osd.18 up 1.00000 1.00000
24 hdd 0.90999 osd.24 up 1.00000 1.00000
26 hdd 0.90999 osd.26 up 1.00000 1.00000
13 ssd 0.43700 osd.13 up 1.00000 1.00000
14 ssd 0.43700 osd.14 up 1.00000 1.00000
15 ssd 0.43700 osd.15 up 1.00000 1.00000
16 ssd 0.43700 osd.16 up 1.00000 1.00000
-25 5.38765 host node1004
23 hdd 0.90999 osd.23 up 1.00000 1.00000
25 hdd 0.90999 osd.25 up 1.00000 1.00000
27 hdd 0.90999 osd.27 up 1.00000 1.00000
28 hdd 0.90970 osd.28 up 1.00000 1.00000
19 ssd 0.43700 osd.19 up 1.00000 1.00000
20 ssd 0.43700 osd.20 up 1.00000 1.00000
21 ssd 0.43700 osd.21 up 1.00000 1.00000
22 ssd 0.43700 osd.22 up 1.00000 1.00000
-30 2.61978 host node1007
38 ssd 0.43660 osd.38 up 1.00000 1.00000
39 ssd 0.43660 osd.39 up 1.00000 1.00000
40 ssd 0.87329 osd.40 up 1.00000 1.00000
41 ssd 0.87329 osd.41 up 1.00000 1.00000
========================================================
root@adminnode:~# ceph osd df tree
ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS
TYPE NAME
-1 30.82903 - 29.9TiB 14.0TiB 16.0TiB 46.65 1.00 -
root default
-16 30.82903 - 29.9TiB 14.0TiB 16.0TiB 46.65 1.00 -
datacenter dc01
-19 30.82903 - 29.9TiB 14.0TiB 16.0TiB 46.65 1.00 -
pod dc01-agg01
-10 17.43365 - 16.5TiB 7.31TiB 9.21TiB 44.26 0.95 -
rack dc01-rack02
-4 7.20665 - 6.29TiB 2.76TiB 3.54TiB 43.83 0.94 -
host node1001
0 hdd 0.90999 1.00000 932GiB 356GiB 575GiB 38.22 0.82 95
osd.0
1 hdd 0.90999 1.00000 932GiB 397GiB 534GiB 42.66 0.91 106
osd.1
5 hdd 0.90999 1.00000 932GiB 284GiB 647GiB 30.50 0.65 76
osd.5
29 hdd 0.90970 1.00000 932GiB 366GiB 566GiB 39.29 0.84 98
osd.29
32 hdd 0.90970 0 0B 0B 0B 0 0 0
osd.32
33 hdd 0.90970 1.00000 932GiB 369GiB 563GiB 39.57 0.85 99
osd.33
2 ssd 0.43700 1.00000 447GiB 271GiB 176GiB 60.67 1.30 50
osd.2
3 ssd 0.43700 1.00000 447GiB 249GiB 198GiB 55.62 1.19 58
osd.3
4 ssd 0.43700 1.00000 447GiB 297GiB 150GiB 66.39 1.42 56
osd.4
30 ssd 0.43660 1.00000 447GiB 236GiB 211GiB 52.85 1.13 48
osd.30
-7 6.29724 - 6.29TiB 2.74TiB 3.55TiB 43.53 0.93 -
host node1002
9 hdd 0.90999 1.00000 932GiB 354GiB 578GiB 37.96 0.81 95
osd.9
10 hdd 0.90999 1.00000 932GiB 357GiB 575GiB 38.28 0.82 96
osd.10
11 hdd 0.90999 1.00000 932GiB 318GiB 613GiB 34.18 0.73 86
osd.11
12 hdd 0.90999 1.00000 932GiB 373GiB 558GiB 40.09 0.86 100
osd.12
35 hdd 0.90970 1.00000 932GiB 343GiB 588GiB 36.83 0.79 92
osd.35
6 ssd 0.43700 1.00000 447GiB 269GiB 178GiB 60.20 1.29 60
osd.6
7 ssd 0.43700 1.00000 447GiB 249GiB 198GiB 55.69 1.19 56
osd.7
8 ssd 0.43700 1.00000 447GiB 286GiB 161GiB 63.95 1.37 56
osd.8
31 ssd 0.43660 1.00000 447GiB 257GiB 190GiB 57.47 1.23 55
osd.31
-28 2.18318 - 2.18TiB 968GiB 1.24TiB 43.29 0.93 -
host node1005
34 ssd 0.43660 1.00000 447GiB 202GiB 245GiB 45.14 0.97 47
osd.34
36 ssd 0.87329 1.00000 894GiB 405GiB 489GiB 45.28 0.97 91
osd.36
37 ssd 0.87329 1.00000 894GiB 361GiB 533GiB 40.38 0.87 79
osd.37
-29 1.74658 - 1.75TiB 888GiB 900GiB 49.65 1.06 -
host node1006
42 ssd 0.87329 1.00000 894GiB 417GiB 477GiB 46.68 1.00 92
osd.42
43 ssd 0.87329 1.00000 894GiB 471GiB 424GiB 52.63 1.13 90
osd.43
-11 13.39537 - 13.4TiB 6.64TiB 6.75TiB 49.60 1.06 -
rack dc01-rack03
-22 5.38794 - 5.39TiB 2.70TiB 2.69TiB 50.14 1.07 -
host node1003
17 hdd 0.90999 1.00000 932GiB 371GiB 560GiB 39.83 0.85 100
osd.17
18 hdd 0.90999 1.00000 932GiB 390GiB 542GiB 41.82 0.90 105
osd.18
24 hdd 0.90999 1.00000 932GiB 352GiB 580GiB 37.77 0.81 94
osd.24
26 hdd 0.90999 1.00000 932GiB 387GiB 545GiB 41.54 0.89 104
osd.26
13 ssd 0.43700 1.00000 447GiB 319GiB 128GiB 71.32 1.53 77
osd.13
14 ssd 0.43700 1.00000 447GiB 303GiB 144GiB 67.76 1.45 70
osd.14
15 ssd 0.43700 1.00000 447GiB 361GiB 86.4GiB 80.67 1.73 77
osd.15
16 ssd 0.43700 1.00000 447GiB 283GiB 164GiB 63.29 1.36 71
osd.16
-25 5.38765 - 5.39TiB 2.83TiB 2.56TiB 52.55 1.13 -
host node1004
23 hdd 0.90999 1.00000 932GiB 382GiB 549GiB 41.05 0.88 102
osd.23
25 hdd 0.90999 1.00000 932GiB 412GiB 520GiB 44.20 0.95 111
osd.25
27 hdd 0.90999 1.00000 932GiB 385GiB 546GiB 41.36 0.89 103
osd.27
28 hdd 0.90970 1.00000 932GiB 462GiB 469GiB 49.64 1.06 124
osd.28
19 ssd 0.43700 1.00000 447GiB 314GiB 133GiB 70.22 1.51 75
osd.19
20 ssd 0.43700 1.00000 447GiB 327GiB 120GiB 73.06 1.57 76
osd.20
21 ssd 0.43700 1.00000 447GiB 324GiB 123GiB 72.45 1.55 77
osd.21
22 ssd 0.43700 1.00000 447GiB 292GiB 156GiB 65.21 1.40 68
osd.22
-30 2.61978 - 2.62TiB 1.11TiB 1.51TiB 42.43 0.91 -
host node1007
38 ssd 0.43660 1.00000 447GiB 165GiB 283GiB 36.82 0.79 46
osd.38
39 ssd 0.43660 1.00000 447GiB 156GiB 292GiB 34.79 0.75 42
osd.39
40 ssd 0.87329 1.00000 894GiB 429GiB 466GiB 47.94 1.03 98
osd.40
41 ssd 0.87329 1.00000 894GiB 389GiB 505GiB 43.55 0.93 103
osd.41
TOTAL 29.9TiB 14.0TiB 16.0TiB 46.65
MIN/MAX VAR: 0.65/1.73 STDDEV: 13.30
=============================================================
root@adminnode:~# ceph df && ceph -v
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
29.9TiB 16.0TiB 14.0TiB 46.65
POOLS:
NAME ID USED %USED MAX AVAIL OBJECTS
rbd_vms_ssd 2 986GiB 49.83 993GiB 262606
rbd_vms_hdd 3 3.76TiB 48.94 3.92TiB 992255
rbd_vms_ssd_01 4 372KiB 0 662GiB 148
rbd_vms_ssd_01_ec 6 2.85TiB 68.81 1.29TiB 770506
ceph version 12.2.8 (ae699615bac534ea496ee965ac6192cb7e0e07c0) luminous (stable)
Kevin
Am Sa., 5. Jan. 2019 um 05:12 Uhr schrieb Konstantin Shalygin <[email protected]>:
>
> On 1/5/19 1:51 AM, Kevin Olbrich wrote:
> > PS: Could behttp://tracker.ceph.com/issues/36361
> > There is one HDD OSD that is out (which will not be replaced because
> > the SSD pool will get the images and the hdd pool will be deleted).
>
> Paste your `ceph osd tree`, `ceph osd df tree` please.
>
>
>
> k
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com