Hi Cephers,

We've been testing drive failures and we're just trying to see if the
behaviour of our cluster is normal,  or if we've setup something wrong.

In summary; the OSD is down and out, but the PGs are showing as degraded
and don't seem to want to remap. We'd have assumed once the OSD was marked
out, that a re-map should have happened and we'd see misplaced rather than
degraded PGs.

  cluster bfb7e824-f37d-45c0-a4fc-a98182fed985
     health HEALTH_WARN
            43 pgs degraded
            43 pgs stuck degraded
            44 pgs stuck unclean
            43 pgs stuck undersized
            43 pgs undersized
            recovery 36899/6822836 objects degraded (0.541%)
            recovery 813/6822836 objects misplaced (0.012%)
     monmap e3: 3 mons at
{ceph-admin1=10.66.8.1:6789/0,ceph-store1=10.66.8.2:6789/0,ceph-store2=10.
66.8.3:6789/0}
            election epoch 950, quorum 0,1,2
ceph-admin1,ceph-store1,ceph-store2
     osdmap e6342: 36 osds: 35 up, 35 in; 1 remapped pgs
      pgmap v11805515: 1700 pgs, 3 pools, 13165 GB data, 3331 kobjects
            25941 GB used, 30044 GB / 55986 GB avail
            36899/6822836 objects degraded (0.541%)
            813/6822836 objects misplaced (0.012%)
                1656 active+clean
                  43 active+undersized+degraded
                   1 active+remapped
  client io 491 kB/s rd, 3998 kB/s wr, 480 op/s


# id    weight  type name       up/down reweight
-6      43.56   root hdd
-2      21.78           host ceph-store1-hdd
0       3.63                    osd.0   up      1
2       3.63                    osd.2   up      1
4       3.63                    osd.4   up      1
6       3.63                    osd.6   up      1
8       3.63                    osd.8   up      1
10      3.63                    osd.10  up      1
-3      21.78           host ceph-store2-hdd
1       3.63                    osd.1   up      1
3       3.63                    osd.3   up      1
5       3.63                    osd.5   up      1
7       3.63                    osd.7   up      1
9       3.63                    osd.9   up      1
11      3.63                    osd.11  up      1
-1      11.48   root ssd
-4      5.74            host ceph-store1-ssd
12      0.43                    osd.12  up      1
13      0.43                    osd.13  up      1
14      0.43                    osd.14  up      1
16      0.43                    osd.16  up      1
18      0.43                    osd.18  down    0
19      0.43                    osd.19  up      1
20      0.43                    osd.20  up      1
21      0.43                    osd.21  up      1
32      0.72                    osd.32  up      1
33      0.72                    osd.33  up      1
17      0.43                    osd.17  up      1
15      0.43                    osd.15  up      1
-5      5.74            host ceph-store2-ssd
22      0.43                    osd.22  up      1
23      0.43                    osd.23  up      1
24      0.43                    osd.24  up      1
25      0.43                    osd.25  up      1
26      0.43                    osd.26  up      1
27      0.43                    osd.27  up      1
28      0.43                    osd.28  up      1
29      0.43                    osd.29  up      1
30      0.43                    osd.30  up      1
31      0.43                    osd.31  up      1
34      0.72                    osd.34  up      1
35      0.72                    osd.35  up      1

Are we misunderstanding the default behaviour? Any help you can provide
will be very much appreciated.

Regards,
Daniel

W: www.3ca.com.au
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to