Hello,
There's a number of reasons I can think of why this would happen.
You say "default behavior" but looking at your map it's obvious that you
probably don't have a default cluster and crush map.
Your ceph.conf may help, too.
Regards,
Christian
On Tue, 4 Aug 2015 13:05:54 +1000 Daniel Manzau wrote:
> Hi Cephers,
>
> We've been testing drive failures and we're just trying to see if the
> behaviour of our cluster is normal, or if we've setup something wrong.
>
> In summary; the OSD is down and out, but the PGs are showing as degraded
> and don't seem to want to remap. We'd have assumed once the OSD was
> marked out, that a re-map should have happened and we'd see misplaced
> rather than degraded PGs.
>
> cluster bfb7e824-f37d-45c0-a4fc-a98182fed985
> health HEALTH_WARN
> 43 pgs degraded
> 43 pgs stuck degraded
> 44 pgs stuck unclean
> 43 pgs stuck undersized
> 43 pgs undersized
> recovery 36899/6822836 objects degraded (0.541%)
> recovery 813/6822836 objects misplaced (0.012%)
> monmap e3: 3 mons at
> {ceph-admin1=10.66.8.1:6789/0,ceph-store1=10.66.8.2:6789/0,ceph-store2=10.
> 66.8.3:6789/0}
> election epoch 950, quorum 0,1,2
> ceph-admin1,ceph-store1,ceph-store2
> osdmap e6342: 36 osds: 35 up, 35 in; 1 remapped pgs
> pgmap v11805515: 1700 pgs, 3 pools, 13165 GB data, 3331 kobjects
> 25941 GB used, 30044 GB / 55986 GB avail
> 36899/6822836 objects degraded (0.541%)
> 813/6822836 objects misplaced (0.012%)
> 1656 active+clean
> 43 active+undersized+degraded
> 1 active+remapped
> client io 491 kB/s rd, 3998 kB/s wr, 480 op/s
>
>
> # id weight type name up/down reweight
> -6 43.56 root hdd
> -2 21.78 host ceph-store1-hdd
> 0 3.63 osd.0 up 1
> 2 3.63 osd.2 up 1
> 4 3.63 osd.4 up 1
> 6 3.63 osd.6 up 1
> 8 3.63 osd.8 up 1
> 10 3.63 osd.10 up 1
> -3 21.78 host ceph-store2-hdd
> 1 3.63 osd.1 up 1
> 3 3.63 osd.3 up 1
> 5 3.63 osd.5 up 1
> 7 3.63 osd.7 up 1
> 9 3.63 osd.9 up 1
> 11 3.63 osd.11 up 1
> -1 11.48 root ssd
> -4 5.74 host ceph-store1-ssd
> 12 0.43 osd.12 up 1
> 13 0.43 osd.13 up 1
> 14 0.43 osd.14 up 1
> 16 0.43 osd.16 up 1
> 18 0.43 osd.18 down 0
> 19 0.43 osd.19 up 1
> 20 0.43 osd.20 up 1
> 21 0.43 osd.21 up 1
> 32 0.72 osd.32 up 1
> 33 0.72 osd.33 up 1
> 17 0.43 osd.17 up 1
> 15 0.43 osd.15 up 1
> -5 5.74 host ceph-store2-ssd
> 22 0.43 osd.22 up 1
> 23 0.43 osd.23 up 1
> 24 0.43 osd.24 up 1
> 25 0.43 osd.25 up 1
> 26 0.43 osd.26 up 1
> 27 0.43 osd.27 up 1
> 28 0.43 osd.28 up 1
> 29 0.43 osd.29 up 1
> 30 0.43 osd.30 up 1
> 31 0.43 osd.31 up 1
> 34 0.72 osd.34 up 1
> 35 0.72 osd.35 up 1
>
> Are we misunderstanding the default behaviour? Any help you can provide
> will be very much appreciated.
>
> Regards,
> Daniel
>
> W: www.3ca.com.au
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
--
Christian Balzer Network/Systems Engineer
[email protected] Global OnLine Japan/Fusion Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com