On Wed, Feb 18, 2015 at 7:53 PM, Brian Rak <b...@gameservers.com> wrote:
> We're running ceph version 0.87 (c51c8f9d80fa4e0168aa52685b8de40e42758578),
> and seeing this:
>
> HEALTH_WARN 1 pgs degraded; 1 pgs stuck degraded; 1 pgs stuck unclean; 1 pgs
> stuck undersized; 1 pgs undersized
> pg 4.2af is stuck unclean for 77192.522960, current state
> active+undersized+degraded, last acting [50,42]
> pg 4.2af is stuck undersized for 980.617479, current state
> active+undersized+degraded, last acting [50,42]
> pg 4.2af is stuck degraded for 980.617902, current state
> active+undersized+degraded, last acting [50,42]
> pg 4.2af is active+undersized+degraded, acting [50,42]
>
>
> However, ceph pg query doesn't really show any issues:
> https://gist.githubusercontent.com/devicenull/9d911362e4de83c02e40/raw/565fe18163e261c8105e5493a4e90cc3c461ed9d/gistfile1.txt
> (too long to post here)
>
> I've also tried:
>
> # ceph pg 4.2af mark_unfound_lost revert
> pg has no unfound objects
>
> How can I get Ceph to rebuild here?  The replica count is 3, but I can't
> seem to figure out what's going on here.  Enabling various debug logs
> doesn't reveal anything obvious to me.
>
> I've tried restarting both OSDs, which did nothing.

What does your crushmap look like (ceph osd getcrushmap -o
/tmp/crushmap; crushtool -d /tmp/crushmap)? Does your placement logic
prevent Ceph from selecting an OSD for the third replica?

Cheers,
Florian
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to