Hi,

Thanks, my warning is gone now.

2013/3/13 Jeff Anderson-Lee <[email protected]>

> On 3/13/2013 9:31 AM, Greg Farnum wrote:
>
>> Nope, it's not because you were using the cluster. The "unclean" PGs here
>> are those which are in the "active+remapped" state. That's actually two
>> states — "active" which is good, because it means they're serving reads and
>> writes"; "remapped" which means that for some reason the current set of
>> OSDs handling them isn't the set that CRUSH thinks should be handling them.
>> Given your cluster expansion that probably means that your CRUSH map and
>> rules aren't behaving themselves and are failing to assign the right number
>> of replicas to those PGs. You can check this by looking at the PG dump. If
>> you search for "ceph active remapped" it looks to me like you'll get some
>> useful results; you might also just be able to enable the CRUSH tunables (
>> http://ceph.com/docs/master/rados/operations/crush-map/#tunables).
>>
>
thanks for the hint in the manual, after the run was successful, should i
set the old default values?


>  I experienced this (stuck active+remapped) frequently with the stock 0.41
> apt-get/Ubuntu version of ceph. Less so with Bobtail.
>

i use:
ceph version 0.48.3argonaut
(commit:920f82e805efec2cae05b79c155c07df0f3ed5dd)
on ubuntu 12.04 and a 3.8.2-030802-generic Kernel

i can't upgrade to Bobtail until ganeti 2.7 is out.


Again thanks for the help,

Ansgar
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to