Hmm...... The crushmap looks fine to me. Could you have a pg dump (by
ceph pg dump -o -) and check the pg membership to see if the third osd
is responsible for some pgs? And as Greg asked, it would be more
helpful if you could provide more info/description about this problem.

Thanks,
Henry

2011/4/28 Ajit K Jena <[email protected]>:
> Hi Henry,
>
> Thanks for your quick reply. I believe our crushmap is
> updated correctly to take care of osd node failures. I
> am attaching the text version of the crushmap with this
> mail. Pl help us find the mistake if any.
>
> Regards.
>
> --ajit
>
>
> On Thu, 2011-04-28 at 17:10 +0800, Henry Chang wrote:
>> 2011/4/28 Ajit K Jena <[email protected]>:
>> > Hi All,
>> >
>> > We have been experimenting with ceph since version 0.24 and
>> > found one important issue that is not resolved (or may be
>> > we don't know how to configure properly).
>> >
>> > I will talk about 0.27 since that is the latest release:
>> >
>> >  We have a one mon, one mds, and two osd setup. The data
>> >  stored goes to both the osds. We introduce a third osd.
>> >  the ceph status shows "osd: 3 up, 3 in" but no data goes
>> >  into the third osd. This remains the case even if we
>> >  power off one of the original osds.
>> >
>> > Data reliability thru replication is one of the major goals
>> > of ceph. I am wondering as to what we might be missing to
>> > get this feature going. I am attaching the configuration
>> > file.
>> >
>> > Pl help us solve this problem.
>> >
>> > Regards.
>> >
>> > --ajit
>> >
>>
>> Hi Ajit,
>>
>> The symptom looks like the crushmap has not been updated after you
>> introduced the third osd, so the data never goes into it. If that is
>> the case, you may refer to the wiki page
>> (http://ceph.newdream.net/wiki/OSD_cluster_expansion/contraction#Include_the_new_OSD_in_the_data_placement)
>> to adjust the crushmap of your cluster.
>>
>> Cheers,
>> --
>> Henry
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to [email protected]
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
>



-- 
Henry Chang
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to