Hello,
You how much your PG pools since he first saw you have left too big.

--
Cordialement,
Corentin BONNETON


> Le 7 juin 2016 à 15:21, Sage Weil <[email protected]> a écrit :
> 
> On Tue, 7 Jun 2016, M Ranga Swami Reddy wrote:
>> OK, understood...
>> To fix the nearfull warn, I am reducing the weight of a specific OSD,
>> which filled >85%..
>> Is this work-around advisable?
> 
> Sure.  This is what reweight-by-utilization does for you, but 
> automatically.
> 
> sage
> 
>> 
>> Thanks
>> Swami
>> 
>> On Tue, Jun 7, 2016 at 6:37 PM, Sage Weil <[email protected]> wrote:
>>> On Tue, 7 Jun 2016, M Ranga Swami Reddy wrote:
>>>> Hi Sage,
>>>>> Jewel and the latest hammer point release have an improved
>>>>> reweight-by-utilization (ceph osd test-reweight-by-utilization ... to dry
>>>>> run) to correct this.
>>>> 
>>>> Thank you....But not planning to upgrade the cluster soon.
>>>> So, in this case - are there any tunable options will help? like
>>>> "crush tunable optimal" or so?
>>>> OR any other configuration options change will help?
>>> 
>>> Firefly also has reweight-by-utilization... it's just a bit less friendly
>>> than the newer versions.  CRUSH tunables don't generally help here unless
>>> you have lots of OSDs that are down+out.
>>> 
>>> Note that firefly is no longer supported.
>>> 
>>> sage
>>> 
>>> 
>>>> 
>>>> 
>>>> Thanks
>>>> Swami
>>>> 
>>>> 
>>>> On Tue, Jun 7, 2016 at 6:00 PM, Sage Weil <[email protected]> wrote:
>>>>> On Tue, 7 Jun 2016, M Ranga Swami Reddy wrote:
>>>>>> Hello,
>>>>>> I have aorund 100 OSDs in my ceph cluster. In this a few OSDs filled
>>>>>> with >85% of data and few OSDs filled with ~60%-70% of data.
>>>>>> 
>>>>>> Any reason why the unevenly OSDs filling happned? do I need to any
>>>>>> tweaks on configuration to fix the above? Please advise.
>>>>>> 
>>>>>> PS: Ceph version is - 0.80.7
>>>>> 
>>>>> Jewel and the latest hammer point release have an improved
>>>>> reweight-by-utilization (ceph osd test-reweight-by-utilization ... to dry
>>>>> run) to correct this.
>>>>> 
>>>>> sage
>>>>> 
>>>> --
>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>> the body of a message to [email protected]
>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>> 
>>>> 
>> 
>> 
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to