It's stored in the OSDMap on the monitors.
Software Engineer #42 @ http://inktank.com | http://ceph.com


On Mon, Sep 8, 2014 at 10:50 AM, JIten Shah <[email protected]> wrote:
> So, if it doesn’t refer to the entry in ceph.conf. Where does it actually 
> store the new value?
>
> —Jiten
>
> On Sep 8, 2014, at 10:31 AM, Gregory Farnum <[email protected]> wrote:
>
>> On Mon, Sep 8, 2014 at 10:08 AM, JIten Shah <[email protected]> wrote:
>>> While checking the health of the cluster, I ran to the following error:
>>>
>>> warning: health HEALTH_WARN too few pgs per osd (1< min 20)
>>>
>>> When I checked the pg and php numbers, I saw the value was the default value
>>> of 64
>>>
>>> ceph osd pool get data pg_num
>>> pg_num: 64
>>> ceph osd pool get data pgp_num
>>> pgp_num: 64
>>>
>>> Checking the ceph documents, I updated the numbers to 2000 using the
>>> following commands:
>>>
>>> ceph osd pool set data pg_num 2000
>>> ceph osd pool set data pgp_num 2000
>>>
>>> It started resizing the data and saw health warnings again:
>>>
>>> health HEALTH_WARN 1 requests are blocked > 32 sec; pool data pg_num 2000 >
>>> pgp_num 64
>>>
>>> and then:
>>>
>>> ceph health detail
>>> HEALTH_WARN 6 requests are blocked > 32 sec; 3 osds have slow requests
>>> 5 ops are blocked > 65.536 sec
>>> 1 ops are blocked > 32.768 sec
>>> 1 ops are blocked > 32.768 sec on osd.16
>>> 1 ops are blocked > 65.536 sec on osd.77
>>> 4 ops are blocked > 65.536 sec on osd.98
>>> 3 osds have slow requests
>>>
>>> This error also went away after a day.
>>>
>>> ceph health detail
>>> HEALTH_OK
>>>
>>>
>>> Now, the question I have is, will this pg number remain effective on the
>>> cluster, even if we restart MON or OSD’s on the individual disks?  I haven’t
>>> changed the values in /etc/ceph/ceph.conf. Do I need to make a change to the
>>> ceph.conf and push that change to all the MON, MSD and OSD’s ?
>>
>> It's durable once the commands are successful on the monitors. You're all 
>> done.
>> -Greg
>> Software Engineer #42 @ http://inktank.com | http://ceph.com
>> _______________________________________________
>> ceph-users mailing list
>> [email protected]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to