On 12/07/16 14:58, Wido den Hollander wrote:
>> Op 7 december 2016 om 11:29 schreef Kees Meijs <[email protected]>:
>>
>>
>> Hi Wido,
>>
>> Valid point. At this moment, we're using a cache pool with size = 2 and
>> would like to "upgrade" to size = 3.
>>
>> Again, you're absolutely right... ;-)
>>
>> Anyway, any things to consider or could we just:
>>
>> 1. Run "ceph osd pool set cache size 3".
>> 2. Wait for rebalancing to complete.
>> 3. Run "ceph osd pool set cache min_size 2".
>>
> Indeed! It is a simple as that.
>
> Your cache pool can also contain very valuable data you do not want to loose.
>
> Wido
Almost as simple as that...
First make sure there is free space. Then when you run it, also monitor
that there are no side effects... bad performance, blocked requests, etc.
And if there are issues, be ready to stop it with:
ceph osd set nobackfill
ceph osd set norecover
And then figure out some tuning... eg. (very minimal settings)
# more than likely you can handle more than 1 on a small cluster and
maybe much more
ceph tell osd.* injectargs --osd_max_backfills=1
# manuals/emails/something I read suggest a number like 0.05 ... I
find that does nothing in times of real trouble, but this really slows
down recovery
ceph tell osd.* injectargs --osd_recovery_sleep=0.6
ceph osd set noscrub
ceph osd set nodeep-scrub
# also I think this one is highly relevant, but not sure what to
suggest for it... others suggest 12[1] to 16[2], and so far I found 8
works better than 12-32 for my frequent "blocked requests" small cluster
# --osd_op_threads=...
And then unset those flags to resume it. And when done, consider
unsetting your new settings (I would unset the noscrub at least).
[1]
https://www.openstack.org/summit/vancouver-2015/summit-videos/presentation/a-year-with-cinder-and-ceph-at-twc
[2] http://www.spinics.net/lists/ceph-users/msg32368.html somewhere in
this thread, but can't find it online "We have recently increase osd op
threads from 2 (default value) to 16 because CPU usage on DN was very
low. We have the impression it has increased overall ceph cluster
performances and reduced block ops occurrences."
--
--------------------------------------------
Peter Maloney
Brockmann Consult
Max-Planck-Str. 2
21502 Geesthacht
Germany
Tel: +49 4152 889 300
Fax: +49 4152 889 333
E-mail: [email protected]
Internet: http://www.brockmann-consult.de
--------------------------------------------
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com