On Thu, Dec 21, 2017 at 11:46 PM, Webert de Souza Lima
<webert.b...@gmail.com> wrote:
> Hello Zheng,
>
> Thanks for opening that issue on the bug tracker.
>
> Also thanks for that tip. Caps dropped from 1.6M to 600k for that client.

idle client shouldn't hold so many caps.

> Is it safe to run in a cronjob? Let's say, once or twice a day during
> production?
>

yes. For now, it's better to run "echo 3 >/proc/sys/vm/drop_caches"
after cronjob finishes

> Thanks!
>
>
> Regards,
>
> Webert Lima
> DevOps Engineer at MAV Tecnologia
> Belo Horizonte - Brasil
> IRC NICK - WebertRLZ
>
> On Thu, Dec 21, 2017 at 11:55 AM, Yan, Zheng <uker...@gmail.com> wrote:
>>
>> On Thu, Dec 21, 2017 at 7:33 PM, Webert de Souza Lima
>> <webert.b...@gmail.com> wrote:
>> > I have upgraded the kernel on a client node (one that has close-to-zero
>> > traffic) used for tests.
>> >
>> >    {
>> >       "reconnecting" : false,
>> >       "id" : 1620266,
>> >       "num_leases" : 0,
>> >       "inst" : "client.1620266 10.0.0.111:0/3921220890",
>> >       "state" : "open",
>> >       "completed_requests" : 0,
>> >       "num_caps" : 1402490,
>> >       "client_metadata" : {
>> >          "kernel_version" : "4.4.0-104-generic",
>> >          "hostname" : "suppressed",
>> >          "entity_id" : "admin"
>> >       },
>> >       "replay_requests" : 0
>> >    },
>> >
>> > still 1.4M caps used.
>> >
>> > is upgrading the client kernel enough ?
>> >
>>
>> See http://tracker.ceph.com/issues/22446. We haven't implemented that
>> feature.  "echo 3 >/proc/sys/vm/drop_caches"  should drop most caps.
>>
>> >
>> >
>> > Regards,
>> >
>> > Webert Lima
>> > DevOps Engineer at MAV Tecnologia
>> > Belo Horizonte - Brasil
>> > IRC NICK - WebertRLZ
>> >
>> > On Fri, Dec 15, 2017 at 11:16 AM, Webert de Souza Lima
>> > <webert.b...@gmail.com> wrote:
>> >>
>> >> So,
>> >>
>> >> On Fri, Dec 15, 2017 at 10:58 AM, Yan, Zheng <uker...@gmail.com> wrote:
>> >>>
>> >>>
>> >>> 300k are ready quite a lot. opening them requires long time. does you
>> >>> mail server really open so many files?
>> >>
>> >>
>> >> Yes, probably. It's a commercial solution. A few thousand domains,
>> >> dozens
>> >> of thousands of users and god knows how any mailboxes.
>> >> From the daemonperf you can see the write workload is high, so yes, too
>> >> much files opening (dovecot mdbox stores multiple e-mails per file,
>> >> split
>> >> into many files).
>> >>
>> >>> I checked 4.4 kernel, it includes the code that trim cache when mds
>> >>> recovers.
>> >>
>> >>
>> >> Ok, all nodes are running 4.4.0-75-generic. The fix might have been
>> >> included in a newer version.
>> >> I'll upgrade it asap.
>> >>
>> >>
>> >> Regards,
>> >>
>> >> Webert Lima
>> >> DevOps Engineer at MAV Tecnologia
>> >> Belo Horizonte - Brasil
>> >> IRC NICK - WebertRLZ
>> >
>> >
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to