Just to close this thread up - it looks like all the problems were
related to setting the "mds cache size" option in Luminous instead of
using "mds cache memory limit". The "mds cache size" option
documentation says that "it is recommended to use mds_cache_memory_limit
...", but it looks more
Hi Andras,
On Thu, Jan 18, 2018 at 3:38 AM, Andras Pataki
wrote:
> Hi John,
>
> Some other symptoms of the problem: when the MDS has been running for a few
> days, it starts looking really busy. At this time, listing directories
> becomes really slow. An "ls -l"
Hi John,
Some other symptoms of the problem: when the MDS has been running for a
few days, it starts looking really busy. At this time, listing
directories becomes really slow. An "ls -l" on a directory with about
250 entries takes about 2.5 seconds. All the metadata is on OSDs with
NVMe
On Wed, Jan 17, 2018 at 3:36 PM, Andras Pataki
wrote:
> Hi John,
>
> All our hosts are CentOS 7 hosts, the majority are 7.4 with kernel
> 3.10.0-693.5.2.el7.x86_64, with fuse 2.9.2-8.el7. We have some hosts that
> have slight variations in kernel versions, the
Hi John,
All our hosts are CentOS 7 hosts, the majority are 7.4 with kernel
3.10.0-693.5.2.el7.x86_64, with fuse 2.9.2-8.el7. We have some hosts
that have slight variations in kernel versions, the oldest one are a
handful of CentOS 7.3 hosts with kernel 3.10.0-514.21.1.el7.x86_64 and
fuse
On Tue, Jan 16, 2018 at 8:50 PM, Andras Pataki
wrote:
> Dear Cephers,
>
> We've upgraded the back end of our cluster from Jewel (10.2.10) to Luminous
> (12.2.2). The upgrade went smoothly for the most part, except we seem to be
> hitting an issue with cephfs.
Dear Cephers,
We've upgraded the back end of our cluster from Jewel (10.2.10) to
Luminous (12.2.2). The upgrade went smoothly for the most part, except
we seem to be hitting an issue with cephfs. After about a day or two of
use, the MDS start complaining about clients failing to respond to