[ceph-users] Re: MDS Behind on Trimming...

2024-04-07 Thread Erich Weiler
Ah, I see. Yes, we are already running version 18.2.1 on the server side (we just installed this cluster a few weeks ago from scratch). So I guess if the fix has already been backported to that version, then we still have a problem. Dos that mean it could be the locker order bug

[ceph-users] Re: MDS Behind on Trimming...

2024-04-07 Thread Alexander E. Patrakov
Hi Erich, On Mon, Apr 8, 2024 at 11:51 AM Erich Weiler wrote: > > Hi Xiubo, > > > Thanks for your logs, and it should be the same issue with > > https://tracker.ceph.com/issues/62052, could you try to test with this > > fix again ? > > This sounds good - but I'm not clear on what I should do? I

[ceph-users] Re: MDS Behind on Trimming...

2024-04-07 Thread Erich Weiler
Hi Xiubo, Thanks for your logs, and it should be the same issue with https://tracker.ceph.com/issues/62052, could you try to test with this fix again ? This sounds good - but I'm not clear on what I should do? I see a patch in that tracker page, is that what you are referring to? If so,

[ceph-users] Re: MDS Behind on Trimming...

2024-04-07 Thread Xiubo Li
Hi Erich, Thanks for your logs, and it should be the same issue with https://tracker.ceph.com/issues/62052, could you try to test with this fix again ? Please let me know if you still could see this bug then it should be the locker order bug as https://tracker.ceph.com/issues/62123.

[ceph-users] DB/WALL and RGW index on the same NVME

2024-04-07 Thread Lukasz Borek
Hi! I'm working on a POC cluster setup dedicated to backup app writing objects via s3 (large objects, up to 1TB transferred via multipart upload process). Initial setup is 18 storage nodes (12HDDs + 1 NVME card for DB/WALL) + EC pool. Plan is to use cephadm. I'd like to follow good practice

[ceph-users] Multiple MDS Daemon needed?

2024-04-07 Thread Erich Weiler
Hi All, We have a slurm cluster with 25 clients, each with 256 cores, each mounting a cephfs filesystem as their main storage target. The workload can be heavy at times. We have two active MDS daemons and one standby. A lot of the time everything is healthy but we sometimes get warnings