Hi,
I'm running cephfs since 2 months now,
and my active msd memory usage is around 20G now (still growing).
ceph 1521539 10.8 31.2 20929836 20534868 ? Ssl janv.26 8573:34
/usr/bin/ceph-mds -f --cluster ceph --id 2 --setuser ceph --setgroup ceph
USER PID %CPU %MEM VSZ RSS TTY STAT START TIME COMMAND
this is on luminous 12.2.2
only tuning done is:
mds_cache_memory_limit = 5368709120
(5GB). I known it's a soft limit, but 20G seem quite huge vs 5GB ....
Is it normal ?
# ceph daemon mds.2 perf dump mds
{
"mds": {
"request": 1444009197,
"reply": 1443999870,
"reply_latency": {
"avgcount": 1443999870,
"sum": 1657849.656122933,
"avgtime": 0.001148095
},
"forward": 0,
"dir_fetch": 51740910,
"dir_commit": 9069568,
"dir_split": 64367,
"dir_merge": 58016,
"inode_max": 2147483647,
"inodes": 2042975,
"inodes_top": 152783,
"inodes_bottom": 138781,
"inodes_pin_tail": 1751411,
"inodes_pinned": 1824714,
"inodes_expired": 7258145573,
"inodes_with_caps": 1812018,
"caps": 2538233,
"subtrees": 2,
"traverse": 1591668547,
"traverse_hit": 1259482170,
"traverse_forward": 0,
"traverse_discover": 0,
"traverse_dir_fetch": 30827836,
"traverse_remote_ino": 7510,
"traverse_lock": 86236,
"load_cent": 144401980319,
"q": 49,
"exported": 0,
"exported_inodes": 0,
"imported": 0,
"imported_inodes": 0
}
}
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com