Hi Brett,

Can you enable debug_bluestore = 5 and debug_prioritycache = 5 on one of the OSDs that's showing the behavior?  You'll want to look in the logs for lines that look like this:


2019-07-18T19:34:42.587-0400 7f4048b8d700  5 prioritycache tune_memory target: 4294967296 mapped: 4260962304 unmapped: 856948736 heap: 5117911040 old mem: 2845415707 new mem: 2845415707 2019-07-18T19:34:33.527-0400 7f4048b8d700  5 bluestore.MempoolThread(0x55a6d330ead0) _resize_shards cache_size: 2845415707 kv_alloc: 1241513984 kv_used: 874833889 meta_alloc: 1258291200 meta_used: 889040246 data_alloc: 318767104 data_used: 0

The first line will tell you what your memory target is set to, how much memory is currently mapped, how much is unmapped (ie what's been freed but the kernel hasn't reclaimed), the total heap size, and the old and new aggregate size for all of bluestores caches.  The second line also tells you the aggregate cache size, and then how much space is being allocated and used for the kv, meta, and data caches.  If there's a leak somewhere in the OSD or bluestore the autotuner will shrink the cache way down but eventually won't be able to contain it and eventually your process will start growing beyond the target size despite having a tiny amount of bluestore cache.  If it's something else like a huge amount of freed memory not being reclaimed by the kernel, you'll see large amount of unmapped memory and a big heap size despite the mapped memory staying near the target.  If it's a bug in the autotuner, we might see the mapped memory greatly exceeding the target.


Mark


On 7/18/19 4:02 PM, Brett Kelly wrote:

Hello,

We have a Nautilus cluster exhibiting what looks like this bug: https://tracker.ceph.com/issues/39618

No matter what is set as the osd_memory_target (currently 2147483648 ), each OSD process will surpass this value and peak around ~4.0GB then eventually start using swap. Cluster stays stable for about a week and then starts running into OOM issues, kills off OSDs and requires a reboot of each node to get back to a stable state.

Has anyone run into similar/workarounds ?

Ceph version: 14.2.1, RGW Clients

CentOS Linux release 7.6.1810 (Core)

Kernel: 3.10.0-957.12.1.el7.x86_64

256GB RAM per OSD node, 60 OSD's in each node.


Thanks,

--
Brett Kelly


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to