IMHO the autotuner is awesome.

1TB of RAM is an embarrassment of riches -- are these hosts perhaps converged 
compute+storage?



> On Jul 31, 2025, at 10:17 AM, Steven Vacaroaia <ste...@gmail.com> wrote:
> 
> Hi
> 
> What is the best practice / your expert advice about using
> osd_memory_target_autotune
> on hosts with lots of RAM  ?
> 
> My hosts have 1 TB RAM , only 3 NVMEs , 12 HDD and 12 SSD

Remember that NVMe devices *are* SSDs ;)  I'm guessing those are used for 
WAL+DB offload, and thus you have 24x OSDs per host?

> Should I disable autotune and allocate more RAM?

The autotuner by default will divide 70% of physmem across all the OSDs it 
finds on a given host, with 30% allocated for the OS and other daemons.  I 
*think* any RGWs, mons, etc. are assumed to be part of that 30% but am not 
positive.

> 
> I saw some suggestion for 16GB to NVME , 8GB to SSD and 6 to HDD

I personally have a growing sense that more RAM actually can help slower OSDs 
more, at least with respect to rebalancing without rampant slow ops.  ymmv.

This implies that your NVMe devices are standalone OSDs, so that would mean 27 
OSDs per node?  I'm curious what manner of chassis this is.

I then would think that the autotuner would set ~~ 26TB to osd_memory_size, 
which is ample by any measure.  ~307TB will be available for non-OSD processes.


If you're running compute or other significant non-Ceph workloads on the same 
nodes, you can adjust the reservation factor by setting ceph config set mgr 
mgr/cephadm/autotune_memory_target_ratio xxx.  So if you want to reserve less 
for non-OSD processes, something like

ceph config set mgr mgr/cephadm/autotune_memory_target_ratio 0.1

If yo do have hungry compute colocated, a good value might be something like 
0.25, which would give each OSD > 9GB for osd_memory_target.  If you do want to 
allot different amounts to different device classes, you can instead set static 
values, using central config device class masks.



> 
> Many thanks
> Steven
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to