Hello Peter I hope you are not keeping images on EC Pool? In my case my all data and images majorly on ec pools
Regards Dev On Wed, 2 Jul 2025 at 1:29 PM, Anthony D'Atri <a...@dreamsnake.net> wrote: > We need to see `ceph osd crush rule dump` and `ceph osd pool ls detail` to > see which pools are using which CRUSH rule. > > Since there are two device classes, *every* pool should specify a CRUSH > rule that constrains to one or the other device class. > > If this is not done, e.g. > > root@cmigsdsc-m18-33:~# ceph osd crush rule dump > [ > { > "rule_id": 0, > "rule_name": "replicated_rule", > "type": 1, > "steps": [ > { > "op": "take", > "item": -1, > "item_name": “default” <——————————————==<<<< > }, > { > "op": "chooseleaf_firstn", > "num": 0, > "type": "host" > }, > { > "op": "emit" > } > ] > }, > > then pools whose rule specifies an item_name “default” will be placed on > both the “nvme” and “ssd” device classes. Which sort of works, but is > usually not the best approach, and may result in the balancer and pg > autoscaler not working properly if at all. Segregating the NVMe and > SAS/SATA SSDs to separate pools is usually the better option since the > former are usually faster. > > Below is an example CRUSH rule that will constrain pools specifying it to > only the `nvme` device class. > > { > "rule_id": 6, > "rule_name": "ssd_nvme_replicated", > "type": 1, > "steps": [ > { > "op": "take", > "item": -33, > "item_name": “default~nvme" > }, > { > "op": "chooseleaf_firstn", > "num": 0, > "type": "host" > }, > { > "op": "emit" > } > ] > }, > > If the pools are using only one or the other device class, weighting to > favor the NVMe SSDs as such isn’t in scope. Messing with CRUSH or legacy > override reweights is more likely to just result in OSDs filling up > prematurely. > > I’ve counseled the OP re PGs, but I think the underlying concern here is > that “utilization” is being interpreted as two different things, one of > which is not meaningfully represented for SSDs by that graph that was > shared. > > > > > > On Jul 2, 2025, at 11:59 AM, Peter Eisch <pe...@boku.net> wrote: > > Would you consider increasing the pgs on your larger pool(s)? You might > find things balance better. Then you could look at weighting to favor NVMe > if needed. > > > _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io