Yes it is on:
# ceph balancer status
{
"active": true,
"last_optimize_duration": "0:00:00.001867",
"last_optimize_started": "Mon Nov 22 13:10:24 2021",
"mode": "upmap",
"optimize_result": "Unable to find further optimization, or pool(s)
pg_num is decreasing, or distribution is already perfect",
"plans": []
}
On Mon, Nov 22, 2021 at 10:17 AM Stefan Kooman <[email protected]> wrote:
> On 11/22/21 08:12, David Tinker wrote:
> > I set osd.7 as "in", uncordened the node, scaled the OSD deployment back
> > up and things are recovering with cluster status HEALTH_OK.
> >
> > I found this message from the archives:
> > https://www.mail-archive.com/[email protected]/msg47071.html
> > <https://www.mail-archive.com/[email protected]/msg47071.html>
> >
> > "You have a large difference in the capacities of the nodes. This
> > resultsin a different host weight, which in turn might lead to problems
> > withthe crush algorithm. It is not able to get three different hosts for
> > OSDplacement for some of the PGs.
> >
> > CEPH and crush do not cope well with heterogenous setups. I wouldsuggest
> > to move one of the OSDs from host ceph1 to ceph4 to equalize thehost
> > weight."
> >
> > My nodes do have very different weights. What I am trying to do is
> > re-install each node in the cluster so they all have the same amount of
> > space for Ceph (much less than before .. we need more space for hostpath
> > stuff).
> >
> > # ceph osd tree
> > ID CLASS WEIGHT TYPE NAME STATUS
> REWEIGHT PRI-AFF
> > -1 13.77573 root default
> > -5 13.77573 region FSN1
> > -22 0.73419 zone FSN1-DC13
> > -21 0 host node5-redacted-com
> > -27 0.73419 host node7-redacted-com
> > 1 ssd 0.36710 osd.1 up
> 1.00000 1.00000
> > 5 ssd 0.36710 osd.5 up
> 1.00000 1.00000
> > -10 6.20297 zone FSN1-DC14
> > -9 6.20297 host node3-redacted-com
> > 2 ssd*3.10149* osd.2 up
> 1.00000 1.00000
> > 4 ssd*3.10149* osd.4 up
> 1.00000 1.00000
> > -18 3.19919 zone FSN1-DC15
> > -17*3.19919* host node4-redacted-com
> > 7 ssd*3.19919* osd.7 down
> 0 1.00000
> > -4 2.90518 zone FSN1-DC16
> > -3 2.90518 host node1-redacted-com
> > 0 ssd*1.45259* osd.0 up
> 1.00000 1.00000
> > 3 ssd*1.45259* osd.3 up
> 1.00000 1.00000
> > -14 0.73419 zone FSN1-DC18
> > -13 0 host node2-redacted-com
> > -25 0.73419 host node6-redacted-com
> > 10 ssd 0.36710 osd.10 up
> 1.00000 1.00000
> > 11 ssd 0.36710 osd.11 up
> 1.00000 1.00000
> >
> >
> > Should I just change the weights before/after removing OSD 7?
> >
> > With something like "ceph osd crush reweight osd.7 1.0"?
>
> The ceph balancer is there to balance PGs across all nodes. Do you have
> it enabled?
>
> ceph balancer status
>
> The most efficient way is to use mode upmap (should work with modern
> clients):
>
> ceph balancer mode upmap
>
> Gr. Stefn
>
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]