Hi,

I attached to the osd container itself and found that ceph.conf there is still wrong.

A test cluster is at the moment difficult because most of the utility nodes are still in transition, we wanted to get data back up and running first.

# minimal ceph.conf for 10489760
[global]
        fsid = 10489760
        mon_host = [v2:129.217.31.171:3300/0,v1:129.217.31.171:6789/0] [v2:129.217.31.172:3300/0,v1:129.217.31.172:6789/0] [v2:129.217.31.175:3300/0,v1:129.217.31.175:6789/0]


Should I go manually through all osd container and change the entries in the conf?


Cheers
Dominik


Am 19.12.2022 um 15:49 schrieb Dominik Baack:
Hi,

ceph.conf on the host as well as in "cephadm shell" are similar and have the correct IPs set. Entries for osd are not present.

# minimal ceph.conf for 10489760
[global]
        fsid = 10489760
        mon_host = [v2:129.217.31.186:3300/0,v1:129.217.31.186:6789/0] [v2:129.217.31.188:3300/0,v1:129.217.31.188:6789/0] [v2:129.217.31.189:3300/0,v1:129.217.31.189:6789/0] [v2:129.217.31.190:3300/0,v1:129.217.31.190:6789/0]


Ceph health detail does not help much either, only that there is an "sn05" artifact still present somewhere that should be removed:


Cheers
Dominik


HEALTH_WARN 1 filesystem is degraded; 1 MDSs report slow metadata IOs; 16 osds down; 3 hosts (18 osds) down; Reduced data availability: 545 pgs inactive
[WRN] FS_DEGRADED: 1 filesystem is degraded
    fs ml2r_storage is degraded
[WRN] MDS_SLOW_METADATA_IO: 1 MDSs report slow metadata IOs
    mds.ml2r_storage.ml2rsn06.clizsn(mds.0): 4 slow metadata IOs are blocked > 30 secs, oldest blocked for 4951 secs
[WRN] OSD_DOWN: 16 osds down
    osd.28 (root=default,host=ml2rsn05) is down
    osd.29 (root=default,host=ml2rsn05) is down
    osd.30 (root=default,host=ml2rsn05) is down
    osd.31 (root=default,host=ml2rsn05) is down
    osd.32 (root=default,host=ml2rsn05) is down
    osd.33 (root=default,host=ml2rsn05) is down
    osd.34 (root=default,host=ml2rsn05) is down
    osd.35 (root=default,host=ml2rsn05) is down
    osd.36 (root=default,host=ml2rsn06) is down
    osd.37 (root=default,host=ml2rsn06) is down
    osd.38 (root=default,host=ml2rsn06) is down
    osd.39 (root=default,host=ml2rsn06) is down
    osd.40 (root=default,host=ml2rsn06) is down
    osd.41 (root=default,host=ml2rsn06) is down
    osd.42 (root=default,host=ml2rsn06) is down
    osd.43 (root=default,host=ml2rsn06) is down
[WRN] OSD_HOST_DOWN: 3 hosts (18 osds) down
    host ml2rsn03 (root=default) (9 osds) is down
    host ml2rsn05 (root=default) (9 osds) is down
    host sn05 (root=default) (0 osds) is down
[WRN] PG_AVAILABILITY: Reduced data availability: 545 pgs inactive


Am 19.12.2022 um 14:02 schrieb Eugen Block:
I re-read a thread from last year [1] but didn't find anything extraordinary required for this to work. If your MONs already run with the new addresses, does the ceph.conf on the OSD nodes still reflect the old entries? The docs do mention it explicitly:

Ceph clients and other Ceph daemons use ceph.conf to discover monitors.

In that case a restart of ceph.target wouldn't be enough. I assume cephadm should update the conf file but maybe it requires a 'ceph orch reconfigure osd'? I haven't tried that yet, so be careful and test the effects in a test cluster.

Regards,
Eugen

[1] https://lists.ceph.io/hyperkitty/list/[email protected]/thread/OWSIXB5O3Y7LXC3D2JYETEAFMRC3K7OY/

Zitat von Dominik Baack <[email protected]>:

Hi,

min_mon_release 17 (quincy)
election_strategy: 1
0: [v2:129.217.31.189:3300/0,v1:129.217.31.189:6789/0] mon.ml2rsn06
1: [v2:129.217.31.186:3300/0,v1:129.217.31.186:6789/0] mon.ml2rsn03
2: [v2:129.217.31.188:3300/0,v1:129.217.31.188:6789/0] mon.ml2rsn05
3: [v2:129.217.31.190:3300/0,v1:129.217.31.190:6789/0] mon.ml2rsn07
dumped monmap epoch 21

seems fine for me (with the exception that there is an even number of mons).

ceph osd metadata 18

 "back_addr": "[v2:129.217.31.173:6817/1187092418,v1:129.217.31.173:6819/1187092418]",


Still shows an old address after restarting of ceph.target.


Cheers
Dominik

Am 19.12.2022 um 09:24 schrieb Eugen Block:
If you did it "the right way" then 'ceph mon dump' should reflect the changes. Does that output show the new IP addresses?


Zitat von Dominik Baack <[email protected]>:

Hi,

I removed/added the new mons as explained in:

https://docs.ceph.com/en/latest/rados/operations/add-or-rm-mons/#changing-a-monitor-s-ip-address-the-right-way but did not change the mon map manually.


I can retry the manual way as well and report afterwards:

https://docs.ceph.com/en/latest/rados/operations/add-or-rm-mons/#changing-a-monitor-s-ip-address-the-messy-way Cheers
Dominik Baack


Am 17.12.2022 um 10:06 schrieb Eugen Block:
Hi,

did you also change the monmap as described in the docs [1]? There have been multiple threads in this list with the same topic. Simply changing the IP address of the MONs is not sufficient, but after fixing the monmap the OSDs should connect successfully.

[1] https://docs.ceph.com/en/quincy/rados/operations/add-or-rm-mons.html

Zitat von Dominik Baack <[email protected]>:

Hi,

we have switched our network from IP over IB to an Ethernet configuration.

I could move the mons over to the new address range and they connect into the cluster network. The OSD create more of a problem, even after setting

public_network129.217.31.176/29
cluster_network 129.217.31.184/29

to their values and restarting ceph.target on all nodes but they still report the wrong information when checked with ceph osd metadata :

back_addr
front_addr
hb_back_addr
hb_front_addr

and are therefore not able to connect.

Where do I need to modify the address to get them connecting again?

Cheers
Dominik

_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]



_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]



_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]



_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to