Eugen,

Thanks for the hint. Here is the osd_remove_queue:

[root@ijc-mon1 ~]# ceph config-key get mgr/cephadm/osd_remove_queue|jq
[
  {
    "osd_id": 253,
    "started": true,
    "draining": false,
    "stopped": false,
    "replace": true,
    "force": false,
    "zap": true,
    "hostname": "dig-osd4",
    "drain_started_at": null,
    "drain_stopped_at": null,
    "drain_done_at": "2025-04-15T14:09:30.521534Z",
    "process_started_at": "2025-04-15T14:09:14.091592Z"
  },
  {
    "osd_id": 381,
    "started": true,
    "draining": false,
    "stopped": false,
    "replace": true,
    "force": false,
    "zap": false,
    "hostname": "dig-osd6",
    "drain_started_at": "2025-04-23T11:56:09.864724Z",
    "drain_stopped_at": null,
    "drain_done_at": "2025-04-25T06:53:03.678729Z",
    "process_started_at": "2025-04-23T11:56:05.924923Z"
  }
]

It is not empty the two stray daemons are listed. Not sure it these entries are expected as I specified --replace... A similar issue was reported in https://tracker.ceph.com/issues/67018 so before Reef but the cause may be different. Still not clear for me how to get out of this, except may be replacing the OSDs but this will take some time...

Best regards,

Michel

Le 27/04/2025 à 10:21, Eugen Block a écrit :
Hi,

what's the current ceph status? Wasn't there a bug in early Reef versions preventing upgrades if there were removed OSDs in the queue? But IIRC, the cephadm module would crash. Can you check

ceph config-key get mgr/cephadm/osd_remove_queue

And then I would check the mgr log, maybe set it to a higher debug level to see what's blocking it.

Zitat von Michel Jouvin <michel.jou...@ijclab.in2p3.fr>:

Hi,

I tried to restart all the mgrs (we have 3, 1 active, 2 standby) by executing 3 times the `ceph mgr fail`, no impact. I don't really understand why I get these stray daemons after doing a 'ceph orch osd rm --replace` but I think I have always seen this. I tried to mute rather than disable the stray daemon check but it doesn't help either. And I find strange this message every 10s about one of the destroyed OSD and only one, reporting it is down and already destroyed and saying it'll zap it (I think I didn't add --zap when I removed it as the underlying disk is dead).

I'm completely stuck with this upgrade and I don't remember having this kind of problems in previous upgrades with cephadm... Any idea where to look for the cause and/or how to fix it?

Best regards,

Michel

Le 24/04/2025 à 23:34, Michel Jouvin a écrit :
Hi,

I'm trying to upgrade a (cephadm) cluster from 18.2.2 to 18.2.6, using 'ceph orch upgrade'. When I enter the command 'ceph orch upgrade start --ceph-version 18.2.6', I receive a message saying that the upgrade has been initiated, with a similar message in the logs but nothing happens after this. 'ceph orch upgrade status' says:

-------

[root@ijc-mon1 ~]# ceph orch upgrade status
{
    "target_image": "quay.io/ceph/ceph:v18.2.6",
    "in_progress": true,
    "which": "Upgrading all daemon types on all hosts",
    "services_complete": [],
    "progress": "",
    "message": "",
    "is_paused": false
}
-------

The first time I entered the command, the cluster status was HEALTH_WARN because of 2 stray daemons (caused by destroyed OSDs, rm --replace). I set mgr/cephadm/warn_on_stray_daemons to false to ignore these 2 daemons, the cluster is now HEALTH_OK but it doesn't help. Following a Red Hat KB entry, I tried to failover the mgr, stopped an restarted the upgrade but without any improvement. I have not seen anything in the logs, except that there is an INF entry every 10s about the destroyed OSD saying:

------

2025-04-24T21:30:54.161988+0000 mgr.ijc-mon1.yyfnhz (mgr.55376028) 14079 : cephadm [INF] osd.253 now down 2025-04-24T21:30:54.162601+0000 mgr.ijc-mon1.yyfnhz (mgr.55376028) 14080 : cephadm [INF] Daemon osd.253 on dig-osd4 was already removed 2025-04-24T21:30:54.164440+0000 mgr.ijc-mon1.yyfnhz (mgr.55376028) 14081 : cephadm [INF] Successfully destroyed old osd.253 on dig-osd4; ready for replacement 2025-04-24T21:30:54.164536+0000 mgr.ijc-mon1.yyfnhz (mgr.55376028) 14082 : cephadm [INF] Zapping devices for osd.253 on dig-osd4
-----

The message seems to be only for one of the 2 destroyed OSDs since I restarted the mgr. May this be the cause for the stucked upgrade? What can I do for fixing this?

Thanks in advance for any hint. Best regards,

Michel

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to