HI Guys, Good Morning! I have re-deployed the ceph cluster using cephadm because of some issues with osds. Everything is working fine now with the new cluster except that Cephadm is still trying to deploy node_exporter daemon on one of the monitor hosts using the old container which no longer exists.
node_exporter is running fine on that host in the new container. But ceph is still trying to deploy it on that host using an old container and since it couldn't find that podman container it throws the warning. I did try to remove that node_exporter service on that node and added it again but no luck. When I reboot that host after its coming ceph status shows "Health OK" but after sometime it starts showing same waring ceph health detail HEALTH_WARN Failed to place 1 daemon(s) [WRN] CEPHADM_DAEMON_PLACE_FAIL: Failed to place 1 daemon(s) Failed while placing node-exporter.ceph-mon2 on ceph-mon2: cephadm exited with an error code: 1, stderr: Non-zero exit code 125 from /usr/bin/podman container inspect --format {{.State.Status}} ceph-1aa23d18-6eb0-11f0-823c-cc28aa448c14-node-exporter-ceph-mon2 /usr/bin/podman: stderr Error: no such container ceph-1aa23d18-6eb0-11f0-823c-cc28aa448c14-node-exporter-ceph-mon2 Non-zero exit code 125 from /usr/bin/podman container inspect --format {{.State.Status}} ceph-1aa23d18-6eb0-11f0-823c-cc28aa448c14-node-exporter.ceph-mon2 /usr/bin/podman: stderr Error: no such container ceph-1aa23d18-6eb0-11f0-823c-cc28aa448c14-node-exporter.ceph-mon2 Deploy daemon node-exporter.ceph-mon2 ... Verifying port 0.0.0.0:9100 ... Cannot bind to IP 0.0.0.0 port 9100: [Errno 98] Address already in use ERROR: TCP Port(s) '0.0.0.0:9100' required for node-exporter already in use Please let me know how to fix it. Thanks, Gagan _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io