HI Guys,
                    Good Morning!

I have re-deployed the ceph cluster using cephadm because of some issues
with osds.
Everything is working fine now with the new cluster except that Cephadm is
still trying to deploy  node_exporter daemon on one of the monitor hosts
using the old container which no longer exists.

node_exporter is running fine on that host in the new container.

But ceph is still trying to deploy it on that host using an old container
and since it couldn't find that podman container it throws the warning.

I did try to remove that node_exporter service on that node and added it
again but no luck.  When I reboot that host after its coming ceph status
shows "Health OK"  but after sometime it starts showing same waring

ceph health detail
HEALTH_WARN Failed to place 1 daemon(s)
[WRN] CEPHADM_DAEMON_PLACE_FAIL: Failed to place 1 daemon(s)
    Failed while placing node-exporter.ceph-mon2 on ceph-mon2: cephadm
exited with an error code: 1, stderr: Non-zero exit code 125 from
/usr/bin/podman container inspect --format {{.State.Status}}
ceph-1aa23d18-6eb0-11f0-823c-cc28aa448c14-node-exporter-ceph-mon2
/usr/bin/podman: stderr Error: no such container
ceph-1aa23d18-6eb0-11f0-823c-cc28aa448c14-node-exporter-ceph-mon2
Non-zero exit code 125 from /usr/bin/podman container inspect --format
{{.State.Status}}
ceph-1aa23d18-6eb0-11f0-823c-cc28aa448c14-node-exporter.ceph-mon2
/usr/bin/podman: stderr Error: no such container
ceph-1aa23d18-6eb0-11f0-823c-cc28aa448c14-node-exporter.ceph-mon2
Deploy daemon node-exporter.ceph-mon2 ...
Verifying port 0.0.0.0:9100 ...
Cannot bind to IP 0.0.0.0 port 9100: [Errno 98] Address already in use
ERROR: TCP Port(s) '0.0.0.0:9100' required for node-exporter already in use

Please let me know how to fix it.


Thanks,
Gagan
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to