My Ceph cluster(s) usually run on openSUSE, in this case Leap 15.6. The host directory /var/lib/ceph/crash/posted is created by the 'ceph-base' package, so I removed that (this is a test machine for all kinds of purposes). I redeployed with cephadm (I don't use FQDN) and it creates the directory /var/lib/ceph/{FSID}/crash/posted for me. I don't see a warning about this directory. So just to clarify, is it complaining about the host directory /var/lib/ceph/crash/posted not being present or the "mapped" directory within the ceph-crash container?

There have been a couple of releases which don't handle crash well in general. I haven't look too deep into the latest though.

As for the cephadm shell, what das 'cephadm ls --no-detail | grep crash' show you? If you deployed with FQDN, I'd expect that crash is also deployed with FQDN. But $HOSTNAME could be the shortname, so you probably should use a different approach, for example 'cephadm shell --name crash.$(hostname -f) or something similar.

Zitat von lejeczek <pelj...@yahoo.co.uk>:

no 'ceph-crush' installed as far as I can tell.

-> $ rpm -qa \*ceph\*
centos-release-ceph-squid-1.0-1.el9s.noarch
python3-ceph-argparse-19.2.3-1.el9s.x86_64
libcephfs2-19.2.3-1.el9s.x86_64
python3-cephfs-19.2.3-1.el9s.x86_64
python3-ceph-common-19.2.3-1.el9s.x86_64
ceph-common-19.2.3-1.el9s.x86_64
cephadm-19.2.3-1.el9s.noarch

cluster was deployed with
-> $ cephadm bootstrap ...

What also is "interesting" is:
-> $ cephadm shell --name crash.$HOSTNAME
Inferring fsid 9f4f9dba-72c7-11f0-8052-525400519d29
Inferring config /var/lib/ceph/9f4f9dba-72c7-11f0-8052-525400519d29/crash.podster1.mine.priv/config Using ceph image with id 'aade1b12b8e6' and tag 'v19' created on 2025-07-17 19:53:27 +0000 UTC
quay.io/ceph/ceph@sha256:af0c5903e901e329adabe219dfc8d0c3efc1f05102a753902f33ee16c26b6cee
Error: statfs /var/lib/ceph/9f4f9dba-72c7-11f0-8052-525400519d29/crash.podster1.mine.priv/keyring: no such file or directory

and that is for all nodes.
But I find:
/var/lib/ceph/9f4f9dba-72c7-11f0-8052-525400519d29/crash/posted/
on all/each node - which path _was_ created by the cluster (as opposed to human)
also, I find:
/var/lib/ceph/9f4f9dba-72c7-11f0-8052-525400519d29/crash.podster3/keyring
also on each node, which seems to contain other configs - as 'shell --name ...' above, was looking for.

No, failing mon I thought I should have mentioned it only as trace of originating issue - but I did not think of causality. Is this some hostnames / FQDN misconfiguration ? but if so, then that was/is cluster/cephadm on its own.
OS resolver check /etc/hosts - DNS also involved later - which contains:
-> $ cat /etc/hosts
127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1         localhost localhost.localdomain localhost6 localhost6.localdomain6

10.1.1.61   podster1.mine.priv podster1
10.1.1.62   podster2.mine.priv podster2
10.1.1.63   podster3.mine.priv podster3

I added hosts to the cluster - and 'bootstraped' -  with their FQDNs.

thanks, L.
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to