The other hosts are still online and the cluster only lost 1/3 of its services.
> Am 16.09.2023 um 12:53 schrieb Eugen Block <ebl...@nde.ag>: > > I don’t have time to look into all the details, but I’m wondering how you > seem to be able to start mgr services with the orchestrator if all mgr > daemons are down. The orchestrator is a mgr module, so that’s a bit weird, > isn’t it? > > Zitat von Boris Behrens <b...@kervyn.de>: > >> Hi Eugen, >> the test-test cluster where we started with simple ceph and the adoption >> when straight forward are working fine. >> >> But this test cluster was all over the place. >> We had an old running update via orchestrator which was still in the >> pipeline, the adoption process was stopped a year ago and now got picked up >> again, and so on and so forth. >> >> But now we have it clean, at least we think it's clean. >> >> After a reboot, the services are not available. I have to start the via >> ceph orch. >> root@0cc47a6df14e:~# systemctl list-units | grep ceph >> ceph-crash.service >> loaded active running Ceph crash dump collector >> ceph-fuse.target >> loaded active active ceph target allowing to start/stop >> all ceph-fuse@.service instances at once >> ceph-mds.target >> loaded active active ceph target allowing to start/stop >> all ceph-mds@.service instances at once >> ceph-mgr.target >> loaded active active ceph target allowing to start/stop >> all ceph-mgr@.service instances at once >> ceph-mon.target >> loaded active active ceph target allowing to start/stop >> all ceph-mon@.service instances at once >> ceph-osd.target >> loaded active active ceph target allowing to start/stop >> all ceph-osd@.service instances at once >> ceph-radosgw.target >> loaded active active ceph target allowing to start/stop >> all ceph-radosgw@.service instances at once >> ceph.target >> loaded active active All Ceph clusters and services >> root@0cc47a6df14e:~# ceph orch start mgr >> Scheduled to start mgr.0cc47a6df14e.nvjlcx on host '0cc47a6df14e' >> Scheduled to start mgr.0cc47a6df330.aznjao on host '0cc47a6df330' >> Scheduled to start mgr.0cc47aad8ce8.ifiydp on host '0cc47aad8ce8' >> root@0cc47a6df14e:~# ceph orch start mon >> Scheduled to start mon.0cc47a6df14e on host '0cc47a6df14e' >> Scheduled to start mon.0cc47a6df330 on host '0cc47a6df330' >> Scheduled to start mon.0cc47aad8ce8 on host '0cc47aad8ce8' >> root@0cc47a6df14e:~# ceph orch start osd.all-flash-over-1tb >> Scheduled to start osd.2 on host '0cc47a6df14e' >> Scheduled to start osd.5 on host '0cc47a6df14e' >> Scheduled to start osd.3 on host '0cc47a6df330' >> Scheduled to start osd.0 on host '0cc47a6df330' >> Scheduled to start osd.4 on host '0cc47aad8ce8' >> Scheduled to start osd.1 on host '0cc47aad8ce8' >> root@0cc47a6df14e:~# systemctl list-units | grep ceph >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@mgr.0cc47a6df14e.nvjlcx.service >> loaded active running Ceph >> mgr.0cc47a6df14e.nvjlcx for 03977a23-f00f-4bb0-b9a7-de57f40ba853 >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@mon.0cc47a6df14e.service >> loaded active running Ceph >> mon.0cc47a6df14e for 03977a23-f00f-4bb0-b9a7-de57f40ba853 >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@osd.2.service >> loaded active running Ceph osd.2 >> for 03977a23-f00f-4bb0-b9a7-de57f40ba853 >> ceph-crash.service >> loaded active running Ceph crash >> dump collector >> system-ceph\x2d03977a23\x2df00f\x2d4bb0\x2db9a7\x2dde57f40ba853.slice >> loaded active active >> system-ceph\x2d03977a23\x2df00f\x2d4bb0\x2db9a7\x2dde57f40ba853.slice >> ceph-fuse.target >> loaded active active ceph target >> allowing to start/stop all ceph-fuse@.service instances at once >> ceph-mds.target >> loaded active active ceph target >> allowing to start/stop all ceph-mds@.service instances at once >> ceph-mgr.target >> loaded active active ceph target >> allowing to start/stop all ceph-mgr@.service instances at once >> ceph-mon.target >> loaded active active ceph target >> allowing to start/stop all ceph-mon@.service instances at once >> ceph-osd.target >> loaded active active ceph target >> allowing to start/stop all ceph-osd@.service instances at once >> ceph-radosgw.target >> loaded active active ceph target >> allowing to start/stop all ceph-radosgw@.service instances at once >> ceph.target >> loaded active active All Ceph >> clusters and services >> root@0cc47a6df14e:~# systemctl status >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@mgr.0cc47a6df14e.nvjlcx.service >> ● ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@mgr.0cc47a6df14e.nvjlcx.service >> - Ceph mgr.0cc47a6df14e.nvjlcx for 03977a23-f00f-4bb0-b9a7-de57f40ba853 >> Loaded: loaded >> (/etc/systemd/system/ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@.service; >> enabled; vendor preset: enabled) >> Active: active (running) since Sat 2023-09-16 09:18:53 UTC; 51s ago >> Process: 4828 ExecStartPre=/bin/rm -f >> /run/ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@mgr.0cc47a6df14e.nvjlcx.service-pid >> /run/ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@mgr.0cc47a6df1> >> Process: 4829 ExecStart=/bin/bash >> /var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/mgr.0cc47a6df14e.nvjlcx/unit.run >> (code=exited, status=0/SUCCESS) >> Main PID: 5132 (conmon) >> Tasks: 36 (limit: 309227) >> Memory: 512.0M >> CGroup: >> /system.slice/system-ceph\x2d03977a23\x2df00f\x2d4bb0\x2db9a7\x2dde57f40ba853.slice/ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@mgr.0cc47a6df14e.nvjlcx.service >> ├─container >> │ ├─5136 /dev/init -- /usr/bin/ceph-mgr -n >> mgr.0cc47a6df14e.nvjlcx -f --setuser ceph --setgroup ceph >> --default-log-to-file=false --default-log-to-journald=true --default-log> >> │ └─5139 /usr/bin/ceph-mgr -n mgr.0cc47a6df14e.nvjlcx -f >> --setuser ceph --setgroup ceph --default-log-to-file=false >> --default-log-to-journald=true --default-log-to-stderr=fa> >> └─supervisor >> └─5132 /usr/libexec/podman/conmon --api-version 1 -c >> 0165b4f78867ad284cc65fbece46013e6547a2f3ecf99cc7ffb8b720f705ee66 -u >> 0165b4f78867ad284cc65fbece46013e6547a2f3ecf99cc7ff> >> >> Sep 16 09:19:04 0cc47a6df14e.f00f.gridscale.dev >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-mgr-0cc47a6df14e-nvjlcx[5132]: >> 2023-09-16T09:19:04.333+0000 7f4fcc0a91c0 -1 mgr[py] Module alert> >> Sep 16 09:19:04 0cc47a6df14e.f00f.gridscale.dev >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-mgr-0cc47a6df14e-nvjlcx[5132]: >> 2023-09-16T09:19:04.501+0000 7f4fcc0a91c0 -1 mgr[py] Module iosta> >> Sep 16 09:19:05 0cc47a6df14e.f00f.gridscale.dev >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-mgr-0cc47a6df14e-nvjlcx[5132]: >> 2023-09-16T09:19:05.249+0000 7f4fcc0a91c0 -1 mgr[py] Module orche> >> Sep 16 09:19:05 0cc47a6df14e.f00f.gridscale.dev >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-mgr-0cc47a6df14e-nvjlcx[5132]: >> 2023-09-16T09:19:05.481+0000 7f4fcc0a91c0 -1 mgr[py] Module rbd_s> >> Sep 16 09:19:06 0cc47a6df14e.f00f.gridscale.dev >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-mgr-0cc47a6df14e-nvjlcx[5132]: >> [16/Sep/2023:09:19:06] ENGINE Bus STARTING >> Sep 16 09:19:06 0cc47a6df14e.f00f.gridscale.dev >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-mgr-0cc47a6df14e-nvjlcx[5132]: >> CherryPy Checker: >> Sep 16 09:19:06 0cc47a6df14e.f00f.gridscale.dev >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-mgr-0cc47a6df14e-nvjlcx[5132]: >> The Application mounted at '' has an empty config. >> Sep 16 09:19:06 0cc47a6df14e.f00f.gridscale.dev >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-mgr-0cc47a6df14e-nvjlcx[5132]: >> Sep 16 09:19:06 0cc47a6df14e.f00f.gridscale.dev >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-mgr-0cc47a6df14e-nvjlcx[5132]: >> [16/Sep/2023:09:19:06] ENGINE Serving on http://:::9283 >> Sep 16 09:19:06 0cc47a6df14e.f00f.gridscale.dev >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-mgr-0cc47a6df14e-nvjlcx[5132]: >> [16/Sep/2023:09:19:06] ENGINE Bus STARTED >> >> This seems to be the cephadm log: >> >> cephadm ['adopt', '--style', 'legacy', '--name', 'osd.3'] >> 2023-09-15 11:32:44,290 7fef7b041740 INFO Pulling container image >> quay.io/ceph/ceph:v17... >> 2023-09-15 11:32:47,128 7fef7b041740 INFO Found online OSD at >> //var/lib/ceph/osd/ceph-3/fsid >> 2023-09-15 11:32:47,129 7fef7b041740 INFO objectstore_type is bluestore >> 2023-09-15 11:32:47,150 7fef7b041740 INFO Stopping old systemd unit >> ceph-osd@3... >> 2023-09-15 11:32:48,560 7fef7b041740 INFO Disabling old systemd unit >> ceph-osd@3... >> 2023-09-15 11:32:49,157 7fef7b041740 INFO Moving data... >> 2023-09-15 11:32:49,158 7fef7b041740 DEBUG move file >> '//var/lib/ceph/osd/ceph-3/require_osd_release' -> >> '/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/require_osd_release' >> 2023-09-15 11:32:49,158 7fef7b041740 DEBUG chown 167:167 >> `/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/require_osd_release` >> 2023-09-15 11:32:49,158 7fef7b041740 DEBUG symlink >> '/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/block' -> >> '/dev/ceph-66d3bb27-cd5c-4897-aa76-684bc46d1c8b/osd-block-4bfc2101-e9b2-468d-8f54-a05f080ebdfe' >> 2023-09-15 11:32:49,158 7fef7b041740 DEBUG move file >> '//var/lib/ceph/osd/ceph-3/ready' -> >> '/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/ready' >> 2023-09-15 11:32:49,159 7fef7b041740 DEBUG chown 167:167 >> `/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/ready` >> 2023-09-15 11:32:49,159 7fef7b041740 DEBUG move file >> '//var/lib/ceph/osd/ceph-3/type' -> >> '/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/type' >> 2023-09-15 11:32:49,159 7fef7b041740 DEBUG chown 167:167 >> `/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/type` >> 2023-09-15 11:32:49,159 7fef7b041740 DEBUG move file >> '//var/lib/ceph/osd/ceph-3/fsid' -> >> '/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/fsid' >> 2023-09-15 11:32:49,159 7fef7b041740 DEBUG chown 167:167 >> `/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/fsid` >> 2023-09-15 11:32:49,160 7fef7b041740 DEBUG move file >> '//var/lib/ceph/osd/ceph-3/ceph_fsid' -> >> '/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/ceph_fsid' >> 2023-09-15 11:32:49,160 7fef7b041740 DEBUG chown 167:167 >> `/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/ceph_fsid` >> 2023-09-15 11:32:49,160 7fef7b041740 DEBUG move file >> '//var/lib/ceph/osd/ceph-3/keyring' -> >> '/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/keyring' >> 2023-09-15 11:32:49,160 7fef7b041740 DEBUG chown 167:167 >> `/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/keyring` >> 2023-09-15 11:32:49,160 7fef7b041740 DEBUG move file >> '//var/lib/ceph/osd/ceph-3/whoami' -> >> '/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/whoami' >> 2023-09-15 11:32:49,161 7fef7b041740 DEBUG chown 167:167 >> `/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/whoami` >> 2023-09-15 11:32:49,161 7fef7b041740 DEBUG Remove dir >> `//var/lib/ceph/osd/ceph-3` >> 2023-09-15 11:32:49,166 7fef7b041740 INFO Chowning content... >> 2023-09-15 11:32:49,171 7fef7b041740 DEBUG chown: stdout changed ownership >> of '/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/block' from >> root:root to 167:167 >> 2023-09-15 11:32:49,172 7fef7b041740 INFO Chowning >> /var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/block... >> 2023-09-15 11:32:49,172 7fef7b041740 INFO Disabling host unit ceph-volume@ >> lvm unit... >> 2023-09-15 11:32:49,649 7fef7b041740 DEBUG systemctl: stderr Removed >> /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-3-4bfc2101-e9b2-468d-8f54-a05f080ebdfe.service. >> 2023-09-15 11:32:49,650 7fef7b041740 DEBUG copy file `//etc/ceph/ceph.conf` >> -> `/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/config` >> 2023-09-15 11:32:49,650 7fef7b041740 DEBUG chown 167:167 >> `/var/lib/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/osd.3/config` >> 2023-09-15 11:32:49,650 7fef7b041740 INFO Moving logs... >> 2023-09-15 11:32:49,651 7fef7b041740 DEBUG move file >> '//var/log/ceph/ceph-osd.3.log' -> >> '/var/log/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/ceph-osd.3.log' >> 2023-09-15 11:32:49,651 7fef7b041740 DEBUG chown 167:167 >> `/var/log/ceph/03977a23-f00f-4bb0-b9a7-de57f40ba853/ceph-osd.3.log` >> 2023-09-15 11:32:49,651 7fef7b041740 INFO Creating new units... >> 2023-09-15 11:32:50,803 7fef7b041740 DEBUG sysctl: stdout * Applying >> /etc/sysctl.d/10-console-messages.conf ... >> 2023-09-15 11:32:50,803 7fef7b041740 DEBUG sysctl: stdout kernel.printk = 4 >> 4 1 7 >> 2023-09-15 11:32:50,803 7fef7b041740 DEBUG sysctl: stdout * Applying >> /etc/sysctl.d/10-ipv6-privacy.conf ... >> 2023-09-15 11:32:50,803 7fef7b041740 DEBUG sysctl: stdout >> net.ipv6.conf.all.use_tempaddr = 2 >> 2023-09-15 11:32:50,803 7fef7b041740 DEBUG sysctl: stdout >> net.ipv6.conf.default.use_tempaddr = 2 >> 2023-09-15 11:32:50,803 7fef7b041740 DEBUG sysctl: stdout * Applying >> /etc/sysctl.d/10-kernel-hardening.conf ... >> 2023-09-15 11:32:50,803 7fef7b041740 DEBUG sysctl: stdout >> kernel.kptr_restrict = 1 >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout * Applying >> /etc/sysctl.d/10-link-restrictions.conf ... >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout >> fs.protected_hardlinks = 1 >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout >> fs.protected_symlinks = 1 >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout * Applying >> /etc/sysctl.d/10-magic-sysrq.conf ... >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout kernel.sysrq = 176 >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout * Applying >> /etc/sysctl.d/10-network-security.conf ... >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout >> net.ipv4.conf.default.rp_filter = 2 >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout >> net.ipv4.conf.all.rp_filter = 2 >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout * Applying >> /etc/sysctl.d/10-ptrace.conf ... >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout >> kernel.yama.ptrace_scope = 1 >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout * Applying >> /etc/sysctl.d/10-zeropage.conf ... >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout vm.mmap_min_addr >> = 65536 >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout * Applying >> /etc/sysctl.d/30-ceph-osd.conf ... >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout fs.aio-max-nr = >> 1048576 >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout kernel.pid_max = >> 4194304 >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout * Applying >> /usr/lib/sysctl.d/50-coredump.conf ... >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout >> kernel.core_pattern = |/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c >> %h >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout * Applying >> /usr/lib/sysctl.d/50-default.conf ... >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout >> net.ipv4.conf.default.promote_secondaries = 1 >> 2023-09-15 11:32:50,804 7fef7b041740 DEBUG sysctl: stdout >> net.ipv4.ping_group_range = 0 2147483647 >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout >> net.core.default_qdisc = fq_codel >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout >> fs.protected_regular = 1 >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout >> fs.protected_fifos = 1 >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout * Applying >> /usr/lib/sysctl.d/50-pid-max.conf ... >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout kernel.pid_max = >> 4194304 >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout * Applying >> /etc/sysctl.d/90-ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853-osd.conf ... >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout fs.aio-max-nr = >> 1048576 >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout kernel.pid_max = >> 4194304 >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout * Applying >> /etc/sysctl.d/99-sysctl.conf ... >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout * Applying >> /usr/lib/sysctl.d/protect-links.conf ... >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout >> fs.protected_fifos = 1 >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout >> fs.protected_hardlinks = 1 >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout >> fs.protected_regular = 2 >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout >> fs.protected_symlinks = 1 >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stdout * Applying >> /etc/sysctl.conf ... >> 2023-09-15 11:32:50,805 7fef7b041740 DEBUG sysctl: stderr sysctl: setting >> key "net.ipv4.conf.all.promote_secondaries": Invalid argument >> 2023-09-15 11:32:51,469 7fef7b041740 DEBUG Non-zero exit code 1 from >> systemctl reset-failed ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@osd.3 >> 2023-09-15 11:32:51,469 7fef7b041740 DEBUG systemctl: stderr Failed to >> reset failed state of unit >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@osd.3.service: Unit >> ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@osd.3.service not loaded. >> 2023-09-15 11:32:51,954 7fef7b041740 DEBUG systemctl: stderr Created >> symlink >> /etc/systemd/system/ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853.target.wants/ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@osd.3.service >> → /etc/systemd/system/ceph-03977a23-f00f-4bb0-b9a7-de57f40ba853@.service. >> 2023-09-15 11:32:54,331 7fef7b041740 DEBUG firewalld does not appear to be >> present >> >>> Am Sa., 16. Sept. 2023 um 10:25 Uhr schrieb Eugen Block <ebl...@nde.ag>: >>> >>> That sounds a bit strange to me, because all clusters we adopted so >>> far successfully converted the previous systemd-units into systemd >>> units targeting the pods. This process also should have been logged >>> (stdout, probably in the cephadm.log as well), resulting in "enabled" >>> systemd units. Can you paste the output of 'systemctl status >>> ceph-<FSID>@mon.<MON>'? If you have it, please also share the logs >>> from the adoption process. >>> What I did notice in a test cluster a while ago was that I had to >>> reboot a node where I had to "play around" a bit with removed and >>> redeployed osd containers. At some point they didn't react to >>> systemctl commands anymore, but a reboot fixed that. But I haven't >>> seen that in a production cluster yet, so some more details would be >>> useful. >>> >>> Zitat von Boris Behrens <b...@kervyn.de>: >>> >>> > Hi, >>> > is there a way to have the pods start again after reboot? >>> > Currently I need to start them by hand via ceph orch start >>> mon/mgr/osd/... >>> > >>> > I imagine this will lead to a lot of headache when the ceph cluster gets >>> a >>> > powercycle and the mon pods will not start automatically. >>> > >>> > I've spun up a test cluster and there the pods start very fast. On the >>> > legacy test cluster, which got adopted to cephadm, it does not. >>> > >>> > Cheers >>> > Boris >>> > _______________________________________________ >>> > ceph-users mailing list -- ceph-users@ceph.io >>> > To unsubscribe send an email to ceph-users-le...@ceph.io >>> >>> >>> _______________________________________________ >>> ceph-users mailing list -- ceph-users@ceph.io >>> To unsubscribe send an email to ceph-users-le...@ceph.io >>> >> >> >> -- >> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im >> groüen Saal. > > > _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io