Hi,
i like to answer to myself :-) I finally found the rest of my documentation...
So after reinstalling the OS also the osd config must be created.
Here is what i have done, maybe this helps someone:
------------------
Get the informations:
```
cephadm ceph-volume lvm list
ceph config generate-minimal-conf
ceph auth get osd.[ID]
```
Now create a minimal osd config:
```
vi osd.[ID].json
```
```
{
"config": "# minimal ceph.conf for
6d0ecf22-9155-4684-971a-2f6cde8628c8\n[global]\n\tfsid =
6d0ecf22-9155-4684-971a-2f6cde8628c8\n\tmon_host =
[v2:192.168.6.21:3300/0,v1:192.168.6.21:6789/0]
[v2:192.168.6.22:3300/0,v1:192.168.6.22:6789/0]
[v2:192.168.6.23:3300/0,v1:192.168.6.23:6789/0]
[v2:192.168.6.24:3300/0,v1:192.168.6.24:6789/0]
[v2:192.168.6.25:3300/0,v1:192.168.6.25:6789/0]\n",
"keyring": "[osd.XXX]\n\tkey = XXXXXXXXXXXXXXXXXXXX\n"
}
```
Deploy the OSD daemon:
```
cephadm deploy --fsid 6d0ecf22-9155-4684-971a-2f6cde8628c8 --osd-fsid [ID]
--name osd.[ID] --config-json osd.[ID].json
```
Yours,
bbk
On Thu, 2021-12-09 at 18:35 +0100, bbk wrote:
> After reading my mail it may not be clear that i reinstalled the OS of
> a node with OSDs.
>
> On Thu, 2021-12-09 at 18:10 +0100, bbk wrote:
> > Hi,
> >
> > the last time i have reinstalled a node with OSDs, i added the disks
> > with the following command. But unfortunatly this time i ran into a
> > error.
> >
> > It seems like this time the command doesn't create the container, i
> > am able to run `cephadm shell`, and other daemons (mon,mgr,mds) are
> > running.
> >
> > I don't know if that is the right way to do it?
> >
> >
> > ~# cephadm deploy --fsid 6d0ecf22-9155-4684-971a-2f6cde8628c8 --osd-
> > fsid 941c6cb6-6898-4aa2-a33a-cec3b6a95cf1 --name osd.9
> >
> > Non-zero exit code 125 from /usr/bin/podman container inspect --
> > format {{.State.Status}} ceph-6d0ecf22-9155-4684-971a-2f6cde8628c8-
> > osd-9
> > /usr/bin/podman: stderr Error: error inspecting object: no such
> > container ceph-6d0ecf22-9155-4684-971a-2f6cde8628c8-osd-9
> > Non-zero exit code 125 from /usr/bin/podman container inspect --
> > format {{.State.Status}} ceph-6d0ecf22-9155-4684-971a-2f6cde8628c8-
> > osd.9
> > /usr/bin/podman: stderr Error: error inspecting object: no such
> > container ceph-6d0ecf22-9155-4684-971a-2f6cde8628c8-osd.9
> > Deploy daemon osd.9 ...
> > Non-zero exit code 1 from systemctl start
> > [email protected]
> > systemctl: stderr Job for
> > [email protected] failed
> > because the control process exited with error code.
> > systemctl: stderr See "systemctl status
> > [email protected]" and
> > "journalctl -xe" for details.
> > Traceback (most recent call last):
> > File "/usr/sbin/cephadm", line 8571, in <module>
> > main()
> > File "/usr/sbin/cephadm", line 8559, in main
> > r = ctx.func(ctx)
> > File "/usr/sbin/cephadm", line 1787, in _default_image
> > return func(ctx)
> > File "/usr/sbin/cephadm", line 4549, in command_deploy
> > ports=daemon_ports)
> > File "/usr/sbin/cephadm", line 2677, in deploy_daemon
> > c, osd_fsid=osd_fsid, ports=ports)
> > File "/usr/sbin/cephadm", line 2906, in deploy_daemon_units
> > call_throws(ctx, ['systemctl', 'start', unit_name])
> > File "/usr/sbin/cephadm", line 1467, in call_throws
> > raise RuntimeError('Failed command: %s' % ' '.join(command))
> > RuntimeError: Failed command: systemctl start
> > [email protected]
> >
> >
> > ~# cephadm ceph-volume lvm list
> >
> > ====== osd.9 =======
> >
> > [block] /dev/ceph-07fa2bb7-628f-40c0-8725-0266926371c0/osd-
> > block-941c6cb6-6898-4aa2-a33a-cec3b6a95cf1
> >
> > block device /dev/ceph-07fa2bb7-628f-40c0-8725-
> > 0266926371c0/osd-block-941c6cb6-6898-4aa2-a33a-cec3b6a95cf1
> > block uuid mVEhfF-LK4E-Dtmb-Jj23-tn8x-lpLy-
> > KiUy1a
> > cephx lockbox secret
> > cluster fsid 6d0ecf22-9155-4684-971a-2f6cde8628c8
> > cluster name ceph
> > crush device class None
> > encrypted 0
> > osd fsid 941c6cb6-6898-4aa2-a33a-cec3b6a95cf1
> > osd id 9
> > type block
> > vdo 0
> > devices /dev/sdd
> >
> >
> > ~# podman --version
> > podman version 3.2.3
> >
> >
> > ~# cephadm version
> > Using recent ceph image
> > quay.io/ceph/ceph@sha256:2f7f0af8663e73a422f797de605e769ae44eb0297f2a
> > 79324739404cc1765728
> > ceph version 16.2.7 (dd0603118f56ab514f133c8d2e3adfc983942503)
> > pacific (stable)
> >
> >
> > ~# lsb_release -a
> > LSB Version: :core-4.1-amd64:core-4.1-noarch
> > Distributor ID: RedHatEnterprise
> > Description: Red Hat Enterprise Linux release 8.5 (Ootpa)
> > Release: 8.5
> > Codename: Ootpa
> >
> >
> > ~# cephadm shell
> > Inferring fsid 6d0ecf22-9155-4684-971a-2f6cde8628c8
> > Using recent ceph image
> > quay.io/ceph/ceph@sha256:2f7f0af8663e73a422f797de605e769ae44eb0297f2a
> > 79324739404cc1765728
> > [ceph: root@hobro /]#
> >
> >
> > Yours,
> > bbk
> >
> > _______________________________________________
> > ceph-users mailing list -- [email protected]
> > To unsubscribe send an email to [email protected]
>
>
> _______________________________________________
> ceph-users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]