Hello list,

i upgraded to Debian 10, after that i upgraded from luminous to nautilus.
I restarted the mons, then the OSDs.

Everything was up and healthy.
After rebooting a node, only 3/10 OSD start up:

-4       20.07686     host ceph03
 4   hdd  2.67020         osd.4     down  1.00000 1.00000
 5   hdd  1.71660         osd.5       up  1.00000 1.00000
 6   hdd  1.71660         osd.6       up  1.00000 1.00000
10   hdd  2.67029         osd.10      up  1.00000 1.00000
15   hdd  2.00000         osd.15    down  1.00000 1.00000
17   hdd  1.20000         osd.17    down  1.00000 1.00000
20   hdd  1.71649         osd.20    down  1.00000 1.00000
21   hdd  2.00000         osd.21    down  1.00000 1.00000
27   hdd  1.71649         osd.27    down  1.00000 1.00000
32   hdd  2.67020         osd.32    down  1.00000 1.00000

root@ceph03:~# /usr/bin/ceph-osd -f --cluster ceph --id 32 --setuser
ceph --setgroup ceph
2020-03-25 15:46:36.330 7efddde5ec80 -1 auth: unable to find a keyring
on /var/lib/ceph/osd/ceph-32/keyring: (2) No such file or directory
2020-03-25 15:46:36.330 7efddde5ec80 -1 AuthRegistry(0x56531c50a140)
no keyring found at /var/lib/ceph/osd/ceph-32/keyring, disabling cephx
2020-03-25 15:46:36.330 7efddde5ec80 -1 auth: unable to find a keyring
on /var/lib/ceph/osd/ceph-32/keyring: (2) No such file or directory
2020-03-25 15:46:36.330 7efddde5ec80 -1 AuthRegistry(0x7ffd04120468)
no keyring found at /var/lib/ceph/osd/ceph-32/keyring, disabling cephx
failed to fetch mon config (--no-mon-config to skip)

root@ceph03:~# df
Filesystem     1K-blocks    Used Available Use% Mounted on
udev            24624580       0  24624580   0% /dev
tmpfs            4928216    9544   4918672   1% /run
/dev/sda3       47930248 5209760  40262684  12% /
tmpfs           24641068       0  24641068   0% /dev/shm
tmpfs               5120       0      5120   0% /run/lock
tmpfs           24641068       0  24641068   0% /sys/fs/cgroup
/dev/sda1         944120  144752    734192  17% /boot
tmpfs           24641068      24  24641044   1% /var/lib/ceph/osd/ceph-1
tmpfs           24641068      24  24641044   1% /var/lib/ceph/osd/ceph-6
tmpfs           24641068      24  24641044   1% /var/lib/ceph/osd/ceph-5
tmpfs           24641068      24  24641044   1% /var/lib/ceph/osd/ceph-10
tmpfs            4928212       0   4928212   0% /run/user/0

root@ceph03:~# ceph-volume lvm list


====== osd.1 =======

  [block]
/dev/ceph-9af8fc69-cab8-4c12-b51e-5746a0f0fc51/osd-block-b4987093-4fa5-47bd-8ddc-102b98444067

      block device
/dev/ceph-9af8fc69-cab8-4c12-b51e-5746a0f0fc51/osd-block-b4987093-4fa5-47bd-8ddc-102b98444067
      block uuid                HSK6Da-elP2-CFYz-s0RH-UNiw-bey0-dVcml1
      cephx lockbox secret
      cluster fsid              5436dd5d-83d4-4dc8-a93b-60ab5db145df
      cluster name              ceph
      crush device class        None
      encrypted                 0
      osd fsid                  b4987093-4fa5-47bd-8ddc-102b98444067
      osd id                    1
      type                      block
      vdo                       0
      devices                   /dev/sdj

====== osd.10 ======

  [block]
/dev/ceph-78f2730d-7277-4d1f-8909-449b45339f80/osd-block-fa241441-1758-4b85-9799-988eee3b2b3f

      block device
/dev/ceph-78f2730d-7277-4d1f-8909-449b45339f80/osd-block-fa241441-1758-4b85-9799-988eee3b2b3f
      block uuid                440fNG-guO2-l1WJ-m5cR-GUkz-ZTUd-Fcz5Ml
      cephx lockbox secret
      cluster fsid              5436dd5d-83d4-4dc8-a93b-60ab5db145df
      cluster name              ceph
      crush device class        None
      encrypted                 0
      osd fsid                  fa241441-1758-4b85-9799-988eee3b2b3f
      osd id                    10
      type                      block
      vdo                       0
      devices                   /dev/sdl

====== osd.5 =======

  [block]
/dev/ceph-793608ca-9dd1-4a4f-a776-c1e292127899/osd-block-112e0c75-f61b-4e50-9bb5-775bacd854af

      block device
/dev/ceph-793608ca-9dd1-4a4f-a776-c1e292127899/osd-block-112e0c75-f61b-4e50-9bb5-775bacd854af
      block uuid                Z6VeNx-S9sg-ZOsh-HTw9-ykTc-YBrh-qFwz5i
      cephx lockbox secret
      cluster fsid              5436dd5d-83d4-4dc8-a93b-60ab5db145df
      cluster name              ceph
      crush device class        None
      encrypted                 0
      osd fsid                  112e0c75-f61b-4e50-9bb5-775bacd854af
      osd id                    5
      type                      block
      vdo                       0
      devices                   /dev/sdb

====== osd.6 =======

  [block]
/dev/ceph-4b0cee89-03f4-4853-bc1d-09e0eb772799/osd-block-35288829-c1f6-42ab-aeb0-f2915a389e48

      block device
/dev/ceph-4b0cee89-03f4-4853-bc1d-09e0eb772799/osd-block-35288829-c1f6-42ab-aeb0-f2915a389e48
      block uuid                G9qHxC-dN0b-XBes-QVss-Bzwa-7Xtw-ikksgM
      cephx lockbox secret
      cluster fsid              5436dd5d-83d4-4dc8-a93b-60ab5db145df
      cluster name              ceph
      crush device class        None
      encrypted                 0
      osd fsid                  35288829-c1f6-42ab-aeb0-f2915a389e48
      osd id                    6
      type                      block
      vdo                       0
      devices                   /dev/sdc

I would mount it and run the osd deamon manually, but ceph-disk list
seems to be gone in Nautilus. Therefore i dont know where to mount
what.

Any ideas on that?


Cheers,
Michael
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to