Hi,

I would appreciate some guidance here

I am trying to reconfigure OSD to put DB/WAL on NVME
so I "out" and "ceph osd rm" them ( because  they were automatically
configured
and I did not "set-unamage" them properly ) then "wipefs -af " them

Now CEPH is seeing drives available for deploying OSD
 on but it fails doing it due to "device or resource busy"

Attempting to destroy and zap it fails due to "device or resource busy"
even after I "wipefs -af /dev/sdaa" it

this is what it is on it ( output of lsblk)

sdaa
                           18.2T disk
└─ceph--dfaa321c--b04d--48e8--b78b--7c39436be0e5-osd--block--36473043--6d2c--4c7a--b61c--b057e83f0169
18.2T lvm

lvs,vgs does not find any logical disks /volume
cephadm ceph-volume inventory is seeing the disks as "available = True"
cephadm ceph-volume lvm list  is NOT finding any LVM on /dev/sdaa
docker ps -a does not show any "left over" OSD docker

There is a reference ( block link) to the
/dev/ceph--dfaa321c--b04d--48e8--b78b--7c39436be0e5-osd--block--36473043--6d2c--4c7a--b61c--b057e83f016
in /var/lib/cephFSID/removed/osd.NUMER_DATE

How can I 'clean up'  the disk ?

Many thanks
Steven

 cephadm ceph-volume lvm zap --destroy /dev/sdaa
Inferring fsid 0cfa836d-68b5-11f0-90bf-7cc2558e5ce8
Not using image
'sha256:4892a7ef541bbfe6181ff8fd5c8e03957338f7dd73de94986a5f15e185dacd51'
as it's not in list of non-dangling images with ceph=True label
Non-zero exit code 1 from /usr/bin/docker run --rm --ipc=host
--stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint
/usr/sbin/ceph-volume --privileged --group-add=disk --init -e
CONTAINER_IMAGE=quay.io/ceph/ceph:v19 -e NODE_NAME=ceph-host-1 -e
CEPH_VOLUME_SKIP_RESTORECON=yes -e CEPH_VOLUME_DEBUG=1 -v
/var/run/ceph/0cfa836d-68b5-11f0-90bf-7cc2558e5ce8:/var/run/ceph:z -v
/var/log/ceph/0cfa836d-68b5-11f0-90bf-7cc2558e5ce8:/var/log/ceph:z -v
/var/lib/ceph/0cfa836d-68b5-11f0-90bf-7cc2558e5ce8/crash:/var/lib/ceph/crash:z
-v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v
/run/lock/lvm:/run/lock/lvm -v /:/rootfs:rslave -v
/tmp/ceph-tmp419ql4q7:/etc/ceph/ceph.conf:z quay.io/ceph/ceph:v19 lvm zap
--destroy /dev/sdaa
/usr/bin/docker: stderr --> Zapping: /dev/sdaa
/usr/bin/docker: stderr --> Removing all BlueStore signature on /dev/sdaa
if any...
/usr/bin/docker: stderr Running command: /usr/bin/ceph-bluestore-tool
zap-device --dev /dev/sdaa --yes-i-really-really-mean-it
/usr/bin/docker: stderr  stderr: error from zap: (16) Device or resource
busy
/usr/bin/docker: stderr 2025-09-02T17:41:02.385+0000 72def157f9c0 -1
bdev(0x57b37fe52000 /dev/sdaa) open open got: (16) Device or resource busy



Some extra info


cephadm ceph-volume inventory output


Device Path               Size         Device nodes    rotates available
Model name
/dev/nvme3n1              1.46 TB      nvme3n1         False   True
 MTFDKCC1T6TGQ-1BK1DABYY
/dev/sdaa                 18.19 TB     sdaa            True    True
 ST20000NM007D-3D
/dev/sdv                  18.19 TB     sdv             True    True
 ST20000NM007D-3D
/dev/sdw                  18.19 TB     sdw             True    True
 ST20000NM007D-3D
/dev/sdx                  18.19 TB     sdx             True    True
 ST20000NM007D-3D
/dev/sdy                  18.19 TB     sdy             True    True
 ST20000NM007D-3D
/dev/sdz                  18.19 TB     sdz             True    True
 ST20000NM007D-3D


lsblk output
NAME
                            SIZE TYPE MOUNTPOINT
sdaa
                           18.2T disk
└─ceph--dfaa321c--b04d--48e8--b78b--7c39436be0e5-osd--block--36473043--6d2c--4c7a--b61c--b057e83f0169
18.2T lvm

lvs output ( greped for lsblk /dev/sdaa and couldn't find it )
 osd-block-29a5b1f5-d442-4756-a61d-e230832a3577
ceph-013dc6ae-90c7-4271-b308-895bbafe8840 -wi-ao----  <6.99t

  osd-block-5f7471f4-22ee-4cf9-9a16-2854a0c26496
ceph-27dfe22b-8316-40e3-aeee-1686bf0a95b7 -wi-ao----  <6.99t

  osd-block-ff6ca83d-80c5-4fff-a50d-399d94ba1b1b
ceph-2de80ef9-8095-44be-9df2-91f458b1f1ea -wi-ao----  18.19t

  osd-block-fc022b7e-3d85-4435-998e-6513020bbdf3
ceph-57f7e6bf-cf30-4d81-8c97-d0cb503cb8f8 -wi-ao----  13.97t

  osd-block-347afb43-812c-4330-a795-060e76f33f89
ceph-5815ca32-056b-476a-8a16-64c06b9c3fe9 -wi-ao----  <6.99t

  osd-block-c96901b7-8b7e-4a94-be0a-ab1316fd114e
ceph-58a9cfda-5090-47f5-bfe8-b993d6133218 -wi-ao----  <6.99t

  osd-block-5e069648-e0ec-4490-a3a9-e758d6e3e1d4
ceph-59b5ad1b-a35b-4c46-b737-7328eee6ec45 -wi-ao----  13.97t

  osd-block-5d7dc85e-dcc8-4d87-b7ba-d61201dccd5e
ceph-6aedba8c-3c5a-476d-9e0c-9fd159c8c09e -wi-ao----  <6.99t

  osd-block-d18a4abc-f723-4ac8-9d05-d02edd7e6018
ceph-6f7e544b-9ead-40c4-82ce-646272461b36 -wi-ao----  <6.99t

  osd-block-33404537-6f45-4f10-af8a-8de491e8bcc9
ceph-7060eb95-cfb9-441b-9094-1803348ca3f2 -wi-ao----  18.19t

  osd-block-153d1ed5-007c-439e-ad9e-c20d1880ff4d
ceph-9729cc29-705c-42e5-bcc6-5d7cfd62a4ac -wi-ao----  18.19t

  osd-block-15f14591-43c9-42ed-8410-d7f216019b55
ceph-9d628846-1c8f-4c73-9c7f-6e34c69210a3 -wi-ao----  18.19t

  osd-block-44c310eb-2002-4817-8b2e-beacb0a70806
ceph-ab1ca6f2-30c8-45a7-bfec-4e09089ec856 -wi-ao----  <6.99t

  osd-block-b55d7d9f-8e2a-4da5-b9c5-ace96975286d
ceph-ba4098bf-6808-4a65-93a6-e84f3ff7936a -wi-ao----  13.97t

  osd-block-a97d2620-d07a-4bd7-8f19-66801f5d613b
ceph-c93d909a-b944-4dc5-a9b2-ed3441226c46 -wi-ao----  18.19t

  osd-block-5c134f31-3a17-4fe9-989f-fb3a1dc2c0e9
ceph-c9e00bac-501b-4449-b23c-d914c61a6838 -wi-ao----  <6.99t

  osd-block-89390a0d-6bc3-41a5-a79b-ebac922d540f
ceph-d8a23938-be11-4cee-ba13-951996ca9328 -wi-ao----  <6.99t

  osd-block-cfabfe08-5e42-46ba-8412-1d264e9312d3
ceph-dcff97c0-5caa-4cd2-87f5-baa552b0b172 -wi-ao----  <6.99t

  osd-block-2e86866b-b537-4b29-86e7-490e627f82be
ceph-de7ee41e-a121-4ecd-a838-cf284e811b7e -wi-ao----  18.19t

  osd-block-ab0b170e-91e0-41d4-a5ea-b5c4fef6c3d8
ceph-e6d22c3b-bfb6-494c-b09d-15a667816442 -wi-ao----  <6.99t

  osd-db-012f9187-1596-4db8-96ba-809a96bce8fd
 ceph-f0e77cf5-9878-4a9b-bdc4-22211aadb951 -wi-ao---- 248.39g

  osd-db-124f39e2-f28d-47a8-acce-fb1364f7a638
 ceph-f0e77cf5-9878-4a9b-bdc4-22211aadb951 -wi-ao---- 248.39g

  osd-db-17edf308-b36b-42be-8855-55833d35f983
 ceph-f0e77cf5-9878-4a9b-bdc4-22211aadb951 -wi-ao---- 248.39g

  osd-db-55eb88b3-89a0-4931-a1be-d73e2aaf527f
 ceph-f0e77cf5-9878-4a9b-bdc4-22211aadb951 -wi-ao---- 248.39g

  osd-db-5e03ba4b-5641-49c6-a477-1b37dffcd398
 ceph-f0e77cf5-9878-4a9b-bdc4-22211aadb951 -wi-ao---- 248.39g

  osd-db-b22a1142-8dec-4035-96b1-7717c4587bcb
 ceph-f0e77cf5-9878-4a9b-bdc4-22211aadb951 -wi-ao---- 248.39g

  osd-block-f9c81b89-a903-4a11-9c2a-9171b4a5760a
ceph-f8276850-dfb2-4689-a5b1-e54925605dc4 -wi-ao----  <6.99t


VGS output ( ( greped for lsblk /dev/sdaa result and couldn't find it )

VG                                        #PV #LV #SN Attr   VSize  VFree
  ceph-013dc6ae-90c7-4271-b308-895bbafe8840   1   1   0 wz--n- <6.99t    0
  ceph-27dfe22b-8316-40e3-aeee-1686bf0a95b7   1   1   0 wz--n- <6.99t    0
  ceph-2de80ef9-8095-44be-9df2-91f458b1f1ea   1   1   0 wz--n- 18.19t    0
  ceph-57f7e6bf-cf30-4d81-8c97-d0cb503cb8f8   1   1   0 wz--n- 13.97t    0
  ceph-5815ca32-056b-476a-8a16-64c06b9c3fe9   1   1   0 wz--n- <6.99t    0
  ceph-58a9cfda-5090-47f5-bfe8-b993d6133218   1   1   0 wz--n- <6.99t    0
  ceph-59b5ad1b-a35b-4c46-b737-7328eee6ec45   1   1   0 wz--n- 13.97t    0
  ceph-6aedba8c-3c5a-476d-9e0c-9fd159c8c09e   1   1   0 wz--n- <6.99t    0
  ceph-6f7e544b-9ead-40c4-82ce-646272461b36   1   1   0 wz--n- <6.99t    0
  ceph-7060eb95-cfb9-441b-9094-1803348ca3f2   1   1   0 wz--n- 18.19t    0
  ceph-9729cc29-705c-42e5-bcc6-5d7cfd62a4ac   1   1   0 wz--n- 18.19t    0
  ceph-9d628846-1c8f-4c73-9c7f-6e34c69210a3   1   1   0 wz--n- 18.19t    0
  ceph-ab1ca6f2-30c8-45a7-bfec-4e09089ec856   1   1   0 wz--n- <6.99t    0
  ceph-ba4098bf-6808-4a65-93a6-e84f3ff7936a   1   1   0 wz--n- 13.97t    0
  ceph-c93d909a-b944-4dc5-a9b2-ed3441226c46   1   1   0 wz--n- 18.19t    0
  ceph-c9e00bac-501b-4449-b23c-d914c61a6838   1   1   0 wz--n- <6.99t    0
  ceph-d8a23938-be11-4cee-ba13-951996ca9328   1   1   0 wz--n- <6.99t    0
  ceph-dcff97c0-5caa-4cd2-87f5-baa552b0b172   1   1   0 wz--n- <6.99t    0
  ceph-de7ee41e-a121-4ecd-a838-cf284e811b7e   1   1   0 wz--n- 18.19t    0
  ceph-e6d22c3b-bfb6-494c-b09d-15a667816442   1   1   0 wz--n- <6.99t    0
  ceph-f0e77cf5-9878-4a9b-bdc4-22211aadb951   1   6   0 wz--n- <1.46t 8.00m
  ceph-f8276850-dfb2-4689-a5b1-e54925605dc4   1   1   0 wz--n- <6.99t    0
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to