No. They are new disks. Not used before. I am setting up a new cluster.

I did run this command to zap "ceph orch device zap od-node1 /dev/sda
--force"

Here is lsblk output.

 lsblk
NAME
MAJ:MIN RM  SIZE RO TYPE MOUNTPOINTS
sda
 8:0    0   14T  0 disk
└─ceph--54ce8bd1--384b--4e94--a9ed--b231ff649bd2-osd--block--ace35922--b31d--49db--ad86--839acca0990c

 253:0    0   14T  0 lvm
sdb
 8:16   0   14T  0 disk
└─ceph--ce627d0c--5a73--4df4--9690--752fa20d614e-osd--block--bbe3ce9e--ae82--453b--bca8--191f2300f780

 253:1    0   14T  0 lvm
sdc
 8:32   0   14T  0 disk
└─ceph--543e9cff--9447--4a3d--a6e8--2c3dd7e1810b-osd--block--a013c7f6--aa1b--4c96--be71--9fa3bf2a910c

 253:2    0   14T  0 lvm
sdd
 8:48   0   14T  0 disk
└─ceph--d3e34042--0a3d--476d--acdb--24f48d7c59e1-osd--block--c77a03ea--9280--46b4--8ff4--2e11ad25eddc

 253:4    0   14T  0 lvm
sde
 8:64   0   14T  0 disk
└─ceph--d3e7308a--09a4--4e91--8d12--8f3b1cc94c6f-osd--block--1f83e5df--4e6c--4613--b8dd--5fea7fdcec4d

 253:3    0   14T  0 lvm
sdf
 8:80   0   14T  0 disk
└─ceph--6a86dae1--f879--414a--8115--03fbef2f36e4-osd--block--debb0534--a12d--48e1--8d4c--d0c07e44ae6c

 253:5    0   14T  0 lvm
sdg
 8:96   0   14T  0 disk
└─ceph--f52542ec--0aa7--4007--8d42--f81723e946e3-osd--block--07837bc6--d868--48b5--b950--9cef1ce54eb4

 253:6    0   14T  0 lvm
sdh
 8:112  0   14T  0 disk



Here is fdisk output :

 fdisk -l /dev/sda
Disk /dev/sda: 13.97 TiB, 15360950534144 bytes, 30001856512 sectors
Disk model: WUS5EA1A1ESP5E3
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes


fdisk -l /dev/sdb
Disk /dev/sdb: 13.97 TiB, 15360950534144 bytes, 30001856512 sectors
Disk model: WUS5EA1A1ESP5E3
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes

Thanks,
Gagan

On Thu, Jul 31, 2025 at 9:21 PM Anthony D'Atri <anthony.da...@gmail.com>
wrote:

> How did you zap the new drives?  Had they been used before?
>
> Does `lsblk` show anything on the drives?  What about fdisk?
>
> # fdisk -l /dev/sda
> Disk /dev/sda: 744.63 GiB, 799535005696 bytes, 1561591808 sectors
> Disk model: PERC H730 Mini
> Units: sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 4096 bytes
> I/O size (minimum/optimal): 4096 bytes / 4096 bytes
> Disklabel type: gpt
> Disk identifier: A00D586C-45F0-49BE-8290-A2C8B872145F
>
> Device     Start        End    Sectors   Size Type
> /dev/sda1   2048       4095       2048     1M BIOS boot
> #
> # lsblk
> NAME   MAJ:MIN RM   SIZE RO TYPE MOUNTPOINTS
> sda      8:0    0 744.6G  0 disk
> ├─sda1   8:1    0     1M  0 part
> └─sda2   8:2    0 744.6G  0 part /
>
>
>
> > On Jul 31, 2025, at 11:48 AM, gagan tiwari <
> gagan.tiw...@mathisys-india.com> wrote:
> >
> > Hi Guys,
> >                   I am setting up a new ceph cluster. After adding osd
> > devices when I did run *ceph orch device ls*
> >
> > It shows * Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM
> > detected  for all osd devices.*
> >
> > All the OSD devices are new nvme disks.
> >
> > I tried to rm , zap , destroy disks and re-added them but again same
> > messages for all disks.
> >
> > Please let me know how to fix this.
> >
> > Thanks,
> > Gagan
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@ceph.io
> > To unsubscribe send an email to ceph-users-le...@ceph.io
>
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to