> -2 21.81000 host node24 > 0 hdd 7.26999 osd.0 destroyed 0 > 1.00000 > 8 hdd 7.26999 osd.8 up 1.00000 > 1.00000 > 16 hdd 7.26999 osd.16 up 1.00000 > 1.00000
Should I do these prior to running without the osd-id specified? > # ceph osd crush remove osd.$ID > # ceph auth del osd.$ID > # ceph osd rm osd.$ID And then it fill in the missing osd.0. Will set norebalance flag first to prevent data reshuffle upon the osd being removed from the crush map. Thanks, Reed > On Jan 9, 2018, at 2:05 PM, Alfredo Deza <[email protected]> wrote: > > On Tue, Jan 9, 2018 at 2:19 PM, Reed Dier <[email protected] > <mailto:[email protected]>> wrote: >> Hi ceph-users, >> >> Hoping that this is something small that I am overlooking, but could use the >> group mind to help. >> >> Ceph 12.2.2, Ubuntu 16.04 environment. >> OSD (0) is an 8TB spinner (/dev/sda) and I am moving from a filestore >> journal to a blocks.db and WAL device on an NVMe partition (/dev/nvme0n1p5). >> >> I have an OSD that I am trying to convert to bluestore and running into some >> trouble. >> >> Started here until the ceps-volume create statement, which doesn’t work. >> http://docs.ceph.com/docs/master/rados/operations/bluestore-migration/ >> Worth mentioning I also flushed the journal on the nvme partition before >> nuking the OSD. >> >> $ sudo ceph-osd -i 0 --flush-journal >> >> >> So I first started with this command: >> >> $ sudo ceph-volume lvm create --bluestore --data /dev/sda --block.db >> /dev/nvme0n1p5 --osd-id 0 >> >> >> Pastebin to the ceph-volume log: https://pastebin.com/epkM3aP6 >> >> However the OSD doesn’t start. > > I was just able to replicate this by using an ID that doesn't exist in > the cluster. On a cluster with just one OSD (with an ID of 0) I > created > an OSD with --osd-id 3, and had the exact same results. > >> >> Pastebin to ceph-osd log: https://pastebin.com/9qEsAJzA >> >> I tried restarting the process, by deleting the LVM structures, zapping the >> disk using ceph-volume. >> This time using prepare and activate instead of create. >> >> $ sudo ceph-volume lvm prepare --bluestore --data /dev/sda --block.db >> /dev/nvme0n1p5 --osd-id 0 >> >> $ sudo ceph-volume lvm activate --bluestore 0 >> 227e1721-cd2e-4d7e-bb48-bc2bb715a038 >> >> >> Also ran the enable on the ceph-volume systemd unit per >> http://docs.ceph.com/docs/master/install/manual-deployment/ >> >> $ sudo systemctl enable >> ceph-volume@lvm-0-227e1721-cd2e-4d7e-bb48-bc2bb715a038 >> >> >> Same results. >> >> Any help is greatly appreciated. > > Could you try without passing --osd-id ? >> >> Thanks, >> >> Reed >> >> _______________________________________________ >> ceph-users mailing list >> [email protected] <mailto:[email protected]> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
