Does the block and/or wal partition need to be an LV? I just passed ceph-volume the raw partition and it seems to be working fine.
On Fri, Aug 17, 2018 at 2:54 PM Alfredo Deza <[email protected]> wrote: > On Fri, Aug 17, 2018 at 10:24 AM, Robert Stanford > <[email protected]> wrote: > > > > I was using the ceph-volume create command, which I understand combines > the > > prepare and activate functions. > > > > ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc --block.db > > /dev/sdb --block.wal /dev/sdb > > > > That is the command context I've found on the web. Is it wrong? > > It is very wrong :( > > If this was coming from our docs, it needs to be fixed because it will > never work. > > If you really want to place both block.db and block.wal on /dev/sdb, > you will need to create one LV for each. ceph-volume will not do this > for you. > > And then you can pass those newly created LVs like: > > ceph-volume lvm create --osd-id 0 --bluestore --data /dev/sdc > --block.db sdb-vg/block-lv --block.wal sdb-vg/wal-lv > > > > > > > Thanks > > R > > > > On Fri, Aug 17, 2018 at 5:55 AM Alfredo Deza <[email protected]> wrote: > >> > >> On Thu, Aug 16, 2018 at 9:00 PM, Robert Stanford > >> <[email protected]> wrote: > >> > > >> > I am following the steps to my filestore journal with a bluestore > >> > journal > >> > ( > http://docs.ceph.com/docs/mimic/rados/operations/bluestore-migration/). > >> > It > >> > is broken at ceph-volume lvm create. Here is my error: > >> > > >> > --> Zapping successful for: /dev/sdc > >> > Preparing sdc > >> > Running command: /bin/ceph-authtool --gen-print-key > >> > Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd > >> > --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json > >> > Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd > >> > --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new > >> > ff523216-350d-4ca0-9022-0c17662c2c3b 10 > >> > Running command: vgcreate --force --yes > >> > ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a /dev/sdc > >> > stdout: Physical volume "/dev/sdc" successfully created. > >> > stdout: Volume group "ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a" > >> > successfully created > >> > Running command: lvcreate --yes -l 100%FREE -n > >> > osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b > >> > ceph-459b4fbe-e3c4-4f28-b58e-3496bf3ea95a > >> > stdout: Logical volume > "osd-block-ff523216-350d-4ca0-9022-0c17662c2c3b" > >> > created. > >> > --> blkid could not detect a PARTUUID for device: sdb > >> > --> Was unable to complete a new OSD, will rollback changes > >> > --> OSD will be destroyed, keeping the ID because it was provided with > >> > --osd-id > >> > Running command: ceph osd destroy osd.10 --yes-i-really-mean-it > >> > stderr: destroyed osd.10 > >> > --> RuntimeError: unable to use device > >> > > >> > Note that SDB is the SSD journal. It has been zapped prior. > >> > >> I can't see what the actual command you used is, but I am guessing you > >> did something like: > >> > >> ceph-volume lvm prepare --filestore --data /dev/sdb --journal /dev/sdb > >> > >> Which is not possible. There are a few ways you can do this (see: > >> http://docs.ceph.com/docs/master/ceph-volume/lvm/prepare/#filestore ) > >> > >> With a raw device and a pre-created partition (must have a PARTUUID): > >> > >> ceph-volume lvm prepare --data /dev/sdb --journal /dev/sdc1 > >> > >> With LVs: > >> > >> ceph-volume lvm prepare --data vg/my-data --journal vg/my-journal > >> > >> With an LV for data and a partition: > >> > >> ceph-volume lvm prepare --data vg/my-data --journal /dev/sdc1 > >> > >> > > >> > What is going wrong, and how can I fix it? > >> > > >> > Thank you > >> > R > >> > > >> > > >> > _______________________________________________ > >> > ceph-users mailing list > >> > [email protected] > >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> > > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
