Hello,

I mount my Bluestore OSDs in /etc/fstab:

vi /etc/fstab

tmpfs   /var/lib/ceph/osd/ceph-12  tmpfs   rw,relatime 0 0
=====================================================
Then mount everyting in fstab with:
mount -a
======================================================
I activate my OSDs this way on startup: You can find the fsid with

cat /var/lib/ceph/osd/ceph-12/fsid

Then add file named ceph.start so ceph-volume will be run at startup.

vi /etc/local.d/ceph.start
ceph-volume lvm activate 12 827f4a2c-8c1b-427b-bd6c-66d31a0468ac
======================================================
Make it excitable:
chmod 700 /etc/local.d/ceph.start
======================================================
cd /etc/local.d/
./ceph.start
======================================================
I am a Gentoo user and use OpenRC, so this may not apply to you.
======================================================
cd /etc/init.d/
ln -s ceph ceph-osd.12
/etc/init.d/ceph-osd.12 start
rc-update add ceph-osd.12 default

Cary

On Fri, Dec 29, 2017 at 8:47 AM, 赵赵贺东 <zhaohed...@gmail.com> wrote:
> Hello Cary!
> It’s really big surprise for me to receive your reply!
> Sincere thanks to you!
> I know it’s a fake execute file, but it works!
>
> ====================>
> $ cat /usr/sbin/systemctl
> #!/bin/bash
> exit 0
> <====================
>
> I can start my osd by following command
> /usr/bin/ceph-osd --cluster=ceph -i 12 -f --setuser ceph --setgroup ceph
>
> But, threre are still problems.
> 1.Though ceph-osd can start successfully, prepare log and activate log looks
> like errors occurred.
>
> Prepare log:
> =======================================>
> # ceph-volume lvm prepare --bluestore --data vggroup/lv
> Running command: sudo mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-12
> Running command: chown -R ceph:ceph /dev/dm-0
> Running command: sudo ln -s /dev/vggroup/lv /var/lib/ceph/osd/ceph-12/block
> Running command: sudo ceph --cluster ceph --name client.bootstrap-osd
> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o
> /var/lib/ceph/osd/ceph-12/activate.monmap
>  stderr: got monmap epoch 1
> Running command: ceph-authtool /var/lib/ceph/osd/ceph-12/keyring
> --create-keyring --name osd.12 --add-key
> AQAQ+UVa4z2ANRAAmmuAExQauFinuJuL6A56ww==
>  stdout: creating /var/lib/ceph/osd/ceph-12/keyring
>  stdout: added entity osd.12 auth auth(auid = 18446744073709551615
> key=AQAQ+UVa4z2ANRAAmmuAExQauFinuJuL6A56ww== with 0 caps)
> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-12/keyring
> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-12/
> Running command: sudo ceph-osd --cluster ceph --osd-objectstore bluestore
> --mkfs -i 12 --monmap /var/lib/ceph/osd/ceph-12/activate.monmap --key
> **************************************** --osd-data
> /var/lib/ceph/osd/ceph-12/ --osd-uuid 827f4a2c-8c1b-427b-bd6c-66d31a0468ac
> --setuser ceph --setgroup ceph
>  stderr: warning: unable to create /var/run/ceph: (13) Permission denied
>  stderr: 2017-12-29 08:13:08.609127 b66f3000 -1 asok(0x850c62a0)
> AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed to
> bind the UNIX domain socket to '/var/run/ceph/ceph-osd.12.asok': (2) No such
> file or directory
>  stderr:
>  stderr: 2017-12-29 08:13:08.643410 b66f3000 -1
> bluestore(/var/lib/ceph/osd/ceph-12//block) _read_bdev_label unable to
> decode label at offset 66: buffer::malformed_input: void
> bluestore_bdev_label_t::decode(ceph::buffer::list::iterator&) decode past
> end of struct encoding
>  stderr: 2017-12-29 08:13:08.644055 b66f3000 -1
> bluestore(/var/lib/ceph/osd/ceph-12//block) _read_bdev_label unable to
> decode label at offset 66: buffer::malformed_input: void
> bluestore_bdev_label_t::decode(ceph::buffer::list::iterator&) decode past
> end of struct encoding
>  stderr: 2017-12-29 08:13:08.644722 b66f3000 -1
> bluestore(/var/lib/ceph/osd/ceph-12//block) _read_bdev_label unable to
> decode label at offset 66: buffer::malformed_input: void
> bluestore_bdev_label_t::decode(ceph::buffer::list::iterator&) decode past
> end of struct encoding
>  stderr: 2017-12-29 08:13:08.646722 b66f3000 -1
> bluestore(/var/lib/ceph/osd/ceph-12/) _read_fsid unparsable uuid
>  stderr: 2017-12-29 08:14:00.697028 b66f3000 -1 key
> AQAQ+UVa4z2ANRAAmmuAExQauFinuJuL6A56ww==
>  stderr: 2017-12-29 08:14:01.261659 b66f3000 -1 created object store
> /var/lib/ceph/osd/ceph-12/ for osd.12 fsid
> 4e5adad0-784c-41b4-ab72-5f4fae499b3a
> <=======================================
>
> Activate log:
> =======================================>
> # ceph-volume lvm activate --bluestore 12
> 827f4a2c-8c1b-427b-bd6c-66d31a0468ac
> Running command: sudo ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev
> /dev/vggroup/lv --path /var/lib/ceph/osd/ceph-12
> Running command: sudo ln -snf /dev/vggroup/lv
> /var/lib/ceph/osd/ceph-12/block
> Running command: chown -R ceph:ceph /dev/dm-0
> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-12
> Running command: sudo systemctl enable
> ceph-volume@lvm-12-827f4a2c-8c1b-427b-bd6c-66d31a0468ac
> Running command: sudo systemctl start ceph-osd@12
> <=======================================
>
> After activate operation , osd did not start , but I can start osd by the
> following command(before restart host)
> /usr/bin/ceph-osd --cluster=ceph -i 12 -f --setuser ceph --setgroup ceph
>
>
>
> 2. After host reboot, everything about ceph-osd has been lost.
>     Because before I reboot, /var/lib/ceph/osd/ceph-12 was mounted on tmpfs,
> so after reboot I lost everything about ceph-osd.
> #df -h
>  /dev/root        15G  2.4G   12G  18% /
> devtmpfs       1009M  4.0K 1009M   1% /dev
> none            4.0K     0  4.0K   0% /sys/fs/cgroup
> none            202M  156K  202M   1% /run
> none            5.0M     0  5.0M   0% /run/lock
> none           1009M     0 1009M   0% /run/shm
> none            100M     0  100M   0% /run/user
> tmpfs          1009M   48K 1009M   1% /var/lib/ceph/osd/ceph-12
>
>
> 3. ceph-osd can not start automatically.
> I think there are something wrong in osd upstart, I should add some
> operation about osd upstart.
>
>
> It seems that this ceph-volume for ubuntu14.04 is not a easy problem for me
> , so any suggestions  or hints about the problems will be appreciated!
>
>
>
>
> 在 2017年12月29日,下午2:06,Cary <dynamic.c...@gmail.com> 写道:
>
>
> You could add a file named  /usr/sbin/systemctl and add:
> exit 0
> to it.
>
> Cary
>
> On Dec 28, 2017, at 18:45, 赵赵贺东 <zhaohed...@gmail.com> wrote:
>
>
> Hello ceph-users!
>
> I am a ceph user from china.
> Our company deploy ceph on arm ubuntu 14.04.
> Ceph Version is luminous 12.2.2.
> When I try to activate osd by ceph-volume, I got the following error.(osd
> prepare stage seems work normally)
> It seems that ceph-volume only work under systemd, but ubuntu 14.04 does not
> support systemd.
> How can I deploy osd in ubuntu 14.04 by ceph-volume?
> Will ceph-volume support upstart in the future?
>
> ===============================================================>
> # ceph-volume lvm activate --bluestore 12
> 03fa2757-412d-4892-af8a-f2260294a2dc
> Running command: sudo ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev
> /dev/vggroup/lvdata --path /var/lib/ceph/osd/ceph-12
> Running command: sudo ln -snf /dev/vggroup/lvdata
> /var/lib/ceph/osd/ceph-12/block
> Running command: chown -R ceph:ceph /dev/dm-2
> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-12
> Running command: sudo systemctl enable
> ceph-volume@lvm-12-03fa2757-412d-4892-af8a-f2260294a2dc
>  stderr: sudo: systemctl: command not found
> -->  RuntimeError: command returned non-zero exit status: 1
> <================================================================
>
>
> Your reply will be appreciated!
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to