I just tried it, and that does indeed get the OSD to start. However, it doesn't add it to the appropriate place so it would survive a reboot. In my case, running 'service ceph status osd.16' still results in the same line I posted above.
There's still something broken such that 'ceph-disk activate' works correctly, but using the long-form version does not. On Wed, Feb 25, 2015 at 4:03 PM, Robert LeBlanc <rob...@leblancnet.us> wrote: > Step #6 in > http://ceph.com/docs/master/install/manual-deployment/#long-form > only set-ups the file structure for the OSD, it doesn't start the long > running process. > > On Wed, Feb 25, 2015 at 2:59 PM, Kyle Hutson <kylehut...@ksu.edu> wrote: > > But I already issued that command (back in step 6). > > > > The interesting part is that "ceph-disk activate" apparently does it > > correctly. Even after reboot, the services start as they should. > > > > On Wed, Feb 25, 2015 at 3:54 PM, Robert LeBlanc <rob...@leblancnet.us> > > wrote: > >> > >> I think that your problem lies with systemd (even though you are using > >> SysV syntax, systemd is really doing the work). Systemd does not like > >> multiple arguments and I think this is why it is failing. There is > >> supposed to be some work done to get systemd working ok, but I think > >> it has the limitation of only working with a cluster named 'ceph' > >> currently. > >> > >> What I did to get around the problem was to run the osd command > manually: > >> > >> ceph-osd -i <osd#> > >> > >> Once I understand the under-the-hood stuff, I moved to ceph-disk and > >> now because of the GPT partition IDs, udev automatically starts up the > >> OSD process at boot/creation and moves to the appropiate CRUSH > >> location (configuratble in ceph.conf > >> http://ceph.com/docs/master/rados/operations/crush-map/#crush-location, > >> an example: crush location = host=test rack=rack3 row=row8 > >> datacenter=local region=na-west root=default). To restart an OSD > >> process, I just kill the PID for the OSD then issue ceph-disk activate > >> /dev/sdx1 to restart the OSD process. You probably could stop it with > >> systemctl since I believe udev creates a resource for it (I should > >> probably look into that now that this system will be going production > >> soon). > >> > >> On Wed, Feb 25, 2015 at 2:13 PM, Kyle Hutson <kylehut...@ksu.edu> > wrote: > >> > I'm having a similar issue. > >> > > >> > I'm following http://ceph.com/docs/master/install/manual-deployment/ > to > >> > a T. > >> > > >> > I have OSDs on the same host deployed with the short-form and they > work > >> > fine. I am trying to deploy some more via the long form (because I > want > >> > them > >> > to appear in a different location in the crush map). Everything > through > >> > step > >> > 10 (i.e. ceph osd crush add {id-or-name} {weight} > >> > [{bucket-type}={bucket-name} ...] ) works just fine. When I go to step > >> > 11 > >> > (sudo /etc/init.d/ceph start osd.{osd-num}) I get: > >> > /etc/init.d/ceph: osd.16 not found (/etc/ceph/ceph.conf defines > >> > mon.hobbit01 > >> > osd.7 osd.15 osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 > >> > osd.6 > >> > osd.11 osd.5 osd.4 osd.0 , /var/lib/ceph defines mon.hobbit01 osd.7 > >> > osd.15 > >> > osd.10 osd.9 osd.1 osd.14 osd.2 osd.3 osd.13 osd.8 osd.12 osd.6 osd.11 > >> > osd.5 > >> > osd.4 osd.0) > >> > > >> > > >> > > >> > On Wed, Feb 25, 2015 at 11:55 AM, Travis Rhoden <trho...@gmail.com> > >> > wrote: > >> >> > >> >> Also, did you successfully start your monitor(s), and define/create > the > >> >> OSDs within the Ceph cluster itself? > >> >> > >> >> There are several steps to creating a Ceph cluster manually. I'm > >> >> unsure > >> >> if you have done the steps to actually create and register the OSDs > >> >> with the > >> >> cluster. > >> >> > >> >> - Travis > >> >> > >> >> On Wed, Feb 25, 2015 at 9:49 AM, Leszek Master <keks...@gmail.com> > >> >> wrote: > >> >>> > >> >>> Check firewall rules and selinux. It sometimes is a pain in the ... > :) > >> >>> > >> >>> 25 lut 2015 01:46 "Barclay Jameson" <almightybe...@gmail.com> > >> >>> napisaĆ(a): > >> >>> > >> >>>> I have tried to install ceph using ceph-deploy but sgdisk seems to > >> >>>> have too many issues so I did a manual install. After mkfs.btrfs on > >> >>>> the disks and journals and mounted them I then tried to start the > >> >>>> osds > >> >>>> which failed. The first error was: > >> >>>> #/etc/init.d/ceph start osd.0 > >> >>>> /etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines , > >> >>>> /var/lib/ceph defines ) > >> >>>> > >> >>>> I then manually added the osds to the conf file with the following > as > >> >>>> an example: > >> >>>> [osd.0] > >> >>>> osd_host = node01 > >> >>>> > >> >>>> Now when I run the command : > >> >>>> # /etc/init.d/ceph start osd.0 > >> >>>> > >> >>>> There is no error or output from the command and in fact when I do > a > >> >>>> ceph -s no osds are listed as being up. > >> >>>> Doing as ps aux | grep -i ceph or ps aux | grep -i osd shows there > >> >>>> are > >> >>>> no osd running. > >> >>>> I also have done htop to see if any process are running and none > are > >> >>>> shown. > >> >>>> > >> >>>> I had this working on SL6.5 with Firefly but Giant on Centos 7 has > >> >>>> been nothing but a giant pain. > >> >>>> _______________________________________________ > >> >>>> ceph-users mailing list > >> >>>> ceph-users@lists.ceph.com > >> >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> >>> > >> >>> > >> >>> _______________________________________________ > >> >>> ceph-users mailing list > >> >>> ceph-users@lists.ceph.com > >> >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> >>> > >> >> > >> >> > >> >> _______________________________________________ > >> >> ceph-users mailing list > >> >> ceph-users@lists.ceph.com > >> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> >> > >> > > >> > > >> > _______________________________________________ > >> > ceph-users mailing list > >> > ceph-users@lists.ceph.com > >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> > > > > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com