Thanks for the reply, the service is still showing as failed. How to bring
the osds service up. Ceph osd tree shows all osds as UP.

 [root@Admin ceph]# systemctl restart ceph-osd@osd.2.service
[root@Admin ceph]# systemctl status ceph-osd@osd.2.service
● ceph-osd@osd.2.service - Ceph object storage daemon
   Loaded: loaded (/usr/lib/systemd/system/ceph-osd@.service; enabled;
vendor preset: disabled)
   Active: failed (Result: start-limit) since Fri 2016-06-17 14:34:25 IST;
3s ago
  Process: 8112 ExecStart=/usr/bin/ceph-osd -f --cluster ${CLUSTER} --id %i
--setuser ceph --setgroup ceph (code=exited, status=1/FAILURE)
  Process: 8071 ExecStartPre=/usr/lib/ceph/ceph-osd-prestart.sh --cluster
${CLUSTER} --id %i (code=exited, status=0/SUCCESS)
 Main PID: 8112 (code=exited, status=1/FAILURE)

Jun 17 14:34:25 Admin ceph-osd[8112]: --debug_ms N      set message debug
level (e.g. 1)
Jun 17 14:34:25 Admin systemd[1]: ceph-osd@osd.2.service: main process
exited, code=exited, status=...ILURE
Jun 17 14:34:25 Admin systemd[1]: Unit ceph-osd@osd.2.service entered
failed state.
Jun 17 14:34:25 Admin systemd[1]: ceph-osd@osd.2.service failed.
Jun 17 14:34:25 Admin ceph-osd[8112]: 2016-06-17 14:34:25.003696
7f3f58664800 -1 must specify '-i #'...mber
Jun 17 14:34:25 Admin systemd[1]: ceph-osd@osd.2.service holdoff time over,
scheduling restart.
Jun 17 14:34:25 Admin systemd[1]: start request repeated too quickly for
ceph-osd@osd.2.service
Jun 17 14:34:25 Admin systemd[1]: Failed to start Ceph object storage
daemon.
Jun 17 14:34:25 Admin systemd[1]: Unit ceph-osd@osd.2.service entered
failed state.
Jun 17 14:34:25 Admin systemd[1]: ceph-osd@osd.2.service failed.
Hint: Some lines were ellipsized, use -l to show in full.


[root@Admin ceph]# systemctl -a | grep ceph
  ceph-osd@osd.0.service
loaded    inactive dead      Ceph object storage daemon
● ceph-osd@osd.1.service
loaded    failed   failed    Ceph object storage daemon
  ceph-osd@osd.10.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.11.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.12.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.13.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.14.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.15.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.16.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.17.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.18.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.19.service
loaded    inactive dead      Ceph object storage daemon
● ceph-osd@osd.2.service
loaded    failed   failed    Ceph object storage daemon
● ceph-osd@osd.3.service
loaded    failed   failed    Ceph object storage daemon
  ceph-osd@osd.4.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.5.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.6.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.7.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.8.service
loaded    inactive dead      Ceph object storage daemon
  ceph-osd@osd.9.service
loaded    inactive dead      Ceph object storage daemon


On Thu, Jun 16, 2016 at 8:03 PM, Joshua M. Boniface <jos...@boniface.me>
wrote:

> RHEL 7.2 and Jewel should be using the systemd unit files by default, so
> you'd do something like:
>
> > sudo systemctl stop ceph-osd@<OSDID>
>
> and then
>
> > sudo systemctl start ceph-osd@<OSDID>
>
> when you're done.
>
> --
> Joshua M. Boniface
> Linux System Ærchitect
> Sigmentation fault. Core dumped.
>
> On 16/06/16 09:44 AM, Kanchana. P wrote:
> >
> > Hi,
> >
> > How can I down a osd and bring it back in RHEL 7.2 with ceph verison
> 10.2.2
> >
> > sudo start ceph-osd id=1 fails with “sudo: start: command not found”.
> >
> > I have 5 osds in each node and i want to down one particular osd (sudo
> stop ceph-sd id=1 also fails) and see whether replicas are written to other
> osds without any issues.
> >
> > Thanks in advance.
> >
> > –kanchana.
> >
> >
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to