This isn't a solution to fix them not starting at boot time, but a fix to
not having to reboot the node again.  `ceph-disk activate-all` should go
through and start up the rest of your osds without another reboot.

On Wed, Aug 23, 2017 at 9:36 AM Sean Purdy <[email protected]> wrote:

> Hi,
>
> Luminous 12.1.1
>
> I've had a couple of servers where at cold boot time, one or two of the
> OSDs haven't mounted/been detected.  Or been partially detected.  These are
> luminous Bluestore OSDs.  Often a warm boot fixes it, but I'd rather not
> have to reboot the node again.
>
> Sometimes /var/lib/ceph/osd/ceph-NN is empty - i.e. not mounted.  And
> sometimes /var/lib/ceph/osd/ceph-NN is mounted, but the
> /var/lib/ceph/osd/ceph-NN/block symlink is pointing to a /dev/mapper UUID
> path that doesn't exist.  Those partitions have to be mounted before
> "systemctl start [email protected]" will work.
>
> What happens at disk detect and mount time?  Is there a timeout somewhere
> I can extend?
>
> How can I tell udev to have another go at mounting the disks?
>
> If it's in the docs and I've missed it, apologies.
>
>
> Thanks in advance,
>
> Sean Purdy
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to