On Mon, 27 Jan 2014, Derek Yarnell wrote:
> Hi Sage,
>
> Our clusters are slightly different but no the monitors start just fine.
> On our test and rgw clusters we run monitors co-located with our OSDs.
> The monitors start just fine. My understanding is that when booting
> the hosts detect a disk hot-plug event in udev via the
> /lib/udev/rules.d/95-ceph-osd.rules file. Running the
> '/usr/sbin/ceph-disk-activate /dev/vdb1' command for example in our test
> cluster does the right thing
>
> # /usr/sbin/ceph-disk-activate /dev/vdb1
> === osd.2 ===
> create-or-move updated item name 'osd.2' weight 0.02 at location
> {host=ceph02,root=default} to crush map
> Starting Ceph osd.2 on ceph02...
> starting osd.2 at :/0 osd_data /var/lib/ceph/osd/ceph-2
> /var/lib/ceph/osd/ceph-2/journal
>
>
> Our best guess so far is that this line is not matching the underlying
> disk that is getting hotplugged (95-ceph-osd.rules). Is
> ID_PART_ENTRY_TYPE just the partition UUID or are we not understanding
> identifier correctly?
>
> ENV{ID_PART_ENTRY_TYPE}=="4fbd7e29-9d25-41b8-afd0-062c0ceff05d",
These rules do not work with RHEL6 because it is an old version of udev
and/or blkid (I forget the details now) that doesn't natively expose the
GPT UUIDs. There is a 95-ceph-osd-alt.rules file that should be installed
that kludges around this by shelling out to the ceph-disk-udev helper.
That's the helper to test in this case...
sage
>
> Thanks,
> derek
>
> On 1/27/14, 11:29 AM, Sage Weil wrote:
> > Hi Derek,
> >
> > Would like to get to the bottom of your problem. Is it that the monitors
> > don't start after a reboot? Is there an error in
> > /var/log/ceph/ceph-mon.`hostname`.log?
> >
> > sage
> >
> > On Mon, 27 Jan 2014, Derek Yarnell wrote:
> >
> >> Hi,
> >>
> >> Would I take this to understand that this may be a known issue with udev
> >> on RHEL then? We will for now add them to the fstab.
> >>
> >> Thanks,
> >> derek
> >>
> >> On 1/25/14, 9:23 PM, Michael J. Kidd wrote:
> >>> While clearly not optimal for long term flexibility, I've found that
> >>> adding my OSD's to fstab allows the OSDs to mount during boot, and they
> >>> start automatically when they're already mounted during boot.
> >>>
> >>> Hope this helps until a permanent fix is available.
> >>>
> >>> Michael J. Kidd
> >>> Sr. Storage Consultant
> >>> Inktank Professional Services
> >>>
> >>>
> >>> On Fri, Jan 24, 2014 at 9:08 PM, Derek Yarnell <[email protected]
> >>> <mailto:[email protected]>> wrote:
> >>>
> >>> So we have a test cluster, and two production clusters all running on
> >>> RHEL6.5. Two are running Emperor and one of them running Dumpling.
> >>> On
> >>> all of them our OSDs do not start at boot it seems via the udev rules.
> >>> The OSDs were created with ceph-deploy and are all GPT. The OSDs are
> >>> visable with `ceph-disk list` and running
> >>> `/usr/sbin/ceph-disk-activate
> >>> {device}` mounts and adds them. Running a `partprobe {device}` does
> >>> not
> >>> seem to trigger the udev rule at all.
> >>>
> >>> I had found this issue[1] but we are definitely running code that was
> >>> released after this ticket was closed. Has there been anyone else
> >>> that
> >>> has problems with udev on RHEL mounting their OSDs?
> >>>
> >>> [1] - http://tracker.ceph.com/issues/5194
> >>>
> >>> Thanks,
> >>> derek
> >>>
> >>> --
> >>> Derek T. Yarnell
> >>> University of Maryland
> >>> Institute for Advanced Computer Studies
> >>> _______________________________________________
> >>> ceph-users mailing list
> >>> [email protected] <mailto:[email protected]>
> >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >>>
> >>>
> >>
> >> --
> >> Derek T. Yarnell
> >> University of Maryland
> >> Institute for Advanced Computer Studies
> >> _______________________________________________
> >> ceph-users mailing list
> >> [email protected]
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >>
> >>
> >
>
> --
> Derek T. Yarnell
> University of Maryland
> Institute for Advanced Computer Studies
>
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com