My ceph-deploy came from the download.ceph.com site and it is 1.5.31-0. This 
code is in ceph itself though, the deploy logic is where the code appears to do 
the right thing ;-)

Steve

> On Mar 15, 2016, at 2:38 PM, Vasu Kulkarni <[email protected]> wrote:
> 
> Thanks for the steps that should be enough to test it out, I hope you got the 
> latest ceph-deploy either from pip or throught github.
> 
> On Tue, Mar 15, 2016 at 12:29 PM, Stephen Lord <[email protected]> wrote:
> I would have to nuke my cluster right now, and I do not have a spare one..
> 
> The procedure though is literally this, given a 3 node redhat 7.2 cluster, 
> ceph00, ceph01 and ceph02
> 
> ceph-deploy install --testing ceph00 ceph01 ceph02
> ceph-deploy new ceph00 ceph01 ceph02
> 
> ceph-deploy mon create  ceph00 ceph01 ceph02
> ceph-deploy gatherkeys  ceph00
> 
> ceph-deploy osd create ceph00:sdb:/dev/sdi
> ceph-deploy osd create ceph00:sdc:/dev/sdi
> 
> All devices have their partition tables wiped before this. They are all just 
> SATA devices, no special devices in the way.
> 
> sdi is an ssd and it is being carved up for journals. The first osd create 
> works, the second one gets stuck in a loop in the update_partition call in 
> ceph_disk for the 5 iterations before it gives up. When I look in 
> /sys/block/sdi the partition for the first osd is visible, the one for the 
> second is not. However looking at /proc/partitions it sees the correct thing. 
> So something about partprobe is not kicking udev into doing the right thing 
> when the second partition is added I suspect.
> 
> If I do not use the separate journal device then it usually works, but 
> occasionally I see a single retry in that same loop.
> 
> There is code in ceph_deploy which uses partprobe or partx depending on which 
> distro it detects, that is how I worked out what to change here.
> 
> If I have to tear things down again I will reproduce and post here.
> 
> Steve
> 
> > On Mar 15, 2016, at 2:12 PM, Vasu Kulkarni <[email protected]> wrote:
> >
> > Do you mind giving the full failed logs somewhere in fpaste.org along with 
> > some os version details?
> >  There are some known issues on RHEL,  If you use 'osd prepare' and 'osd 
> > activate'(specifying just the journal partition here) it might work better.
> >
> > On Tue, Mar 15, 2016 at 12:05 PM, Stephen Lord <[email protected]> 
> > wrote:
> > Not multipath if you mean using the multipath driver, just trying to setup 
> > OSDs which use a data disk and a journal ssd. If I run just a disk based 
> > OSD and only specify one device to ceph-deploy then it usually works 
> > although sometimes has to retry. In the case where I am using it to carve 
> > an SSD into several partitions for journals it fails on the second one.
> >
> > Steve
> >
> 
> 
> ----------------------------------------------------------------------
> The information contained in this transmission may be confidential. Any 
> disclosure, copying, or further distribution of confidential information is 
> not permitted unless such privilege is explicitly granted in writing by 
> Quantum. Quantum reserves the right to have electronic communications, 
> including email and attachments, sent across its networks filtered through 
> anti virus and spam software programs and retain such messages in order to 
> comply with applicable data security and retention requirements. Quantum is 
> not responsible for the proper and complete transmission of the substance of 
> this communication or for any delay in its receipt.
> 


_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to