[Please keep all mail on the list.]  

Hmm, that OSD log doesn't show a crash. I thought you said they were all 
crashing? Do they come up okay when you turn them back on again?
-Greg

Software Engineer #42 @ http://inktank.com | http://ceph.com


On Wednesday, April 10, 2013 at 9:27 AM, Witalij Poljatchek wrote:

> the log files.
>  
> thank you ! :)
>  
> On 04/10/2013 06:06 PM, Gregory Farnum wrote:
> > [Re-adding the list.]
> >  
> > When the OSDs crash they will print out to their log a short description of 
> > what happened, with a bunch of function names.
> >  
> > Unfortunately the problem you've run into is probably non-trivial to solve 
> > as you've introduced a bit of a weird situation into the permanent record 
> > that your OSDs need to process. I've created a bug 
> > (http://tracker.ceph.com/issues/4699), you can follow that. :)
> > -Greg
> > Software Engineer #42 @ http://inktank.com | http://ceph.com
> >  
> >  
> > On Wednesday, April 10, 2013 at 8:57 AM, Witalij Poljatchek wrote:
> >  
> > > there are no data.
> > >  
> > > Plain OSDs
> > >  
> > >  
> > > What you mean backtrace ? strace of ceph-osd process ?
> > >  
> > >  
> > > is easy to reproduce.
> > >  
> > > setup plain cluster
> > >  
> > > and then set:
> > >  
> > > ceph osd pool set rbd size 0
> > >  
> > > after minute set:
> > >  
> > > ceph osd pool set rbd size 2
> > >  
> > > that all.
> > >  
> > >  
> > >  
> > >  
> > >  
> > >  
> > > On 04/10/2013 05:24 PM, Gregory Farnum wrote:
> > > > Sounds like they aren't handling the transition very well when trying 
> > > > to calculate old OSDs which might have held the PG. Are you trying to 
> > > > salvage the data that was in it, or can you throw it away?
> > > > Can you post the backtrace they're producing?
> > > > -Greg
> > > > Software Engineer #42 @ http://inktank.com | http://ceph.com
> > > >  
> > > >  
> > > > On Wednesday, April 10, 2013 at 3:59 AM, Witalij Poljatchek wrote:
> > > >  
> > > > > Hello,
> > > > >  
> > > > > need help to solve segfault on all osd in my test cluster.
> > > > >  
> > > > >  
> > > > > Setup ceph from scratch.
> > > > > service ceph -a start
> > > > >  
> > > > > ceph -w
> > > > > health HEALTH_OK
> > > > > monmap e1: 3 mons at 
> > > > > {1=10.200.20.1:6789/0,2=10.200.20.2:6789/0,3=10.200.20.3:6789/0}, 
> > > > > election epoch 6, quorum 0,1,2 1,2,3
> > > > > osdmap e5: 4 osds: 4 up, 4 in
> > > > > pgmap v305: 960 pgs: 960 active+clean; 0 bytes data, 40147 MB used, 
> > > > > 26667 GB / 26706 GB avail
> > > > > mdsmap e1: 0/0/1 up
> > > > >  
> > > > >  
> > > > > if i set replica size to 0 "i know this make no sense"
> > > > > ceph osd pool set rbd size 0
> > > > > and then back to 2
> > > > > ceph osd pool set rbd size 2
> > > > >  
> > > > > then i see that on all OSDs the process ceph-osd crash with segfault
> > > > >  
> > > > > If i stop MONs daemons then i can start OSDs but if i start MONs back 
> > > > > then die all OSDs again.
> > > > >  
> > > > >  
> > > > >  
> > > > > How i cann repair this behavior ?
> > > > >  
> > > > >  
> > > > >  
> > > > >  
> > > > >  
> > > > > My setup
> > > > > Nothing specials:
> > > > >  
> > > > > Centos 6.3
> > > > >  
> > > > > Kernel: 3.8.3-1.el6.elrepo.x86_64
> > > > >  
> > > > > ceph-fuse-0.56.4-0.el6.x86_64
> > > > > ceph-test-0.56.4-0.el6.x86_64
> > > > > libcephfs1-0.56.4-0.el6.x86_64
> > > > > ceph-0.56.4-0.el6.x86_64
> > > > > ceph-release-1-0.el6.noarch
> > > > >  
> > > > > cat /etc/ceph/ceph.conf
> > > > >  
> > > > > [global]
> > > > > auth cluster required = none
> > > > > auth service required = none
> > > > > auth client required = none
> > > > > keyring = /etc/ceph/$name.keyring
> > > > > [mon]
> > > > > [mds]
> > > > > [osd]
> > > > > osd journal size = 10000
> > > > > [mon.1]
> > > > > host = ceph-mon1
> > > > > mon addr = 10.200.20.1:6789
> > > > > [mon.2]
> > > > > host = ceph-mon2
> > > > > mon addr = 10.200.20.2:6789
> > > > > [mon.3]
> > > > > host = ceph-mon3
> > > > > mon addr = 10.200.20.3:6789
> > > > >  
> > > > > [osd.0]
> > > > > host = ceph-osd1
> > > > > [osd.1]
> > > > > host = ceph-osd2
> > > > > [osd.2]
> > > > > host = ceph-osd3
> > > > > [osd.3]
> > > > > host = ceph-osd4
> > > > >  
> > > > > [mds.a]
> > > > > host = ceph-mds1
> > > > > [mds.b]
> > > > > host = ceph-mds2
> > > > > [mds.c]
> > > > > host = ceph-mds3
> > > > >  
> > > > > Thanks much.
> > > > > -- AIXIT GmbH - Witalij Poljatchek (T) +49 69 203 4709-13 - (F) +49 
> > > > > 69 203 470 979 [email protected] - http://www.aixit.com AIXIT GmbH 
> > > > > Strahlenbergerstr. 14 63067 Offenbach am Main (T) +49 69 203 470 913 
> > > > > Amtsgericht Offenbach, HRB 43953 Geschäftsführer: Friedhelm Heyer, 
> > > > > Holger Grauer
> > > >  
> > >  
> > >  
> > >  
> > >  
> > >  
> > > --
> > > AIXIT GmbH - Witalij Poljatchek
> > > (T) +49 69 203 4709-13 - (F) +49 69 203 470 979
> > > [email protected] - http://www.aixit.com
> > >  
> > > AIXIT GmbH
> > >  
> > > Strahlenbergerstr. 14
> > > 63067 Offenbach am Main
> > > (T) +49 69 203 470 913
> > >  
> > > Amtsgericht Offenbach, HRB 43953
> > > Geschäftsführer: Friedhelm Heyer, Holger Grauer
> >  
>  
>  
>  
> --  
> AIXIT GmbH - Witalij Poljatchek
> (T) +49 69 203 4709-13 - (F) +49 69 203 470 979
> [email protected] - http://www.aixit.com
>  
> AIXIT GmbH
>  
> Strahlenbergerstr. 14
> 63067 Offenbach am Main
> (T) +49 69 203 470 913
>  
> Amtsgericht Offenbach, HRB 43953
> Geschäftsführer: Friedhelm Heyer, Holger Grauer
>  
>  
> Attachments:  
> - ceph-osd.0.log
>  
> - ceph-mon.1.log
>  
> - ceph.log
>  



_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to