[Re-adding the list.] When the OSDs crash they will print out to their log a short description of what happened, with a bunch of function names.
Unfortunately the problem you've run into is probably non-trivial to solve as you've introduced a bit of a weird situation into the permanent record that your OSDs need to process. I've created a bug (http://tracker.ceph.com/issues/4699), you can follow that. :) -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Wednesday, April 10, 2013 at 8:57 AM, Witalij Poljatchek wrote: > there are no data. > > Plain OSDs > > > What you mean backtrace ? strace of ceph-osd process ? > > > is easy to reproduce. > > setup plain cluster > > and then set: > > ceph osd pool set rbd size 0 > > after minute set: > > ceph osd pool set rbd size 2 > > that all. > > > > > > > On 04/10/2013 05:24 PM, Gregory Farnum wrote: > > Sounds like they aren't handling the transition very well when trying to > > calculate old OSDs which might have held the PG. Are you trying to salvage > > the data that was in it, or can you throw it away? > > Can you post the backtrace they're producing? > > -Greg > > Software Engineer #42 @ http://inktank.com | http://ceph.com > > > > > > On Wednesday, April 10, 2013 at 3:59 AM, Witalij Poljatchek wrote: > > > > > Hello, > > > > > > need help to solve segfault on all osd in my test cluster. > > > > > > > > > Setup ceph from scratch. > > > service ceph -a start > > > > > > ceph -w > > > health HEALTH_OK > > > monmap e1: 3 mons at > > > {1=10.200.20.1:6789/0,2=10.200.20.2:6789/0,3=10.200.20.3:6789/0}, > > > election epoch 6, quorum 0,1,2 1,2,3 > > > osdmap e5: 4 osds: 4 up, 4 in > > > pgmap v305: 960 pgs: 960 active+clean; 0 bytes data, 40147 MB used, 26667 > > > GB / 26706 GB avail > > > mdsmap e1: 0/0/1 up > > > > > > > > > if i set replica size to 0 "i know this make no sense" > > > ceph osd pool set rbd size 0 > > > and then back to 2 > > > ceph osd pool set rbd size 2 > > > > > > then i see that on all OSDs the process ceph-osd crash with segfault > > > > > > If i stop MONs daemons then i can start OSDs but if i start MONs back > > > then die all OSDs again. > > > > > > > > > > > > How i cann repair this behavior ? > > > > > > > > > > > > > > > > > > My setup > > > Nothing specials: > > > > > > Centos 6.3 > > > > > > Kernel: 3.8.3-1.el6.elrepo.x86_64 > > > > > > ceph-fuse-0.56.4-0.el6.x86_64 > > > ceph-test-0.56.4-0.el6.x86_64 > > > libcephfs1-0.56.4-0.el6.x86_64 > > > ceph-0.56.4-0.el6.x86_64 > > > ceph-release-1-0.el6.noarch > > > > > > cat /etc/ceph/ceph.conf > > > > > > [global] > > > auth cluster required = none > > > auth service required = none > > > auth client required = none > > > keyring = /etc/ceph/$name.keyring > > > [mon] > > > [mds] > > > [osd] > > > osd journal size = 10000 > > > [mon.1] > > > host = ceph-mon1 > > > mon addr = 10.200.20.1:6789 > > > [mon.2] > > > host = ceph-mon2 > > > mon addr = 10.200.20.2:6789 > > > [mon.3] > > > host = ceph-mon3 > > > mon addr = 10.200.20.3:6789 > > > > > > [osd.0] > > > host = ceph-osd1 > > > [osd.1] > > > host = ceph-osd2 > > > [osd.2] > > > host = ceph-osd3 > > > [osd.3] > > > host = ceph-osd4 > > > > > > [mds.a] > > > host = ceph-mds1 > > > [mds.b] > > > host = ceph-mds2 > > > [mds.c] > > > host = ceph-mds3 > > > > > > Thanks much. > > > -- AIXIT GmbH - Witalij Poljatchek (T) +49 69 203 4709-13 - (F) +49 69 > > > 203 470 979 [email protected] - http://www.aixit.com AIXIT GmbH > > > Strahlenbergerstr. 14 63067 Offenbach am Main (T) +49 69 203 470 913 > > > Amtsgericht Offenbach, HRB 43953 Geschäftsführer: Friedhelm Heyer, Holger > > > Grauer > > > > > > -- > AIXIT GmbH - Witalij Poljatchek > (T) +49 69 203 4709-13 - (F) +49 69 203 470 979 > [email protected] - http://www.aixit.com > > AIXIT GmbH > > Strahlenbergerstr. 14 > 63067 Offenbach am Main > (T) +49 69 203 470 913 > > Amtsgericht Offenbach, HRB 43953 > Geschäftsführer: Friedhelm Heyer, Holger Grauer _______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
