Try " osd crush chooseleaf type = 0" in /etc/ceph/<clustername>.conf
Regards, CY. -----Original Message----- From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Wah Peng Sent: 2015年10月29日 9:14 To: Robert LeBlanc Cc: Lindsay Mathieson; Gurjar, Unmesh; ceph-users@lists.ceph.com Subject: Re: [ceph-users] creating+incomplete issues wow this sounds hard to me. can you show the details? thanks a lot. On 2015/10/29 星期四 9:01, Robert LeBlanc wrote: > You need to change the CRUSH map to select osd instead of host. > > Robert LeBlanc > > Sent from a mobile device please excuse any typos. > > On Oct 28, 2015 7:00 PM, "Wah Peng" <wah_p...@yahoo.com.sg > <mailto:wah_p...@yahoo.com.sg>> wrote: > > $ ceph osd tree > # id weight type name up/down reweight > -1 0.24 root default > -2 0.24 host ceph2 > 0 0.07999 osd.0 up 1 > 1 0.07999 osd.1 up 1 > 2 0.07999 osd.2 up 1 > > > On 2015/10/29 星期四 8:55, Robert LeBlanc wrote: > > Please paste 'ceph osd tree'. > > Robert LeBlanc > > Sent from a mobile device please excuse any typos. > > On Oct 28, 2015 6:54 PM, "Wah Peng" <wah_p...@yahoo.com.sg > <mailto:wah_p...@yahoo.com.sg> > <mailto:wah_p...@yahoo.com.sg <mailto:wah_p...@yahoo.com.sg>>> > wrote: > > Hello, > > Just did it, but still no good health. can you help? thanks. > > ceph@ceph:~/my-cluster$ ceph osd stat > osdmap e24: 3 osds: 3 up, 3 in > > ceph@ceph:~/my-cluster$ ceph health > HEALTH_WARN 89 pgs degraded; 67 pgs incomplete; 67 pgs stuck > inactive; 192 pgs stuck unclean > > > On 2015/10/29 星期四 8:38, Lindsay Mathieson wrote: > > > On 29 October 2015 at 10:29, Wah Peng > <wah_p...@yahoo.com.sg <mailto:wah_p...@yahoo.com.sg> > <mailto:wah_p...@yahoo.com.sg > <mailto:wah_p...@yahoo.com.sg>> > <mailto:wah_p...@yahoo.com.sg > <mailto:wah_p...@yahoo.com.sg> <mailto:wah_p...@yahoo.com.sg > <mailto:wah_p...@yahoo.com.sg>>>> > wrote: > > $ ceph osd stat > osdmap e18: 2 osds: 2 up, 2 in > > this is what it shows. > does it mean I need to add up to 3 osds? I just > use the > default setup. > > > If you went with the defaults then your pool size will > be 3, > meaning it > needs 3 copies of the data (replica 3) to be valid - as > you only > have > two nodes/osd's that can never happen :) > > Your options are: > - Add another node and osd. > or > - reduce the size to 2.(ceph osd set <poolname> size > 2) > > > > -- > Lindsay > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> > <mailto:ceph-users@lists.ceph.com > <mailto:ceph-users@lists.ceph.com>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com