Hmm, it should be another problem plays. Maybe more logs could explain it. ceph.log ceph-mon.log
On Wed, Apr 30, 2014 at 12:06 PM, Cao, Buddy <[email protected]> wrote: > Thanks your reply, Haomai. What I don't understand is that, why the stuck > unclean pgs keep the same numbers after 12 hours. It's the common behavior or > not? > > > Wei Cao (Buddy) > > -----Original Message----- > From: Haomai Wang [mailto:[email protected]] > Sent: Wednesday, April 30, 2014 11:36 AM > To: Cao, Buddy > Cc: [email protected] > Subject: Re: [ceph-users] mkcephfs questions > > The result of "ceph -s" should tell you the reason. There only exists > 21 OSD up but we need 24 OSDs > > On Wed, Apr 30, 2014 at 11:21 AM, Cao, Buddy <[email protected]> wrote: >> Hi, >> >> >> >> I setup ceph cluster thru mkcephfs command, after I enter “ceph –s”, >> it always returns 4950 stuck unclean pgs. I tried the same “ceph -s” >> after 12 hrs, there still returns the same unclean pgs number, nothing >> changed. >> Does mkcephfs always has the problem or I did something wrong? I >> attached the result of “ceph -s”, “ceph osd tree” and ceph.conf I >> have, please kindly help. >> >> >> >> >> >> [root@ceph]# ceph -s >> >> cluster 99fd4ff8-0fb8-47b9-8179-fefbba1c2503 >> >> health HEALTH_WARN 4950 pgs degraded; 4950 pgs stuck unclean; >> recovery >> 21/42 objects degraded (50.000%); 3/24 in osds are down; clock skew >> detected on mon.1, mon.2 >> >> monmap e1: 3 mons at >> {0=192.168.0.2:6789/0,1=192.168.0.3:6789/0,2=192.168.0.4:6789/0}, >> election epoch 6, quorum 0,1,2 0,1,2 >> >> mdsmap e4: 1/1/1 up {0=0=up:active} >> >> osdmap e6019: 24 osds: 21 up, 24 in >> >> pgmap v16445: 4950 pgs, 6 pools, 9470 bytes data, 21 objects >> >> 4900 MB used, 93118 MB / 98019 MB avail >> >> 21/42 objects degraded (50.000%) >> >> 4950 active+degraded >> >> >> >> [root@ceph]# ceph osd tree //part of returns >> >> # id weight type name up/down reweight >> >> -36 25 root vsm >> >> -31 3.2 storage_group ssd >> >> -16 3 zone zone_a_ssd >> >> -1 1 host vsm2_ssd_zone_a >> >> 2 1 osd.2 up 1 >> >> -6 1 host vsm3_ssd_zone_a >> >> 10 1 osd.10 up 1 >> >> -11 1 host vsm4_ssd_zone_a >> >> 18 1 osd.18 up 1 >> >> -21 0.09999 zone zone_c_ssd >> >> -26 0.09999 zone zone_b_ssd >> >> -33 3.2 storage_group sata >> >> -18 3 zone zone_a_sata >> >> -3 1 host vsm2_sata_zone_a >> >> 1 1 osd.1 up 1 >> >> -8 1 host vsm3_sata_zone_a >> >> 9 1 osd.9 up 1 >> >> -13 1 host vsm4_sata_zone_a >> >> 17 1 osd.17 up 1 >> >> -23 0.09999 zone zone_c_sata >> >> -28 0.09999 zone zone_b_sata >> >> >> >> >> >> Wei Cao (Buddy) >> >> >> >> >> _______________________________________________ >> ceph-users mailing list >> [email protected] >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > > > -- > Best Regards, > > Wheat -- Best Regards, Wheat _______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
