Thanks, actually I create a pool with more pgs also meet this problem. Following is my crush map, please help point how to change the crush ruleset? thanks.
#begin crush map tunable choose_local_tries 0 tunable choose_local_fallback_tries 0 tunable choose_total_tries 50 tunable chooseleaf_descend_once 1 tunable chooseleaf_vary_r 1 tunable straw_calc_version 1 # devices device 0 device0 device 1 device1 device 2 osd.2 device 3 osd.3 device 4 osd.4 # types type 0 osd type 1 host type 2 chassis type 3 rack type 4 row type 5 pdu type 6 pod type 7 room type 8 datacenter type 9 region type 10 root # buckets host redpower-ceph-01 { id -2 # do not change unnecessarily # weight 3.000 alg straw hash 0 # rjenkins1 item osd.2 weight 1.000 item osd.3 weight 1.000 item osd.4 weight 1.000 } root default { id -1 # do not change unnecessarily # weight 3.000 alg straw hash 0 # rjenkins1 item redpower-ceph-01 weight 3.000 } # rules rule replicated_ruleset { ruleset 0 type replicated min_size 1 max_size 10 step take default step chooseleaf firstn 0 type host step emit } # end crush map 2016-06-22 18:27 GMT+08:00 Burkhard Linke < burkhard.li...@computational.bio.uni-giessen.de>: > Hi, > > On 06/22/2016 12:10 PM, min fang wrote: > > Hi, I created a new ceph cluster, and create a pool, but see "stuck > unclean since forever" errors happen(as the following), can help point out > the possible reasons for this? thanks. > > ceph -s > cluster 602176c1-4937-45fc-a246-cc16f1066f65 > health HEALTH_WARN > 8 pgs degraded > 8 pgs stuck unclean > 8 pgs undersized > too few PGs per OSD (2 < min 30) > monmap e1: 1 mons at {ceph-01=172.0.0.11:6789/0} > election epoch 14, quorum 0 ceph-01 > osdmap e89: 3 osds: 3 up, 3 in > flags > pgmap v310: 8 pgs, 1 pools, 0 bytes data, 0 objects > 60112 MB used, 5527 GB / 5586 GB avail > 8 active+undersized+degraded > > > *snipsnap* > > With three OSDs and a single host you need to change the crush ruleset for > the pool, since it tries to distribute the data across 3 different _host_ > by default. > > Regards, > Burkhard > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com