Thanks, actually I create a pool with more pgs also meet this problem.
Following is my crush map, please help point how to change the crush
ruleset? thanks.

#begin crush map
tunable choose_local_tries 0
tunable choose_local_fallback_tries 0
tunable choose_total_tries 50
tunable chooseleaf_descend_once 1
tunable chooseleaf_vary_r 1
tunable straw_calc_version 1

# devices
device 0 device0
device 1 device1
device 2 osd.2
device 3 osd.3
device 4 osd.4

# types
type 0 osd
type 1 host
type 2 chassis
type 3 rack
type 4 row
type 5 pdu
type 6 pod
type 7 room
type 8 datacenter
type 9 region
type 10 root

# buckets
host redpower-ceph-01 {
        id -2           # do not change unnecessarily
        # weight 3.000
        alg straw
        hash 0  # rjenkins1
        item osd.2 weight 1.000
        item osd.3 weight 1.000
        item osd.4 weight 1.000
}
root default {
        id -1           # do not change unnecessarily
        # weight 3.000
        alg straw
        hash 0  # rjenkins1
        item redpower-ceph-01 weight 3.000
}

# rules
rule replicated_ruleset {
        ruleset 0
        type replicated
        min_size 1
        max_size 10
        step take default
        step chooseleaf firstn 0 type host
        step emit
}

# end crush map


2016-06-22 18:27 GMT+08:00 Burkhard Linke <
burkhard.li...@computational.bio.uni-giessen.de>:

> Hi,
>
> On 06/22/2016 12:10 PM, min fang wrote:
>
> Hi, I created a new ceph cluster, and create a pool, but see "stuck
> unclean since forever" errors happen(as the following), can help point out
> the possible reasons for this? thanks.
>
> ceph -s
>     cluster 602176c1-4937-45fc-a246-cc16f1066f65
>      health HEALTH_WARN
>             8 pgs degraded
>             8 pgs stuck unclean
>             8 pgs undersized
>             too few PGs per OSD (2 < min 30)
>      monmap e1: 1 mons at {ceph-01=172.0.0.11:6789/0}
>             election epoch 14, quorum 0 ceph-01
>      osdmap e89: 3 osds: 3 up, 3 in
>             flags
>       pgmap v310: 8 pgs, 1 pools, 0 bytes data, 0 objects
>             60112 MB used, 5527 GB / 5586 GB avail
>                    8 active+undersized+degraded
>
>
> *snipsnap*
>
> With three OSDs and a single host you need to change the crush ruleset for
> the pool, since it tries to distribute the data across 3 different _host_
> by default.
>
> Regards,
> Burkhard
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to