Hi Warren,

a simple:
ceph osd pool set bench2 hashpspool false
solved my problem.

Thank a lot
Giuseppe

2015-10-02 16:18 GMT+02:00 Warren Wang - ISD <[email protected]>:

> You probably don’t want hashpspool automatically set, since your clients
> may still not understand that crush map feature. You can try to unset it
> for that pool and see what happens, or create a new pool without hashpspool
> enabled from the start.  Just a guess.
>
> Warren
>
> From: Giuseppe Civitella <[email protected]<mailto:
> [email protected]>>
> Date: Friday, October 2, 2015 at 10:05 AM
> To: ceph-users <[email protected]<mailto:[email protected]>>
> Subject: [ceph-users] pgs stuck unclean on a new pool despite the pool
> size reconfiguration
>
> Hi all,
> I have a Firefly cluster which has been upgraded from Emperor.
> It has 2 OSD hosts and 3 monitors.
> The cluster has default values for what concerns size and min_size of the
> pools.
> Once upgraded to Firefly, I created a new pool called bench2:
> ceph osd pool create bench2 128 128
> and set its sizes:
> ceph osd pool set bench2 size 2
> ceph osd pool set bench2 min_size 1
>
> this is the state of the pools:
> pool 0 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 1 crash_replay_interval 45
> stripe_width 0
> pool 1 'metadata' replicated size 2 min_size 1 crush_ruleset 1 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 1 stripe_width 0
> pool 2 'rbd' replicated size 2 min_size 1 crush_ruleset 2 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 1 stripe_width 0
> pool 3 'volumes' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 384 pgp_num 384 last_change 2568 stripe_width 0
>         removed_snaps [1~75]
> pool 4 'images' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 384 pgp_num 384 last_change 1895 stripe_width 0
> pool 8 'bench2' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 128 pgp_num 128 last_change 2580 flags hashpspool
> stripe_width 0
>
> despite this I still get a warning about 128 pgs stuck unclean.
> The "ceph health detail" shows me the stuck PGs. So i take one to get the
> involved OSDs:
>
> pg 8.38 is stuck unclean since forever, current state active, last acting
> [22,7]
>
> if I restart the OSD with id 22, the PG 8.38 gets an active+clean state.
>
> This is an incorrect behavior, AFAIK. The cluster should get noticed of
> the new size and min_size values without any manual intervention. So my
> question is: any idea about why this happens and how to restore the default
> behavior? Do I need to restart all of the OSDs to restore an healthy state?
>
> thanks a lot
> Giuseppe
>
> This email and any files transmitted with it are confidential and intended
> solely for the individual or entity to whom they are addressed. If you have
> received this email in error destroy it immediately. *** Walmart
> Confidential ***
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to