Hi Greg,

Sorry I didn't reserve the environment due to urgent needs.

However I think you are right because at that time I just purged all pools and 
re-create them in a short time, thank you very much!

Sincerely,
Craig Chi

On 2016-12-13 14:21, Gregory Farnum<gfar...@redhat.com>wrote:
> On Thu, Dec 1, 2016 at 8:35 AM, Craig Chi<craig...@synology.com>wrote:>Hi 
> list,>>I am testing the Ceph cluster with unpractical pg numbers to do 
> some>experiments.>>But when I use ceph -w to watch my cluster status, I see 
> pg numbers doubled.>From my ceph -w>>root@mon1:~# ceph -w>cluster 
> 1c33bf75-e080-4a70-9fd8-860ff216f595>health HEALTH_WARN>too many PGs per OSD 
> (514>max 300)>noout,noscrub,nodeep-scrub,sortbitwise flag(s) set>monmap e1: 3 
> mons 
> at>{mon1=172.20.1.2:6789/0,mon2=172.20.1.3:6789/0,mon3=172.20.1.4:6789/0}>election
>  epoch 634, quorum 0,1,2 mon1,mon2,mon3>osdmap e48791: 420 osds: 420 up, 420 
> in>flags noout,noscrub,nodeep-scrub,sortbitwise>pgmap v892347: 25600 pgs, 4 
> pools, 14321 GB data, 3579 kobjects>23442 GB used, 3030 TB / 3053 TB 
> avail>25600 active+clean>>2016-12-01 17:26:20.358407 mon.0 [INF] pgmap 
> v892346: 51200 pgs: 51200>active+clean; 16973 GB data, 24609 GB used, 4556 TB 
> / 4580 TB avail>2016-12-01 17:26:22.877765 mon.0 [INF] pgmap v892347: 51200 
> pgs: 51200>a
 ctive+cl

ean; 16973 GB data, 24610 GB used, 4556 TB / 4580 TB avail>>From my ceph osd 
pool ls detail>>pool 81 'vms' replicated size 3 min_size 2 crush_ruleset 0 
object_hash>rjenkins pg_num 512 pgp_num 512 last_change 48503 
flags>hashpspool,nodelete,nopgchange,nosizechange stripe_width 0>pool 82 
'images' replicated size 3 min_size 2 crush_ruleset 0 object_hash>rjenkins 
pg_num 512 pgp_num 512 last_change 48507 
flags>hashpspool,nodelete,nopgchange,nosizechange stripe_width 0>pool 85 
'objects' erasure size 20 min_size 17 crush_ruleset 1 object_hash>rjenkins 
pg_num 8192 pgp_num 8192 last_change 48778 
flags>hashpspool,nodelete,nopgchange,nosizechange stripe_width 4352>pool 86 
'volumes' replicated size 3 min_size 2 crush_ruleset 0 object_hash>rjenkins 
pg_num 16384 pgp_num 16384 last_change 48786 
flags>hashpspool,nodelete,nopgchange,nosizechange stripe_width 0>>I think I 
created 25600 pgs totally, but ceph -s reported 25600 / 51200>randomly. However 
ceph -w always reported 51200 on the latest
  line.>>

If this a kind of bug or just I was doing something wrong? Feel free to let>me 
know if you need more information. Are you still seeing this? It certainly 
sounds like a bug, but I think the output you're seeing could also happen if 
you had some deleted pools which the OSDs hadn't been able to purge yet. -Greg


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to