2015-03-27 18:27 GMT+01:00 Gregory Farnum <[email protected]>:
> Ceph has per-pg and per-OSD metadata overhead. You currently have 26000 PGs,
> suitable for use on a cluster of the order of 260 OSDs. You have placed
> almost 7GB of data into it (21GB replicated) and have about 7GB of
> additional overhead.
>
> You might try putting a suitable amount of data into the cluster before
> worrying about the ratio of space used to data stored. :)
> -Greg
Hello Greg,
I put a suitable amount of data now, and it looks like my ratio is still 1 to 5.
The folder:
/var/lib/ceph/osd/ceph-N/current/meta/
did not grow, so it looks like that is not the problem.
Do you have any hint how to troubleshoot this issue ???
ansible@zrh-srv-m-cph02:~$ ceph osd pool get .rgw.buckets size
size: 3
ansible@zrh-srv-m-cph02:~$ ceph osd pool get .rgw.buckets min_size
min_size: 2
ansible@zrh-srv-m-cph02:~$ ceph -w
cluster 4179fcec-b336-41a1-a7fd-4a19a75420ea
health HEALTH_WARN pool .rgw.buckets has too few pgs
monmap e4: 4 mons at
{rml-srv-m-cph01=10.120.50.20:6789/0,rml-srv-m-cph02=10.120.50.21:6789/0,rml-srv-m-stk03=10.120.50.32:6789/0,zrh-srv-m-cph02=10.120.50.2:6789/0},
election epoch 668, quorum 0,1,2,3
zrh-srv-m-cph02,rml-srv-m-cph01,rml-srv-m-cph02,rml-srv-m-stk03
osdmap e2170: 54 osds: 54 up, 54 in
pgmap v619041: 28684 pgs, 15 pools, 109 GB data, 7358 kobjects
518 GB used, 49756 GB / 50275 GB avail
28684 active+clean
ansible@zrh-srv-m-cph02:~$ ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
50275G 49756G 518G 1.03
POOLS:
NAME ID USED %USED MAX AVAIL OBJECTS
rbd 0 155 0 16461G 2
gianfranco 7 156 0 16461G 2
images 8 257M 0 16461G 38
.rgw.root 9 840 0 16461G 3
.rgw.control 10 0 0 16461G 8
.rgw 11 21334 0 16461G 108
.rgw.gc 12 0 0 16461G 32
.users.uid 13 1575 0 16461G 6
.users 14 72 0 16461G 6
.rgw.buckets.index 15 0 0 16461G 30
.users.swift 17 36 0 16461G 3
.rgw.buckets 18 108G 0.22 16461G 7534745
.intent-log 19 0 0 16461G 0
.rgw.buckets.extra 20 0 0 16461G 0
volumes 21 512M 0 16461G 161
ansible@zrh-srv-m-cph02:~$
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com