Hello Developers
I have encountered some wired output of ceph df command , suddenly
When i was writing some data on cache-pool , and checked its used % , i found
some used as 8E ( don’t know what is this ) and the used % for cache-pool was 0
# ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
1301T 1284T 17518G 1.31
POOLS:
NAME ID USED %USED OBJECTS
data 0 801M 0 2
metadata 1 801M 0 22
rbd 2 0 0 0
.rgw 3 3524 0 26
.rgw.root 4 778 0 3
.rgw.control 5 0 0 8
.rgw.buckets 6 8201M 0 2298
.rgw.buckets.index 7 0 0 13
.users.swift 8 7 0 1
volumes 9 1106G 0.08 283387
images 10 40960k 0 8
backups 11 0 0 0
.rgw.gc 12 0 0 32
.users.uid 13 848 0 5
.users 14 16 0 2
.log 15 153k 0 37
16 0 0 0
hpsl4540 21 110G 0 28152
hpdl380 22 245G 0.02 62688
EC-2-2 23 6338G 0.48 4859
cache-pool 24 8E 0 5849
## What is the meaning of E here , also please note used % for
cache-pool is 0 here
ssd 25 25196M 0 5464
After some time when cache-poo used value changed to 7E and used % as 644301.19
. While there were no objects in the cache-pool
# ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
1301T 1284T 17508G 1.31
POOLS:
NAME ID USED %USED OBJECTS
data 0 801M 0 2
metadata 1 801M 0 22
rbd 2 0 0 0
.rgw 3 3524 0 26
.rgw.root 4 778 0 3
.rgw.control 5 0 0 8
.rgw.buckets 6 8201M 0 2298
.rgw.buckets.index 7 0 0 13
.users.swift 8 7 0 1
volumes 9 1106G 0.08 283387
images 10 40960k 0 8
backups 11 0 0 0
.rgw.gc 12 0 0 32
.users.uid 13 848 0 5
.users 14 16 0 2
.log 15 153k 0 37
16 0 0 0
hpsl4540 21 110G 0 28152
hpdl380 22 245G 0.02 62688
EC-2-2 23 6338G 0.48 4843
cache-pool 24 7E 644301.19 1056
## The used % for cache-pool has become 644301.19
ssd 25 25196M 0 5464
#
# rados -p cache-pool ls
#
Is this a bug , if yes , then is is already known. Do you want me to raise a
bug ticket in tracker.ceph.com ?
****************************************************************
Karan Singh
Cloud computing group
CSC - IT Center for Science,
Keilaranta 14, P. O. Box 405, FIN-02101 Espoo, Finland
tel. +358 9 4572001
fax +358 9 4572302
http://www.csc.fi/
****************************************************************
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com