Hello Developers 

I have encountered some wired output of ceph df command , suddenly


When i was writing some data on cache-pool , and checked its used % , i found 
some used as 8E ( don’t know what is this ) and the used % for cache-pool was 0


# ceph df
GLOBAL:
    SIZE      AVAIL     RAW USED     %RAW USED
    1301T     1284T     17518G       1.31
POOLS:
    NAME                   ID     USED       %USED     OBJECTS
    data                   0      801M       0         2
    metadata               1      801M       0         22
    rbd                    2      0          0         0
    .rgw                   3      3524       0         26
    .rgw.root              4      778        0         3
    .rgw.control           5      0          0         8
    .rgw.buckets           6      8201M      0         2298
    .rgw.buckets.index     7      0          0         13
    .users.swift           8      7          0         1
    volumes                9      1106G      0.08      283387
    images                 10     40960k     0         8
    backups                11     0          0         0
    .rgw.gc                12     0          0         32
    .users.uid             13     848        0         5
    .users                 14     16         0         2
    .log                   15     153k       0         37
                           16     0          0         0
    hpsl4540               21     110G       0         28152
    hpdl380                22     245G       0.02      62688
    EC-2-2                 23     6338G      0.48      4859
    cache-pool             24     8E         0         5849                     
        ## What is the meaning of E here , also please note used % for 
cache-pool is 0 here
    ssd                    25     25196M     0         5464


After some time when cache-poo used value changed to 7E and used % as 644301.19 
 . While there were no objects in the cache-pool


# ceph df
GLOBAL:
    SIZE      AVAIL     RAW USED     %RAW USED
    1301T     1284T     17508G       1.31
POOLS:
    NAME                   ID     USED       %USED         OBJECTS
    data                   0      801M       0             2
    metadata               1      801M       0             22
    rbd                    2      0          0             0
    .rgw                   3      3524       0             26
    .rgw.root              4      778        0             3
    .rgw.control           5      0          0             8
    .rgw.buckets           6      8201M      0             2298
    .rgw.buckets.index     7      0          0             13
    .users.swift           8      7          0             1
    volumes                9      1106G      0.08          283387
    images                 10     40960k     0             8
    backups                11     0          0             0
    .rgw.gc                12     0          0             32
    .users.uid             13     848        0             5
    .users                 14     16         0             2
    .log                   15     153k       0             37
                           16     0          0             0
    hpsl4540               21     110G       0             28152
    hpdl380                22     245G       0.02          62688
    EC-2-2                 23     6338G      0.48          4843
    cache-pool             24     7E         644301.19     1056                 
      ## The used % for cache-pool has become 644301.19 
    ssd                    25     25196M     0             5464
#


# rados -p cache-pool ls                         
#




Is this a bug , if yes , then is is already known. Do you want me to raise a 
bug ticket in tracker.ceph.com ?



****************************************************************
Karan Singh 
Cloud computing group
CSC - IT Center for Science,
Keilaranta 14, P. O. Box 405, FIN-02101 Espoo, Finland
tel. +358 9 4572001
fax +358 9 4572302
http://www.csc.fi/
****************************************************************

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to