Ok our last shoot is buy NVME PCIe cards for index pool and dedicate to it.

Checking how many GB/TB are needed for the pool is not clean by ceph df show 0:

    default.rgw.buckets.index      38         0 B         0       1.8 TiB       
   1056

Any idea for 200M Objects?



-----Mensaje original-----
De: Casey Bodley <cbod...@redhat.com> 
Enviado el: martes, 7 de mayo de 2019 19:13
Para: EDH - Manuel Rios Fernandez <mrios...@easydatahost.com>; 
ceph-users@lists.ceph.com
Asunto: Re: [ceph-users] Ceph Bucket strange issues rgw.none + id and marker 
diferent.


On 5/7/19 11:24 AM, EDH - Manuel Rios Fernandez wrote:
> Hi Casey
>
> ceph version 13.2.5 (cbff874f9007f1869bfd3821b7e33b2a6ffd4988) mimic
> (stable)
>
> Reshard is something than don’t allow us customer to list index?

Reshard does not prevent buckets from being listed, it just spreads the index 
over more rados objects (so more osds). Bucket sharding does have an impact on 
listing performance though, because each request to list the bucket has to read 
from every shard of the bucket index in order to sort the entries. If any of 
those osds have performance issues or slow requests, that would slow down all 
bucket listings.

> Regards
>
>
> -----Mensaje original-----
> De: ceph-users <ceph-users-boun...@lists.ceph.com> En nombre de Casey 
> Bodley Enviado el: martes, 7 de mayo de 2019 17:07
> Para: ceph-users@lists.ceph.com
> Asunto: Re: [ceph-users] Ceph Bucket strange issues rgw.none + id and 
> marker diferent.
>
> When the bucket id is different than the bucket marker, that indicates 
> the bucket has been resharded. Bucket stats shows 128 shards, which is 
> reasonable for that object count. The rgw.none category in bucket 
> stats is nothing to worry about.
>
> What ceph version is this? This reminds me of a fix in 
> https://github.com/ceph/ceph/pull/23940, which I now see never got its 
> backports to mimic or luminous. :(
>
> On 5/7/19 10:20 AM, EDH - Manuel Rios Fernandez wrote:
>> Hi Ceph’s
>>
>> We got an issue that we’re still looking the cause after more than 60 
>> hour searching a misconfiguration.
>>
>> After cheking a lot of documentation and Questions&Answer we find 
>> that bucket id and bucket marker are not the same. We compared all 
>> our other bucket and all got the same id and marker.
>>
>> Also found some bucket with the rgw.none section an another not.
>>
>> This bucket is unable to be listed in a fashionable time. Customer 
>> relaxed usage from 120TB to 93TB , from 7Million objects to 5.8M.
>>
>> We isolated a single petition in a RGW server and check some metric, 
>> just try to list this bucket generate 2-3Gbps traffic from RGW to 
>> OSD/MON’s.
>>
>> I asked at IRC if there’re any problem about index pool be in other 
>> root in the same site at crushmap and we think that shouldn’t be.
>>
>> Any idea or suggestion, however crazy, will be proven.
>>
>> Our relevant configuration that may help :
>>
>> CEPH DF:
>>
>> ceph df
>>
>> GLOBAL:
>>
>>      SIZE AVAIL       RAW USED     %RAW USED
>>
>>      684 TiB     139 TiB      545 TiB         79.70
>>
>> POOLS:
>>
>>      NAME               ID     USED        %USED     MAX AVAIL 
>> OBJECTS
>>
>> volumes                        21     3.3 TiB 63.90       1.9 TiB
>> 831300
>>
>> backups                        22         0 B 0       1.9 TiB
>> 0
>>
>>      images                     23     1.8 TiB     49.33       1.9
> TiB
>> 237066
>>
>> vms                            24     3.4 TiB 64.85       1.9 TiB
>> 811534
>>
>> openstack-volumes-archive      25      30 TiB 47.92        32 TiB
>> 7748864
>>
>> .rgw.root                      26     1.6 KiB 0       1.9 TiB
>> 4
>>
>> default.rgw.control            27         0 B 0       1.9 TiB
>> 100
>>
>> default.rgw.data.root          28      56 KiB 0       1.9 TiB
>> 186
>>
>> default.rgw.gc                 29         0 B 0       1.9 TiB
>> 32
>>
>> default.rgw.log                30         0 B 0       1.9 TiB
>> 175
>>
>> default.rgw.users.uid          31     4.9 KiB 0       1.9 TiB
>>        26
>>
>> default.rgw.users.email        36        12 B 0       1.9 TiB
>> 1
>>
>> default.rgw.users.keys         37       243 B 0       1.9 TiB
>> 14
>>
>> default.rgw.buckets.index      38         0 B 0       1.9 TiB
>> 1056
>>
>> default.rgw.buckets.data       39     245 TiB 93.84        16 TiB
>> 102131428
>>
>> default.rgw.buckets.non-ec     40         0 B 0       1.9 TiB
>> 23046
>>
>> default.rgw.usage              43         0 B         0    1.9
> TiB
>> 6
>>
>> CEPH OSD Distribution:
>>
>> ceph osd tree
>>
>> ID  CLASS   WEIGHT TYPE NAME                 STATUS REWEIGHT PRI-AFF
>>
>> -41         654.84045 root archive
>>
>> -37 130.96848     host CEPH-ARCH-R03-07
>>
>> 100 archive 10.91399         osd.100               up  1.00000 
>> 1.00000
>>
>> 101 archive 10.91399         osd.101               up  1.00000 
>> 1.00000
>>
>> 102 archive 10.91399         osd.102               up  1.00000 
>> 1.00000
>>
>> 103 archive 10.91399         osd.103               up  1.00000 
>> 1.00000
>>
>> 104 archive 10.91399         osd.104               up  1.00000 
>> 1.00000
>>
>> 105 archive 10.91399         osd.105               up  1.00000 
>> 1.00000
>>
>> 106 archive 10.91409         osd.106               up  1.00000 
>> 1.00000
>>
>> 107 archive 10.91409         osd.107               up  1.00000 
>> 1.00000
>>
>> 108 archive 10.91409         osd.108               up  1.00000 
>> 1.00000
>>
>> 109 archive 10.91409         osd.109               up  1.00000 
>> 1.00000
>>
>> 110 archive 10.91409         osd.110               up  1.00000 
>> 1.00000
>>
>> 111 archive 10.91409         osd.111               up  1.00000 
>> 1.00000
>>
>> -23 130.96800     host CEPH005
>>
>>    4 archive 10.91399         osd.4                 up  1.00000 
>> 1.00000
>>
>> 41 archive 10.91399         osd.41                up  1.00000 1.00000
>>
>> 74 archive 10.91399         osd.74                up  1.00000 1.00000
>>
>> 75 archive 10.91399         osd.75                up  1.00000 1.00000
>>
>> 81 archive 10.91399         osd.81                up  1.00000 1.00000
>>
>> 82 archive 10.91399         osd.82                up  1.00000 1.00000
>>
>> 83 archive 10.91399         osd.83                up  1.00000 1.00000
>>
>> 84 archive 10.91399         osd.84                up  1.00000 1.00000
>>
>> 85 archive 10.91399         osd.85                up  1.00000 1.00000
>>
>> 86 archive 10.91399         osd.86                up  1.00000 1.00000
>>
>> 87 archive 10.91399         osd.87                up  1.00000 1.00000
>>
>> 88 archive 10.91399         osd.88                up  1.00000 1.00000
>>
>> -17 130.96800     host CEPH006
>>
>>    7 archive 10.91399         osd.7                 up  1.00000 
>> 1.00000
>>
>>    8 archive 10.91399         osd.8                 up  1.00000 
>> 1.00000
>>
>>    9 archive 10.91399         osd.9                 up  1.00000 
>> 1.00000
>>
>> 10 archive 10.91399         osd.10                up  1.00000 1.00000
>>
>> 12 archive 10.91399         osd.12                up  1.00000 1.00000
>>
>> 13 archive 10.91399         osd.13                up  1.00000 1.00000
>>
>> 42 archive 10.91399         osd.42                up  1.00000 1.00000
>>
>> 43 archive 10.91399         osd.43                up  1.00000 1.00000
>>
>> 51 archive 10.91399         osd.51                up  1.00000 1.00000
>>
>> 53 archive 10.91399         osd.53                up  1.00000 1.00000
>>
>> 76 archive 10.91399         osd.76                up  1.00000 1.00000
>>
>> 80 archive 10.91399         osd.80                up  1.00000 1.00000
>>
>> -26 130.96800     host CEPH007
>>
>> 14 archive 10.91399         osd.14                up  1.00000 1.00000
>>
>> 15 archive 10.91399         osd.15                up  1.00000 1.00000
>>
>> 16 archive 10.91399         osd.16                up  1.00000 1.00000
>>
>> 39 archive 10.91399         osd.39                up  1.00000 1.00000
>>
>> 40 archive 10.91399         osd.40                up  1.00000 1.00000
>>
>> 44 archive 10.91399         osd.44                up  1.00000 1.00000
>>
>> 48 archive 10.91399         osd.48                up  1.00000 1.00000
>>
>> 49 archive 10.91399         osd.49                up  1.00000 1.00000
>>
>> 52 archive 10.91399         osd.52                up  1.00000 1.00000
>>
>> 77 archive 10.91399         osd.77                up  1.00000 1.00000
>>
>> 89 archive 10.91399         osd.89                up  1.00000 1.00000
>>
>> 90 archive 10.91399         osd.90                up  1.00000 1.00000
>>
>> -31 130.96800     host CEPH008
>>
>>    5 archive 10.91399         osd.5                 up  1.00000 
>> 1.00000
>>
>>    6 archive 10.91399         osd.6                 up  1.00000 
>> 1.00000
>>
>> 11 archive 10.91399         osd.11                up  1.00000 1.00000
>>
>> 45 archive 10.91399         osd.45                up  1.00000 1.00000
>>
>> 46 archive 10.91399         osd.46                up  1.00000 1.00000
>>
>> 47 archive 10.91399         osd.47                up  1.00000 1.00000
>>
>> 55 archive 10.91399         osd.55                up  1.00000 1.00000
>>
>> 70 archive 10.91399         osd.70                up  1.00000 1.00000
>>
>> 71 archive 10.91399         osd.71                up  1.00000 1.00000
>>
>> 78 archive 10.91399         osd.78                up  1.00000 1.00000
>>
>> 79 archive 10.91399         osd.79                up  1.00000 1.00000
>>
>> 91 archive 10.91399         osd.91                up  1.00000 1.00000
>>
>> -1          27.91296 root default
>>
>> -30          6.98199     host CEPH-SSD-004
>>
>> 92     ssd 0.87299         osd.92                up  1.00000 1.00000
>>
>> 93     ssd 0.87299         osd.93                up  1.00000 1.00000
>>
>> 94     ssd 0.87299         osd.94                up  1.00000 1.00000
>>
>> 95     ssd 0.87299         osd.95                up  1.00000 1.00000
>>
>> 96     ssd 0.87299         osd.96                up  1.00000 1.00000
>>
>> 97     ssd 0.87299         osd.97                up  1.00000 1.00000
>>
>> 98     ssd 0.87299         osd.98                up  1.00000 1.00000
>>
>> 99     ssd 0.87299         osd.99                up  1.00000 1.00000
>>
>> -3 6.97699     host CEPH001
>>
>>    1     ssd 0.43599         osd.1                 up  1.00000 
>> 1.00000
>>
>> 17     ssd 0.43599         osd.17                up  1.00000 1.00000
>>
>> 18     ssd 0.43599         osd.18                up  1.00000 1.00000
>>
>> 19     ssd 0.43599         osd.19                up  1.00000 1.00000
>>
>> 20     ssd 0.43599         osd.20                up  1.00000 1.00000
>>
>> 21     ssd 0.43599         osd.21                up  1.00000 1.00000
>>
>> 22     ssd 0.43599         osd.22                up  1.00000 1.00000
>>
>> 23     ssd 0.43599         osd.23                up  1.00000 1.00000
>>
>> 37     ssd 0.87299         osd.37                up  1.00000 1.00000
>>
>> 54     ssd 0.87299         osd.54                up  1.00000 1.00000
>>
>> 56     ssd 0.43599         osd.56                up  1.00000 1.00000
>>
>> 60     ssd 0.43599         osd.60                up  1.00000 1.00000
>>
>> 61     ssd 0.43599         osd.61                up  1.00000 1.00000
>>
>> 62     ssd 0.43599         osd.62                up  1.00000 1.00000
>>
>> -5 6.97699     host CEPH002
>>
>>    2     ssd 0.43599         osd.2                 up  1.00000 
>> 1.00000
>>
>> 24     ssd 0         osd.24                up  1.00000 1.00000
>>
>> 25     ssd 0.43599         osd.25                up  1.00000 1.00000
>>
>> 26     ssd 0.43599         osd.26                up  1.00000 1.00000
>>
>> 27     ssd 0.43599         osd.27                up  1.00000 1.00000
>>
>> 28     ssd 0.43599         osd.28                up  1.00000 1.00000
>>
>> 29     ssd 0.43599         osd.29                up  1.00000 1.00000
>>
>> 30     ssd 0.43599         osd.30                up  1.00000 1.00000
>>
>> 38     ssd 0.87299         osd.38                up  1.00000 1.00000
>>
>> 57     ssd 0.43599         osd.57                up  1.00000 1.00000
>>
>> 63     ssd 0.43599         osd.63                up  1.00000 1.00000
>>
>> 64     ssd 0.43599         osd.64                up  1.00000 1.00000
>>
>> 65     ssd 0.43599         osd.65                up  1.00000 1.00000
>>
>> 66     ssd 0.43599         osd.66                up  1.00000 1.00000
>>
>> 72     ssd 0.87299         osd.72                up  1.00000 1.00000
>>
>> -7 6.97699     host CEPH003
>>
>>    0     ssd 0.43599         osd.0                 up  1.00000 
>> 1.00000
>>
>>    3     ssd 0.43599         osd.3                 up  1.00000 
>> 1.00000
>>
>> 31     ssd    0         osd.31                up  1.00000 1.00000
>>
>> 32     ssd 0.43599         osd.32                up  1.00000 1.00000
>>
>> 33     ssd 0.43599         osd.33                up  1.00000 1.00000
>>
>> 34     ssd 0.43599         osd.34                up  1.00000 1.00000
>>
>> 35     ssd 0.43599         osd.35                up  1.00000 1.00000
>>
>> 36     ssd 0.43599         osd.36                up  1.00000 1.00000
>>
>> 50     ssd 0.87299         osd.50                up  1.00000 1.00000
>>
>> 58     ssd   0.43599      osd.58                up  1.00000 1.00000
>>
>> 59     ssd 0.43599         osd.59                up  1.00000 1.00000
>>
>> 67     ssd 0.43599         osd.67                up  1.00000 1.00000
>>
>> 68     ssd 0.43599         osd.68                up  1.00000 1.00000
>>
>> 69     ssd 0.43599         osd.69                up  1.00000 1.00000
>>
>> 73     ssd 0.87299         osd.73                up  1.00000 1.00000
>>
>> CEPH CONF:
>>
>> [global]
>>
>> #Normal-Memory 1/5
>>
>> debug rgw = 1
>>
>> #Disable
>>
>> debug osd = 0
>>
>> debug journal = 0
>>
>> debug ms = 0
>>
>> fsid = e1ee8086-7cce-43fd-a252-3d677af22428
>>
>> mon_initial_members = CEPH001, CEPH002, CEPH003
>>
>> mon_host = 172.16.2.10,172.16.2.11,172.16.2.12
>>
>> auth_cluster_required = cephx
>>
>> auth_service_required = cephx
>>
>> auth_client_required = cephx
>>
>> osd pool default pg num = 1024
>>
>> osd pool default pgp num = 1024
>>
>> public network = 172.16.2.0/24
>>
>> cluster network = 172.16.1.0/24
>>
>> osd pool default size = 2
>>
>> osd pool default min size = 1
>>
>> rgw dynamic resharding = true
>>
>> [osd]
>>
>> osd mkfs type = xfs
>>
>> osd op threads = 12
>>
>> osd disk threads = 12
>>
>> osd recovery threads = 4
>>
>> osd recovery op priority = 1
>>
>> osd recovery max active = 2
>>
>> osd recovery max single start = 1
>>
>> osd max backfills = 4
>>
>> osd backfill scan max = 16
>>
>> osd backfill scan min = 4
>>
>> osd client op priority = 63
>>
>> osd memory target = 2147483648
>>
>> osd scrub begin hour = 23
>>
>> osd scrub end hour = 6
>>
>> osd scrub load threshold = 1 #low load scrubbing
>>
>> osd scrub during recovery = false #scrub during recovery
>>
>> [mon]
>>
>>      mon allow pool delete = true
>>
>>      mon osd min down reporters = 3
>>
>> [mon.a]
>>
>>      host = CEPH001
>>
>>      public bind addr = 172.16.2.10
>>
>>      mon addr = 172.16.2.10:6789
>>
>>      mon allow pool delete = true
>>
>> [mon.b]
>>
>>      host = CEPH002
>>
>>      public bind addr = 172.16.2.11
>>
>>      mon addr = 172.16.2.11:6789
>>
>>      mon allow pool delete = true
>>
>> [mon.c]
>>
>>      host = CEPH003
>>
>>      public bind addr = 172.16.2.12
>>
>>      mon addr = 172.16.2.12:6789
>>
>>      mon allow pool delete = true
>>
>> [client.rgw.ceph-rgw01]
>>
>> host = ceph-rgw01
>>
>> rgw dns name = eu-es-s3gateway.edh-services.com
>>
>> rgw frontends = "beast endpoint=172.16.2.6 port=8080"
>>
>> rgw resolve cname = false
>>
>> rgw thread pool size = 4096
>>
>> rgw op thread timeout = 600
>>
>> rgw num rados handles = 1
>>
>> rgw num control oids = 8
>>
>> rgw cache enabled = true
>>
>> rgw cache lru size = 10000
>>
>> rgw enable usage log = true
>>
>> rgw usage log tick interval = 30
>>
>> rgw usage log flush threshold = 1024
>>
>> rgw usage max shards = 32
>>
>> rgw usage max user shards = 1
>>
>> rgw log http headers = "http_x_forwarded_for"
>>
>> [client.rgw.ceph-rgw03]
>>
>> host = ceph-rgw03
>>
>> rgw dns name = XXXXXXXXXXXXXX
>>
>> rgw frontends = "beast endpoint=172.16.2.8 port=8080"
>>
>> rgw resolve cname = false
>>
>> rgw thread pool size = 4096
>>
>> rgw op thread timeout = 600
>>
>> rgw num rados handles = 1
>>
>> rgw num control oids = 8
>>
>> rgw cache enabled = true
>>
>> rgw cache lru size = 100000
>>
>> rgw enable usage log = true
>>
>> rgw usage log tick interval = 30
>>
>> rgw usage log flush threshold = 1024
>>
>> rgw usage max shards = 32
>>
>> rgw usage max user shards = 1
>>
>> rgw log http headers = "http_x_forwarded_for"
>>
>> [OSD HOST CROP]
>>
>> BUCKET:
>>
>> radosgw-admin bucket stats --bucket=XXXX
>>
>> {
>>
>>      "bucket": "XXXX",
>>
>>      "zonegroup": "4d8c7c5f-ca40-4ee3-b5bb-b2cad90bd007",
>>
>>      "placement_rule": "default-placement",
>>
>> "explicit_placement": {
>>
>>          "data_pool": "default.rgw.buckets.data",
>>
>> "data_extra_pool": "default.rgw.buckets.non-ec",
>>
>>          "index_pool": "default.rgw.buckets.index"
>>
>>      },
>>
>>      "id": "48efb8c3-693c-4fe0-bbe4-fdc16f590a82.14987264.1",
>>
>>      "marker": "48efb8c3-693c-4fe0-bbe4-fdc16f590a82.3856921.7",
>>
>>      "index_type": "Normal",
>>
>>      "owner": "XXXXX",
>>
>>      "ver":
>> "0#38096,1#37913,2#37954,3#37836,4#38134,5#38236,6#37911,7#37879,8#37
>> 8
>> 04,9#38005,10#37584,11#38446,12#37957,13#37537,14#37775,15#37794,16#3
>> 8 
>> 399,17#38162,18#37834,19#37633,20#37403,21#38173,22#37651,23#37303,24
>> #
>> 37992,25#38228,26#38441,27#37975,28#38095,29#37835,30#38264,31#37958,
>> 3
>> 2#37666,33#37517,34#38260,35#38168,36#37689,37#37551,38#37700,39#3805
>> 6
>> ,40#38175,41#37765,42#37721,43#38472,44#37928,45#38451,46#37491,47#37
>> 8
>> 75,48#38405,49#38011,50#38025,51#37983,52#37940,53#38306,54#37908,55#
>> 3
>> 8181,56#37721,57#38366,58#37834,59#38392,60#37928,61#38235,62#37837,6
>> 3 
>> #37940,64#38294,65#37610,66#37974,67#38304,68#37725,69#37301,70#37155
>> ,
>> 71#37681,72#37463,73#37603,74#37323,75#37717,76#37111,77#37688,78#374
>> 7
>> 3,79#37052,80#37413,81#37758,82#36971,83#37327,84#37056,85#37302,86#3
>> 7 
>> 492,87#37250,88#37708,89#36891,90#38019,91#37485,92#37335,93#37712,94
>> #
>> 37754,95#37117,96#37085,97#37694,98#37386,99#37384,100#37668,101#3732
>> 9
>> ,102#37177,103#37494,104#37296,105#37366,106#37853,107#37234,108#3694
>> 5
>> ,109#37040,110#37389,111#37973,112#37092,113#37327,114#37505,115#3754
>> 5
>> ,116#37884,117#37325,118#37401,119#37208,120#37277,121#37087,122#3766
>> 4 ,123#37649,124#37517,125#37618,126#37145,127#37300",
>>
>>      "master_ver":
>> "0#0,1#0,2#0,3#0,4#0,5#0,6#0,7#0,8#0,9#0,10#0,11#0,12#0,13#0,14#0,15#
>> 0
>> ,16#0,17#0,18#0,19#0,20#0,21#0,22#0,23#0,24#0,25#0,26#0,27#0,28#0,29#
>> 0
>> ,30#0,31#0,32#0,33#0,34#0,35#0,36#0,37#0,38#0,39#0,40#0,41#0,42#0,43#
>> 0
>> ,44#0,45#0,46#0,47#0,48#0,49#0,50#0,51#0,52#0,53#0,54#0,55#0,56#0,57#
>> 0
>> ,58#0,59#0,60#0,61#0,62#0,63#0,64#0,65#0,66#0,67#0,68#0,69#0,70#0,71#
>> 0
>> ,72#0,73#0,74#0,75#0,76#0,77#0,78#0,79#0,80#0,81#0,82#0,83#0,84#0,85#
>> 0
>> ,86#0,87#0,88#0,89#0,90#0,91#0,92#0,93#0,94#0,95#0,96#0,97#0,98#0,99#
>> 0
>> ,100#0,101#0,102#0,103#0,104#0,105#0,106#0,107#0,108#0,109#0,110#0,11
>> 1
>> #0,112#0,113#0,114#0,115#0,116#0,117#0,118#0,119#0,120#0,121#0,122#0,
>> 1
>> 23#0,124#0,125#0,126#0,127#0",
>>
>>      "mtime": "2019-04-30 15:14:20.152747",
>>
>>      "max_marker":
>> "0#,1#,2#,3#,4#,5#,6#,7#,8#,9#,10#,11#,12#,13#,14#,15#,16#,17#,18#,19
>> #
>> ,20#,21#,22#,23#,24#,25#,26#,27#,28#,29#,30#,31#,32#,33#,34#,35#,36#,
>> 3 
>> 7#,38#,39#,40#,41#,42#,43#,44#,45#,46#,47#,48#,49#,50#,51#,52#,53#,54
>> #
>> ,55#,56#,57#,58#,59#,60#,61#,62#,63#,64#,65#,66#,67#,68#,69#,70#,71#,
>> 7 
>> 2#,73#,74#,75#,76#,77#,78#,79#,80#,81#,82#,83#,84#,85#,86#,87#,88#,89
>> # 
>> ,90#,91#,92#,93#,94#,95#,96#,97#,98#,99#,100#,101#,102#,103#,104#,105
>> # 
>> ,106#,107#,108#,109#,110#,111#,112#,113#,114#,115#,116#,117#,118#,119
>> # ,120#,121#,122#,123#,124#,125#,126#,127#",
>>
>>      "usage": {
>>
>>          "rgw.none": {
>>
>>              "size": 0,
>>
>> "size_actual": 0,
>>
>> "size_utilized": 0,
>>
>>              "size_kb": 0,
>>
>> "size_kb_actual": 0,
>>
>> "size_kb_utilized": 0,
>>
>> "num_objects": 0
>>
>>          },
>>
>>          "rgw.main": {
>>
>>              "size": 95368837327307,
>>
>> "size_actual": 95379421843456,
>>
>> "size_utilized": 95370098064318,
>>
>>              "size_kb": 93133630203,
>>
>> "size_kb_actual": 93143966644,
>>
>> "size_kb_utilized": 93134861391,
>>
>> "num_objects": 5872260
>>
>>          },
>>
>>          "rgw.multimeta": {
>>
>>              "size": 0,
>>
>> "size_actual": 0,
>>
>> "size_utilized": 0,
>>
>>              "size_kb": 0,
>>
>> "size_kb_actual": 0,
>>
>> "size_kb_utilized": 0,
>>
>> "num_objects": 467
>>
>>          }
>>
>>      },
>>
>>      "bucket_quota": {
>>
>>          "enabled": false,
>>
>>          "check_on_raw": false,
>>
>>          "max_size": -1024,
>>
>>          "max_size_kb": 0,
>>
>>          "max_objects": -1
>>
>>      }
>>
>> }
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to