Home directories probably means lots of small objects. Default minimum
allocation size of BlueStore on HDD is 64 kiB, so there's a lot of overhead
for everything smaller;

Details: google bluestore min alloc size, can only be changed during OSD
creation

Paul

-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90


On Fri, Dec 6, 2019 at 12:57 PM Jochen Schulz <sch...@math.uni-goettingen.de>
wrote:

> Hi!
>
> Thank you!
> The output of both commands are below.
> I still dont understand why there are 21T used data (because 5.5T*3 =
> 16.5T != 21T) and why there seems to be only 4.5 T MAX AVAIL, but the
> osd output tells we have 25T free space.
>
>
> $ sudo ceph df
> RAW STORAGE:
>     CLASS     SIZE        AVAIL       USED        RAW USED     %RAW USED
>     hdd        45 TiB      24 TiB      21 TiB       21 TiB         46.33
>     ssd       596 GiB     524 GiB     1.7 GiB       72 GiB         12.09
>     TOTAL      46 TiB      25 TiB      21 TiB       21 TiB         45.89
>
> POOLS:
>     POOL                ID     STORED      OBJECTS     USED        %USED
>     MAX AVAIL
>     images               8     149 GiB      38.30k     354 GiB      2.52
>       4.5 TiB
>     cephfs_data          9     5.5 TiB      26.61M      20 TiB     60.36
>       4.5 TiB
>     cephfs_metadata     10      12 GiB       3.17M      13 GiB      2.57
>       164 GiB
>
>
> $ sudo ceph osd df
> ID CLASS WEIGHT  REWEIGHT SIZE    RAW USE DATA    OMAP    META    AVAIL
>    %USE  VAR  PGS STATUS
>  0   hdd 0.89000  1.00000 931 GiB 456 GiB 453 GiB 136 MiB 3.5 GiB  475
> GiB 49.01 1.07 103     up
>  1   hdd 0.89000  1.00000 931 GiB 495 GiB 491 GiB 100 MiB 3.9 GiB  436
> GiB 53.14 1.16  89     up
>  4   hdd 0.89000  1.00000 931 GiB 345 GiB 342 GiB 108 MiB 2.9 GiB  586
> GiB 37.05 0.81  87     up
>  5   hdd 0.89000  1.00000 931 GiB 521 GiB 517 GiB 108 MiB 4.1 GiB  410
> GiB 55.96 1.22  98     up
>  6   hdd 0.89000  1.00000 931 GiB 367 GiB 364 GiB  95 MiB 3.2 GiB  564
> GiB 39.44 0.86  95     up
>  7   hdd 0.89000  1.00000 931 GiB 540 GiB 536 GiB  77 MiB 3.7 GiB  392
> GiB 57.96 1.26 111     up
> 20   hdd 0.89000  1.00000 931 GiB 382 GiB 378 GiB  60 MiB 3.3 GiB  550
> GiB 40.96 0.89  85     up
> 23   hdd 1.81929  1.00000 1.8 TiB 706 GiB 701 GiB 113 MiB 4.9 GiB  1.1
> TiB 37.92 0.83 182     up
> 44   hdd 0.89000  1.00000 931 GiB 468 GiB 465 GiB  34 MiB 3.3 GiB  463
> GiB 50.29 1.10  93     up
> 45   hdd 1.78999  1.00000 1.8 TiB 882 GiB 875 GiB 138 MiB 6.3 GiB  981
> GiB 47.33 1.03 179     up
> 46   hdd 1.78999  1.00000 1.8 TiB 910 GiB 903 GiB 127 MiB 6.4 GiB  953
> GiB 48.83 1.06 192     up
> 22   ssd 0.11639  1.00000 119 GiB  15 GiB 357 MiB  12 GiB 2.8 GiB  104
> GiB 12.61 0.27 315     up
> 12   hdd 0.89000  1.00000 931 GiB 499 GiB 494 GiB  64 MiB 4.5 GiB  432
> GiB 53.57 1.17 116     up
> 13   hdd 0.89000  1.00000 931 GiB 536 GiB 532 GiB  48 MiB 4.4 GiB  395
> GiB 57.59 1.26 109     up
> 30   hdd 0.89000  1.00000 931 GiB 510 GiB 506 GiB  33 MiB 3.9 GiB  421
> GiB 54.80 1.19 100     up
> 32   hdd 0.89000  1.00000 931 GiB 495 GiB 491 GiB  56 MiB 4.1 GiB  436
> GiB 53.17 1.16 101     up
> 33   hdd 0.89000  1.00000 931 GiB 333 GiB 330 GiB  56 MiB 3.1 GiB  598
> GiB 35.80 0.78  82     up
> 15   ssd 0.11639  1.00000 119 GiB  14 GiB 336 MiB  11 GiB 2.9 GiB  105
> GiB 12.13 0.26 305     up
> 17   hdd 0.89000  1.00000 931 GiB 577 GiB 573 GiB  77 MiB 4.4 GiB  354
> GiB 61.99 1.35  97     up
> 18   hdd 0.89000  1.00000 931 GiB 413 GiB 409 GiB  70 MiB 4.0 GiB  518
> GiB 44.34 0.97  95     up
> 19   hdd 1.81879  1.00000 1.8 TiB 895 GiB 889 GiB 144 MiB 5.6 GiB  967
> GiB 48.06 1.05 184     up
> 21   hdd 0.89000  1.00000 931 GiB 360 GiB 357 GiB  60 MiB 3.4 GiB  570
> GiB 38.72 0.84 100     up
> 31   hdd 0.90909  1.00000 931 GiB 508 GiB 505 GiB  80 MiB 3.5 GiB  423
> GiB 54.58 1.19 102     up
> 25   ssd 0.11639  1.00000 119 GiB  14 GiB 339 MiB  11 GiB 2.7 GiB  105
> GiB 11.86 0.26 310     up
>  8   hdd 0.89000  1.00000 931 GiB 359 GiB 356 GiB  72 MiB 3.1 GiB  572
> GiB 38.55 0.84  80     up
>  9   hdd 0.89000  1.00000 931 GiB 376 GiB 373 GiB  42 MiB 3.0 GiB  555
> GiB 40.39 0.88  87     up
> 24   hdd 0.89000  1.00000 931 GiB 342 GiB 339 GiB  70 MiB 2.8 GiB  590
> GiB 36.69 0.80  78     up
> 26   hdd 1.78999  1.00000 1.8 TiB 921 GiB 915 GiB 129 MiB 6.1 GiB  942
> GiB 49.45 1.08 177     up
> 27   hdd 1.78999  1.00000 1.8 TiB 891 GiB 885 GiB 125 MiB 5.7 GiB  972
> GiB 47.82 1.04 208     up
> 35   hdd 1.81929  1.00000 1.8 TiB 819 GiB 814 GiB 110 MiB 5.3 GiB  1.0
> TiB 43.99 0.96 184     up
> 29   ssd 0.11638  1.00000 119 GiB  15 GiB 339 MiB  11 GiB 2.9 GiB  105
> GiB 12.25 0.27 311     up
> 14   hdd 1.81929  1.00000 1.8 TiB 889 GiB 884 GiB 1.9 MiB 4.3 GiB  974
> GiB 47.70 1.04 162     up
> 28   hdd 1.81929  1.00000 1.8 TiB 609 GiB 606 GiB 3.4 MiB 2.9 GiB  1.2
> TiB 32.67 0.71 169     up
> 36   hdd 1.36429  1.00000 1.4 TiB 591 GiB 589 GiB 2.0 MiB 2.9 GiB  806
> GiB 42.34 0.92 135     up
> 37   hdd 1.81929  1.00000 1.8 TiB 840 GiB 836 GiB 2.5 MiB 4.1 GiB 1023
> GiB 45.12 0.98 177     up
> 38   hdd 1.81929  1.00000 1.8 TiB 914 GiB 909 GiB 2.2 MiB 4.4 GiB  949
> GiB 49.05 1.07 182     up
>  2   hdd 0.89000  1.00000 931 GiB 449 GiB 445 GiB  60 MiB 3.4 GiB  482
> GiB 48.19 1.05  89     up
>  3   hdd 0.89000  1.00000 931 GiB 333 GiB 329 GiB  75 MiB 3.0 GiB  598
> GiB 35.73 0.78  86     up
> 10   hdd 0.89000  1.00000 931 GiB 416 GiB 412 GiB  54 MiB 3.8 GiB  515
> GiB 44.68 0.97 100     up
> 11   hdd 1.81879  1.00000 1.8 TiB 803 GiB 798 GiB  70 MiB 5.3 GiB  1.0
> TiB 43.14 0.94 188     up
> 34   hdd 0.90909  1.00000 931 GiB 605 GiB 600 GiB  64 MiB 4.2 GiB  326
> GiB 64.95 1.42 106     up
> 16   ssd 0.11639  1.00000 119 GiB  14 GiB 322 MiB  10 GiB 3.1 GiB  105
> GiB 11.59 0.25 295     up
>                     TOTAL  46 TiB  21 TiB  21 TiB  59 GiB 165 GiB   25
> TiB 45.89
> MIN/MAX VAR: 0.25/1.42  STDDEV: 13.76
>
>
>
>
> On 06.12.19 12:27, Aleksey Gutikov wrote:
> > On 6.12.19 13:29, Jochen Schulz wrote:
> >> Hi!
> >>
> >> We have a ceph cluster with 42 OSD in production as a server providing
> >> mainly home-directories of users. Ceph is 14.2.4 nautilus.
> >>
> >> We have 3 pools. One images (for rbd images) a cephfs_metadata and a
> >> cephfs_data pool.
> >>
> >> Our raw data is about 5.6T. All pools have replica size 3 and there are
> >> only very little snapshots in the rbd images pool, the cephfspool doesnt
> >> use snapshots.
> >>
> >> How is it possible that the status tells us, that 21T/46T  is used,
> >> because thats much more than 3 times the raw size.
> >>
> >> Also, to make that more confusing, there as at least half of the cluster
> >> free, and we get pg backfill_toofull after we added some OSDs lately.
> >> The ceph dashboard tells aus the pool ist 82 % full and has only 4.5 T
> >> free.
> >>
> >> The autoscale module seems to calculate the 20T times 3 for the space
> >> needed and thus has wrong numbers (see below).
> >>
> >> Status of the cluster is added below too.
> >>
> >> how can these size/capacity numbers be explained?
> >> and, would be there a recommendation to change something?
> >>
> >> Thank you in advance!
> >>
> >> best
> >> Jochen
> >>
> >>
> >> # ceph -s
> >>
> >>   cluster:
> >>      id:     2b16167f-3f33-4580-a0e9-7a71978f403d
> >>      health: HEALTH_ERR
> >>              Degraded data redundancy (low space): 1 pg backfill_toofull
> >>              1 subtrees have overcommitted pool target_size_bytes
> >>              1 subtrees have overcommitted pool target_size_ratio
> >>              2 pools have too many placement groups
> >>
> >>    services:
> >>      mon: 4 daemons, quorum jade,assam,matcha,jasmine (age 2d)
> >>      mgr: earl(active, since 24h), standbys: assam
> >>      mds: cephfs:1 {0=assam=up:active} 1 up:standby
> >>      osd: 42 osds: 42 up (since 106m), 42 in (since 115m); 30 remapped
> >> pgs
> >>
> >>    data:
> >>      pools:   3 pools, 2048 pgs
> >>      objects: 29.80M objects, 5.6 TiB
> >>      usage:   21 TiB used, 25 TiB / 46 TiB avail
> >>      pgs:     1164396/89411013 objects misplaced (1.302%)
> >>               2018 active+clean
> >>               22   active+remapped+backfill_wait
> >>               7    active+remapped+backfilling
> >>               1    active+remapped+backfill_wait+backfill_toofull
> >>
> >>    io:
> >>      client:   1.7 KiB/s rd, 516 KiB/s wr, 0 op/s rd, 28 op/s wr
> >>      recovery: 9.2 MiB/s, 41 objects/s
> >>
> >>
> >> # ceph osd pool autoscale-status
> >>   POOL               SIZE  TARGET SIZE  RATE  RAW CAPACITY   RATIO
> >> TARGET RATIO  BIAS  PG_NUM  NEW PG_NUM  AUTOSCALE
> >>   images           354.2G                3.0        46100G  0.0231
> >>           1.0    1024          32  warn
> >>   cephfs_metadata  13260M                3.0        595.7G  0.0652
> >>           1.0     512           8  warn
> >>   cephfs_data      20802G                3.0        46100G  1.3537
> >>           1.0     512              warn
> >>
> >>
> >>
> >> _______________________________________________
> >> ceph-users mailing list -- ceph-users@ceph.io
> >> To unsubscribe send an email to ceph-users-le...@ceph.io
> >>
> >
> > Please, provide output of ceph df and ceph osd df - that should explain
> > both questions (21T and 82%).
> >
> >
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to