On 12/3/19 1:30 PM, Lars Täuber wrote:
here it comes:

$ ceph osd df tree
ID  CLASS WEIGHT    REWEIGHT SIZE    RAW USE DATA    OMAP    META    AVAIL    
%USE  VAR  PGS STATUS TYPE NAME
  -1       195.40730        - 195 TiB 130 TiB 128 TiB  58 GiB 476 GiB   66 TiB 
66.45 1.00   -        root default
-25       195.40730        - 195 TiB 130 TiB 128 TiB  58 GiB 476 GiB   66 TiB 
66.45 1.00   -            room PRZ
-26       195.40730        - 195 TiB 130 TiB 128 TiB  58 GiB 476 GiB   66 TiB 
66.45 1.00   -                row rechts
-27        83.74599        -  84 TiB  57 TiB  56 TiB  25 GiB 206 GiB   27 TiB 
67.51 1.02   -                    rack 1-eins
  -3        27.91533        -  28 TiB  18 TiB  17 TiB 8.4 GiB  66 GiB   10 TiB 
62.80 0.95   -                        host onode1
   0   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.8 MiB  14 GiB  2.1 TiB 
62.48 0.94 163     up                     osd.0
   1   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 6.5 MiB  12 GiB  2.1 TiB 
62.47 0.94 163     up                     osd.1
   2   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 7.1 MiB  12 GiB  2.1 TiB 
62.53 0.94 163     up                     osd.2
   3   hdd   5.51459  1.00000 5.5 TiB 3.5 TiB 3.4 TiB 7.5 MiB  12 GiB  2.0 TiB 
62.90 0.95 164     up                     osd.3
  37   hdd   5.51459  1.00000 5.5 TiB 3.7 TiB 3.7 TiB 6.4 MiB  13 GiB  1.8 TiB 
67.32 1.01 176     up                     osd.37
   4   ssd   0.34239  1.00000 351 GiB  11 GiB 187 MiB 8.3 GiB 2.0 GiB  340 GiB  
3.01 0.05 110     up                     osd.4
-13        27.91533        -  28 TiB  17 TiB  17 TiB 8.2 GiB  66 GiB   10 TiB 
62.64 0.94   -                        host onode4
  13   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.0 MiB  13 GiB  2.1 TiB 
62.49 0.94 163     up                     osd.13
  14   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.2 MiB  13 GiB  2.1 TiB 
62.49 0.94 163     up                     osd.14
  15   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.4 MiB  12 GiB  2.1 TiB 
62.43 0.94 163     up                     osd.15
  16   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 5.8 MiB  12 GiB  2.1 TiB 
62.13 0.94 162     up                     osd.16
  40   hdd   5.51459  1.00000 5.5 TiB 3.7 TiB 3.7 TiB 3.6 MiB  13 GiB  1.8 TiB 
67.36 1.01 176     up                     osd.40
  33   ssd   0.34239  1.00000 351 GiB  11 GiB 201 MiB 8.2 GiB 2.2 GiB  340 GiB  
3.02 0.05 110     up                     osd.33
-22        27.91533        -  28 TiB  22 TiB  21 TiB 8.1 GiB  74 GiB  6.4 TiB 
77.10 1.16   -                        host onode7
  25   hdd   5.51459  1.00000 5.5 TiB 4.3 TiB 4.2 TiB 7.2 MiB  14 GiB  1.2 TiB 
77.59 1.17 203     up                     osd.25
  26   hdd   5.51459  1.00000 5.5 TiB 4.3 TiB 4.3 TiB 4.7 MiB  14 GiB  1.2 TiB 
78.40 1.18 205     up                     osd.26
  27   hdd   5.51459  1.00000 5.5 TiB 4.2 TiB 4.1 TiB 3.8 MiB  14 GiB  1.3 TiB 
75.80 1.14 198     up                     osd.27
  28   hdd   5.51459  1.00000 5.5 TiB 4.2 TiB 4.1 TiB 4.5 MiB  14 GiB  1.3 TiB 
76.13 1.15 199     up                     osd.28
  30   hdd   5.51459  1.00000 5.5 TiB 4.5 TiB 4.5 TiB 8.2 MiB  15 GiB 1006 GiB 
82.18 1.24 215     up                     osd.30
  36   ssd   0.34239  1.00000 351 GiB  10 GiB 184 MiB 8.1 GiB 2.0 GiB  340 GiB  
2.92 0.04 110     up                     osd.36
-28        55.83066        -  56 TiB  35 TiB  34 TiB  17 GiB 132 GiB   21 TiB 
62.36 0.94   -                    rack 2-zwei
  -7        27.91533        -  28 TiB  17 TiB  17 TiB 8.2 GiB  66 GiB   11 TiB 
62.27 0.94   -                        host onode2
   5   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.0 MiB  12 GiB  2.1 TiB 
62.08 0.93 162     up                     osd.5
   6   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.9 MiB  13 GiB  2.1 TiB 
62.13 0.93 162     up                     osd.6
   7   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.3 TiB 3.7 MiB  12 GiB  2.1 TiB 
61.77 0.93 161     up                     osd.7
   8   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.3 TiB 3.2 MiB  12 GiB  2.1 TiB 
61.75 0.93 161     up                     osd.8
  38   hdd   5.51459  1.00000 5.5 TiB 3.7 TiB 3.7 TiB 3.7 MiB  14 GiB  1.8 TiB 
67.31 1.01 176     up                     osd.38
  31   ssd   0.34239  1.00000 351 GiB  11 GiB 166 MiB 8.1 GiB 2.4 GiB  340 GiB  
3.04 0.05 110     up                     osd.31
-16        27.91533        -  28 TiB  17 TiB  17 TiB 8.7 GiB  66 GiB   10 TiB 
62.44 0.94   -                        host onode5
  17   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB   4 MiB  12 GiB  2.1 TiB 
62.15 0.94 162     up                     osd.17
  18   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.0 MiB  13 GiB  2.1 TiB 
62.16 0.94 162     up                     osd.18
  19   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.5 MiB  13 GiB  2.1 TiB 
62.14 0.94 162     up                     osd.19
  20   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.5 MiB  13 GiB  2.1 TiB 
62.12 0.93 162     up                     osd.20
  41   hdd   5.51459  1.00000 5.5 TiB 3.7 TiB 3.7 TiB 3.9 MiB  14 GiB  1.8 TiB 
67.31 1.01 176     up                     osd.41
  34   ssd   0.34239  1.00000 351 GiB  11 GiB 192 MiB 8.7 GiB 1.8 GiB  340 GiB  
3.04 0.05 109     up                     osd.34
-29        55.83066        -  56 TiB  38 TiB  38 TiB  16 GiB 138 GiB   17 TiB 
68.95 1.04   -                    rack 3-drei
-10        27.91533        -  28 TiB  17 TiB  17 TiB 8.1 GiB  63 GiB   11 TiB 
61.02 0.92   -                        host onode3
   9   hdd   5.51459  1.00000 5.5 TiB 3.3 TiB 3.3 TiB 3.7 MiB  12 GiB  2.2 TiB 
60.63 0.91 158     up                     osd.9
  10   hdd   5.51459  1.00000 5.5 TiB 3.3 TiB 3.3 TiB 3.1 MiB  12 GiB  2.2 TiB 
60.19 0.91 157     up                     osd.10
  11   hdd   5.51459  1.00000 5.5 TiB 3.3 TiB 3.3 TiB 6.7 MiB  12 GiB  2.2 TiB 
60.27 0.91 157     up                     osd.11
  12   hdd   5.51459  1.00000 5.5 TiB 3.3 TiB 3.3 TiB 4.1 MiB  12 GiB  2.2 TiB 
60.28 0.91 157     up                     osd.12
  39   hdd   5.51459  1.00000 5.5 TiB 3.7 TiB 3.7 TiB 4.6 MiB  13 GiB  1.8 TiB 
67.34 1.01 176     up                     osd.39
  32   ssd   0.34239  1.00000 351 GiB  10 GiB 271 MiB 8.1 GiB 1.8 GiB  341 GiB  
2.88 0.04 109     up                     osd.32
-19        27.91533        -  28 TiB  21 TiB  21 TiB 8.1 GiB  74 GiB  6.5 TiB 
76.89 1.16   -                        host onode6
  21   hdd   5.51459  1.00000 5.5 TiB 4.0 TiB 4.0 TiB 6.2 MiB  13 GiB  1.5 TiB 
72.79 1.10 190     up                     osd.21
  22   hdd   5.51459  1.00000 5.5 TiB 4.5 TiB 4.5 TiB 5.1 MiB  16 GiB  1.0 TiB 
81.79 1.23 214     up                     osd.22
  23   hdd   5.51459  1.00000 5.5 TiB 4.4 TiB 4.4 TiB 4.4 MiB  16 GiB  1.1 TiB 
80.29 1.21 210     up                     osd.23
  24   hdd   5.51459  1.00000 5.5 TiB 4.3 TiB 4.2 TiB 6.7 MiB  14 GiB  1.3 TiB 
77.31 1.16 202     up                     osd.24
  29   hdd   5.51459  1.00000 5.5 TiB 4.2 TiB 4.2 TiB 4.6 MiB  14 GiB  1.3 TiB 
76.86 1.16 201     up                     osd.29
  35   ssd   0.34239  1.00000 351 GiB  10 GiB 208 MiB 8.1 GiB 1.9 GiB  340 GiB  
2.89 0.04 110     up                     osd.35
                        TOTAL 195 TiB 130 TiB 128 TiB  58 GiB 476 GiB   66 TiB 
66.45
MIN/MAX VAR: 0.04/1.24  STDDEV: 26.74


better only for the class hdd

$ ceph osd df tree class hdd
ID  CLASS WEIGHT    REWEIGHT SIZE    RAW USE DATA    OMAP    META    AVAIL    
%USE  VAR  PGS STATUS TYPE NAME
  -1       195.40730        - 193 TiB 130 TiB 128 TiB 169 MiB 462 GiB   63 TiB 
67.24 1.00   -        root default
-25       195.40730        - 193 TiB 130 TiB 128 TiB 169 MiB 462 GiB   63 TiB 
67.24 1.00   -            room PRZ
-26       195.40730        - 193 TiB 130 TiB 128 TiB 169 MiB 462 GiB   63 TiB 
67.24 1.00   -                row rechts
-27        83.74599        -  83 TiB  57 TiB  56 TiB  81 MiB 200 GiB   26 TiB 
68.31 1.02   -                    rack 1-eins
  -3        27.91533        -  28 TiB  18 TiB  17 TiB  31 MiB  64 GiB   10 TiB 
63.54 0.94   -                        host onode1
   0   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.8 MiB  14 GiB  2.1 TiB 
62.48 0.93 163     up                     osd.0
   1   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 6.5 MiB  12 GiB  2.1 TiB 
62.47 0.93 163     up                     osd.1
   2   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 7.1 MiB  12 GiB  2.1 TiB 
62.53 0.93 163     up                     osd.2
   3   hdd   5.51459  1.00000 5.5 TiB 3.5 TiB 3.4 TiB 7.5 MiB  12 GiB  2.0 TiB 
62.90 0.94 164     up                     osd.3
  37   hdd   5.51459  1.00000 5.5 TiB 3.7 TiB 3.7 TiB 6.4 MiB  13 GiB  1.8 TiB 
67.32 1.00 176     up                     osd.37
-13        27.91533        -  28 TiB  17 TiB  17 TiB  21 MiB  64 GiB   10 TiB 
63.38 0.94   -                        host onode4
  13   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.0 MiB  13 GiB  2.1 TiB 
62.49 0.93 163     up                     osd.13
  14   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.2 MiB  13 GiB  2.1 TiB 
62.49 0.93 163     up                     osd.14
  15   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.4 MiB  12 GiB  2.1 TiB 
62.43 0.93 163     up                     osd.15
  16   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 5.8 MiB  12 GiB  2.1 TiB 
62.13 0.92 162     up                     osd.16
  40   hdd   5.51459  1.00000 5.5 TiB 3.7 TiB 3.7 TiB 3.6 MiB  13 GiB  1.8 TiB 
67.36 1.00 176     up                     osd.40
-22        27.91533        -  28 TiB  22 TiB  21 TiB  28 MiB  72 GiB  6.1 TiB 
78.02 1.16   -                        host onode7
  25   hdd   5.51459  1.00000 5.5 TiB 4.3 TiB 4.2 TiB 7.2 MiB  14 GiB  1.2 TiB 
77.59 1.15 203     up                     osd.25
  26   hdd   5.51459  1.00000 5.5 TiB 4.3 TiB 4.3 TiB 4.7 MiB  14 GiB  1.2 TiB 
78.40 1.17 205     up                     osd.26
  27   hdd   5.51459  1.00000 5.5 TiB 4.2 TiB 4.1 TiB 3.8 MiB  14 GiB  1.3 TiB 
75.80 1.13 198     up                     osd.27
  28   hdd   5.51459  1.00000 5.5 TiB 4.2 TiB 4.1 TiB 4.5 MiB  14 GiB  1.3 TiB 
76.13 1.13 199     up                     osd.28
  30   hdd   5.51459  1.00000 5.5 TiB 4.5 TiB 4.5 TiB 8.2 MiB  15 GiB 1006 GiB 
82.18 1.22 215     up                     osd.30
-28        55.83066        -  55 TiB  35 TiB  34 TiB  38 MiB 128 GiB   20 TiB 
63.09 0.94   -                    rack 2-zwei
  -7        27.91533        -  28 TiB  17 TiB  17 TiB  18 MiB  63 GiB   10 TiB 
63.01 0.94   -                        host onode2
   5   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.0 MiB  12 GiB  2.1 TiB 
62.08 0.92 162     up                     osd.5
   6   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.9 MiB  13 GiB  2.1 TiB 
62.13 0.92 162     up                     osd.6
   7   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.3 TiB 3.7 MiB  12 GiB  2.1 TiB 
61.77 0.92 161     up                     osd.7
   8   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.3 TiB 3.2 MiB  12 GiB  2.1 TiB 
61.75 0.92 161     up                     osd.8
  38   hdd   5.51459  1.00000 5.5 TiB 3.7 TiB 3.7 TiB 3.7 MiB  14 GiB  1.8 TiB 
67.31 1.00 176     up                     osd.38
-16        27.91533        -  28 TiB  17 TiB  17 TiB  20 MiB  65 GiB   10 TiB 
63.18 0.94   -                        host onode5
  17   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB   4 MiB  12 GiB  2.1 TiB 
62.15 0.92 162     up                     osd.17
  18   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.0 MiB  13 GiB  2.1 TiB 
62.16 0.92 162     up                     osd.18
  19   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.5 MiB  13 GiB  2.1 TiB 
62.14 0.92 162     up                     osd.19
  20   hdd   5.51459  1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.5 MiB  13 GiB  2.1 TiB 
62.12 0.92 162     up                     osd.20
  41   hdd   5.51459  1.00000 5.5 TiB 3.7 TiB 3.7 TiB 3.9 MiB  14 GiB  1.8 TiB 
67.31 1.00 176     up                     osd.41
-29        55.83066        -  55 TiB  38 TiB  38 TiB  49 MiB 134 GiB   17 TiB 
69.77 1.04   -                    rack 3-drei
-10        27.91533        -  28 TiB  17 TiB  17 TiB  22 MiB  62 GiB   11 TiB 
61.74 0.92   -                        host onode3
   9   hdd   5.51459  1.00000 5.5 TiB 3.3 TiB 3.3 TiB 3.7 MiB  12 GiB  2.2 TiB 
60.63 0.90 158     up                     osd.9
  10   hdd   5.51459  1.00000 5.5 TiB 3.3 TiB 3.3 TiB 3.1 MiB  12 GiB  2.2 TiB 
60.19 0.90 157     up                     osd.10
  11   hdd   5.51459  1.00000 5.5 TiB 3.3 TiB 3.3 TiB 6.7 MiB  12 GiB  2.2 TiB 
60.27 0.90 157     up                     osd.11
  12   hdd   5.51459  1.00000 5.5 TiB 3.3 TiB 3.3 TiB 4.1 MiB  12 GiB  2.2 TiB 
60.28 0.90 157     up                     osd.12
  39   hdd   5.51459  1.00000 5.5 TiB 3.7 TiB 3.7 TiB 4.6 MiB  13 GiB  1.8 TiB 
67.34 1.00 176     up                     osd.39
-19        27.91533        -  28 TiB  21 TiB  21 TiB  27 MiB  72 GiB  6.1 TiB 
77.81 1.16   -                        host onode6
  21   hdd   5.51459  1.00000 5.5 TiB 4.0 TiB 4.0 TiB 6.2 MiB  13 GiB  1.5 TiB 
72.79 1.08 190     up                     osd.21
  22   hdd   5.51459  1.00000 5.5 TiB 4.5 TiB 4.5 TiB 5.1 MiB  16 GiB  1.0 TiB 
81.79 1.22 214     up                     osd.22
  23   hdd   5.51459  1.00000 5.5 TiB 4.4 TiB 4.4 TiB 4.4 MiB  16 GiB  1.1 TiB 
80.29 1.19 210     up                     osd.23
  24   hdd   5.51459  1.00000 5.5 TiB 4.3 TiB 4.2 TiB 6.7 MiB  14 GiB  1.3 TiB 
77.31 1.15 202     up                     osd.24
  29   hdd   5.51459  1.00000 5.5 TiB 4.2 TiB 4.2 TiB 4.6 MiB  14 GiB  1.3 TiB 
76.86 1.14 201     up                     osd.29
                        TOTAL 193 TiB 130 TiB 128 TiB 169 MiB 462 GiB   63 TiB 
67.24
MIN/MAX VAR: 0.90/1.22  STDDEV: 7.17




ceph osd pool ls detail
pool 1 'cephfs_data' erasure size 6 min_size 5 crush_rule 1 object_hash 
rjenkins pg_num 1024 pgp_num 1024 autoscale_mode on last_change 20353 lfor 
0/0/2366 flags hashpspool,ec_overwrites,selfmanaged_snaps max_bytes 
119457034600410 stripe_width 16384 target_size_ratio 0.85 application cephfs
        removed_snaps 
[2~4,7~27,2f~1e,4f~1f,6f~39,a9~5,af~1,b1~1,b3~1,b5~1,b7~1,b9~1,bb~1,bd~1,bf~1,c1~1,c3~1,c5~1,c7~1,c9~1]
pool 2 'cephfs_metadata' replicated size 3 min_size 2 crush_rule 2 object_hash 
rjenkins pg_num 256 pgp_num 256 autoscale_mode on last_change 261 lfor 0/0/259 
flags hashpspool stripe_width 0 pg_autoscale_bias 4 pg_num_min 16 
recovery_priority 5 target_size_ratio 0.3 application cephfs


$ ceph osd crush rule dump
[
     {
         "rule_id": 0,
         "rule_name": "replicated_rule",
         "ruleset": 0,
         "type": 1,
         "min_size": 1,
         "max_size": 10,
         "steps": [
             {
                 "op": "take",
                 "item": -1,
                 "item_name": "default"
             },
             {
                 "op": "chooseleaf_firstn",
                 "num": 0,
                 "type": "host"
             },
             {
                 "op": "emit"
             }
         ]
     },
     {
         "rule_id": 1,
         "rule_name": "cephfs_data",
         "ruleset": 1,
         "type": 3,
         "min_size": 3,
         "max_size": 6,
         "steps": [
             {
                 "op": "set_chooseleaf_tries",
                 "num": 5
             },
             {
                 "op": "set_choose_tries",
                 "num": 100
             },
             {
                 "op": "take",
                 "item": -2,
                 "item_name": "default~hdd"
             },
             {
                 "op": "chooseleaf_indep",
                 "num": 0,
                 "type": "host"
             },
             {
                 "op": "emit"
             }
         ]
     },
     {
         "rule_id": 2,
         "rule_name": "rep_3_ssd",
         "ruleset": 2,
         "type": 1,
         "min_size": 1,
         "max_size": 10,
         "steps": [
             {
                 "op": "take",
                 "item": -6,
                 "item_name": "default~ssd"
             },
             {
                 "op": "chooseleaf_firstn",
                 "num": 0,
                 "type": "host"
             },
             {
                 "op": "emit"
             }
         ]
     }
]

Lars, you have too much PG's for this OSD's. I suggest to disable PG autoscaler and:

- reduce number of PG's for cephfs_metada pool to something like 16 PG's.

- reduce number of PG's for cephfs_data to something like 512.

- update crush rule for cephfs_metadata pool - set domain to 'rack' instead 'host'.


Also please paste your `ceph osd tree`.



k
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to