On hdd failure the number of placement groups on the rest of osds on the
same host goes up. I would expect equal distribution of failed placement
groups across the cluster, not just on the troubled host. Shall the host
weight auto reduce whenever an osd gets out?
Exibit 1: Attached osd-df-tree file. Number of placement groups per osd
on healthy nodes across the cluster is around 160, see osd050 and
osd056. Number of placement groups per osd on nodes with hdd failures
goes noticeably up, more so as more hdd failures happen on the same
node, see osd051 and osd053.
This cluster can handle this case at this moment as it has got plenty of
free space. I wonder how is this going to play out when we get to 90% of
usage on the whole cluster. A single backplane failure in a node takes
four drives out at once; that is 30% of storage space on a node. The
whole cluster would have enough space to host the failed placement
groups but one node would not.
This cluster is running Nautilus 14.2.0 with default settings deployed
using ceph-ansible.
Milan
--
Milan Kupcevic
Senior Cyberinfrastructure Engineer at Project NESE
Harvard University
FAS Research Computing
> ceph osd df tree name osd050
ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL
%USE VAR PGS STATUS TYPE NAME
-130 110.88315 - 111 TiB 6.0 TiB 4.7 TiB 563 MiB 21 GiB 105 TiB
5.39 1.00 - host osd050
517 hdd 9.20389 1.00000 9.2 TiB 442 GiB 329 GiB 16 KiB 1.7 GiB 8.8 TiB
4.69 0.87 157 up osd.517
532 hdd 9.20389 1.00000 9.2 TiB 465 GiB 352 GiB 32 KiB 1.8 GiB 8.7 TiB
4.94 0.92 170 up osd.532
544 hdd 9.20389 1.00000 9.2 TiB 447 GiB 334 GiB 32 KiB 1.8 GiB 8.8 TiB
4.74 0.88 153 up osd.544
562 hdd 9.20389 1.00000 9.2 TiB 440 GiB 328 GiB 64 KiB 1.5 GiB 8.8 TiB
4.67 0.87 159 up osd.562
575 hdd 9.20389 1.00000 9.2 TiB 479 GiB 366 GiB 88 KiB 1.9 GiB 8.7 TiB
5.08 0.94 175 up osd.575
592 hdd 9.20389 1.00000 9.2 TiB 434 GiB 321 GiB 24 KiB 1.4 GiB 8.8 TiB
4.60 0.85 153 up osd.592
605 hdd 9.20389 1.00000 9.2 TiB 456 GiB 343 GiB 0 B 1.5 GiB 8.8 TiB
4.84 0.90 170 up osd.605
618 hdd 9.20389 1.00000 9.2 TiB 473 GiB 360 GiB 16 KiB 1.6 GiB 8.7 TiB
5.01 0.93 172 up osd.618
631 hdd 9.20389 1.00000 9.2 TiB 461 GiB 348 GiB 44 KiB 1.5 GiB 8.8 TiB
4.89 0.91 165 up osd.631
644 hdd 9.20389 1.00000 9.2 TiB 459 GiB 346 GiB 92 KiB 1.7 GiB 8.8 TiB
4.87 0.90 163 up osd.644
656 hdd 9.20389 1.00000 9.2 TiB 433 GiB 320 GiB 68 KiB 1.4 GiB 8.8 TiB
4.59 0.85 156 up osd.656
669 hdd 9.20389 1.00000 9.2 TiB 1.1 TiB 1019 GiB 36 KiB 2.6 GiB 8.1 TiB
12.01 2.23 169 up osd.669
682 ssd 0.43649 1.00000 447 GiB 3.1 GiB 2.1 GiB 562 MiB 462 MiB 444 GiB
0.69 0.13 168 up osd.682
TOTAL 111 TiB 6.0 TiB 4.7 TiB 563 MiB 21 GiB 105 TiB
5.39
MIN/MAX VAR: 0.13/2.23 STDDEV: 2.32
> ceph osd df tree name osd051
ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL
%USE VAR PGS STATUS TYPE NAME
-148 110.88315 - 83 TiB 4.9 TiB 4.0 TiB 573 MiB 20 GiB 78 TiB
5.94 1.00 - host osd051
408 hdd 9.20389 0 0 B 0 B 0 B 0 B 0 B 0 B
0 0 0 down osd.408
538 hdd 9.20389 1.00000 9.2 TiB 542 GiB 429 GiB 24 KiB 2.4 GiB 8.7 TiB
5.75 0.97 212 up osd.538
552 hdd 9.20389 0 0 B 0 B 0 B 0 B 0 B 0 B
0 0 0 down osd.552
565 hdd 9.20389 0 0 B 0 B 0 B 0 B 0 B 0 B
0 0 0 down osd.565
578 hdd 9.20389 1.00000 9.2 TiB 557 GiB 444 GiB 56 KiB 2.0 GiB 8.7 TiB
5.91 0.99 213 up osd.578
590 hdd 9.20389 1.00000 9.2 TiB 533 GiB 420 GiB 34 KiB 2.4 GiB 8.7 TiB
5.66 0.95 212 up osd.590
603 hdd 9.20389 1.00000 9.2 TiB 562 GiB 449 GiB 76 KiB 2.2 GiB 8.7 TiB
5.96 1.00 218 up osd.603
616 hdd 9.20389 1.00000 9.2 TiB 553 GiB 440 GiB 16 KiB 2.2 GiB 8.7 TiB
5.86 0.99 217 up osd.616
629 hdd 9.20389 1.00000 9.2 TiB 579 GiB 466 GiB 40 KiB 2.0 GiB 8.6 TiB
6.14 1.03 228 up osd.629
642 hdd 9.20389 1.00000 9.2 TiB 588 GiB 475 GiB 40 KiB 2.6 GiB 8.6 TiB
6.23 1.05 228 up osd.642
655 hdd 9.20389 1.00000 9.2 TiB 583 GiB 470 GiB 32 KiB 2.3 GiB 8.6 TiB
6.18 1.04 223 up osd.655
668 hdd 9.20389 1.00000 9.2 TiB 570 GiB 457 GiB 32 KiB 1.9 GiB 8.6 TiB
6.05 1.02 229 up osd.668
681 ssd 0.43649 1.00000 447 GiB 3.1 GiB 2.1 GiB 573 MiB 451 MiB 444 GiB
0.69 0.12 167 up osd.681
TOTAL 83 TiB 4.9 TiB 4.0 TiB 573 MiB 20 GiB 78 TiB
5.94
MIN/MAX VAR: 0.12/1.05 STDDEV: 1.67
> ceph osd df tree name osd053
ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL
%USE VAR PGS STATUS TYPE NAME
-136 110.88315 - 74 TiB 4.8 TiB 4.0 TiB 447 MiB 18 GiB 69 TiB
6.53 1.00 - host osd053
519 hdd 9.20389 1.00000 9.2 TiB 665 GiB 552 GiB 52 KiB 2.2 GiB 8.6 TiB
7.05 1.08 256 up osd.519
534 hdd 9.20389 1.00000 9.2 TiB 654 GiB 541 GiB 44 KiB 2.2 GiB 8.6 TiB
6.94 1.06 261 up osd.534
546 hdd 9.20389 1.00000 9.2 TiB 641 GiB 528 GiB 46 KiB 2.2 GiB 8.6 TiB
6.80 1.04 251 up osd.546
558 hdd 9.20389 1.00000 9.2 TiB 581 GiB 468 GiB 20 KiB 2.0 GiB 8.6 TiB
6.17 0.95 232 up osd.558
571 hdd 9.20389 1.00000 9.2 TiB 594 GiB 481 GiB 68 KiB 2.1 GiB 8.6 TiB
6.30 0.97 240 up osd.571
583 hdd 9.20389 1.00000 9.2 TiB 664 GiB 551 GiB 68 KiB 2.5 GiB 8.6 TiB
7.05 1.08 268 up osd.583
596 hdd 9.20389 1.00000 9.2 TiB 569 GiB 456 GiB 8 KiB 2.4 GiB 8.6 TiB
6.04 0.93 218 up osd.596
609 hdd 9.20389 1.00000 9.2 TiB 580 GiB 467 GiB 8 KiB 2.2 GiB 8.6 TiB
6.15 0.94 231 up osd.609
622 hdd 9.20389 0 0 B 0 B 0 B 0 B 0 B 0 B
0 0 0 down osd.622
635 hdd 9.20389 0 0 B 0 B 0 B 0 B 0 B 0 B
0 0 0 down osd.635
648 hdd 9.20389 0 0 B 0 B 0 B 0 B 0 B 0 B
0 0 0 down osd.648
661 hdd 9.20389 0 0 B 0 B 0 B 0 B 0 B 0 B
0 0 0 down osd.661
674 ssd 0.43649 1.00000 447 GiB 3.1 GiB 2.1 GiB 447 MiB 577 MiB 444 GiB
0.70 0.11 143 up osd.674
TOTAL 74 TiB 4.8 TiB 4.0 TiB 447 MiB 18 GiB 69 TiB
6.53
MIN/MAX VAR: 0.11/1.08 STDDEV: 1.98
> ceph osd df tree name osd056
ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL
%USE VAR PGS STATUS TYPE NAME
-160 110.88315 - 111 TiB 5.2 TiB 3.9 TiB 478 MiB 17 GiB 106 TiB
4.70 1.00 - host osd056
528 hdd 9.20389 1.00000 9.2 TiB 450 GiB 337 GiB 0 B 1.5 GiB 8.8 TiB
4.78 1.02 163 up osd.528
542 hdd 9.20389 1.00000 9.2 TiB 422 GiB 309 GiB 28 KiB 1.2 GiB 8.8 TiB
4.48 0.95 145 up osd.542
555 hdd 9.20389 1.00000 9.2 TiB 397 GiB 284 GiB 0 B 1.1 GiB 8.8 TiB
4.21 0.90 147 up osd.555
568 hdd 9.20389 1.00000 9.2 TiB 400 GiB 287 GiB 32 KiB 1.1 GiB 8.8 TiB
4.25 0.90 141 up osd.568
579 hdd 9.20389 1.00000 9.2 TiB 532 GiB 419 GiB 0 B 1.5 GiB 8.7 TiB
5.64 1.20 196 up osd.579
591 hdd 9.20389 1.00000 9.2 TiB 451 GiB 339 GiB 112 KiB 1.5 GiB 8.8 TiB
4.79 1.02 163 up osd.591
604 hdd 9.20389 1.00000 9.2 TiB 463 GiB 350 GiB 64 KiB 1.2 GiB 8.8 TiB
4.92 1.05 168 up osd.604
617 hdd 9.20389 1.00000 9.2 TiB 455 GiB 343 GiB 12 KiB 1.3 GiB 8.8 TiB
4.83 1.03 168 up osd.617
630 hdd 9.20389 1.00000 9.2 TiB 407 GiB 294 GiB 32 KiB 1.6 GiB 8.8 TiB
4.32 0.92 151 up osd.630
643 hdd 9.20389 1.00000 9.2 TiB 447 GiB 335 GiB 16 KiB 1.3 GiB 8.8 TiB
4.75 1.01 152 up osd.643
659 hdd 9.20389 1.00000 9.2 TiB 464 GiB 351 GiB 20 KiB 1.3 GiB 8.8 TiB
4.92 1.05 167 up osd.659
672 hdd 9.20389 1.00000 9.2 TiB 441 GiB 328 GiB 44 KiB 1.4 GiB 8.8 TiB
4.68 1.00 158 up osd.672
685 ssd 0.43649 1.00000 447 GiB 3.1 GiB 2.1 GiB 478 MiB 546 MiB 444 GiB
0.70 0.15 156 up osd.685
TOTAL 111 TiB 5.2 TiB 3.9 TiB 478 MiB 17 GiB 106 TiB
4.70
MIN/MAX VAR: 0.15/1.20 STDDEV: 1.17
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com