On 12/7/19 11:42 AM, Philippe D'Anjou wrote:
> Hi,
> the docs say the upmap mode is trying to achieve perfect distribution as
> to have equal amount of PGs/OSD.
> This is what I got(v14.2.4):
>
> 0 ssd 3.49219 1.00000 3.5 TiB 794 GiB 753 GiB 38 GiB 3.4 GiB 2.7
> TiB 22.20 0.32 82 up
> 1 ssd 3.49219 1.00000 3.5 TiB 800 GiB 751 GiB 45 GiB 3.7 GiB 2.7
> TiB 22.37 0.33 84 up
> 2 ssd 3.49219 1.00000 3.5 TiB 846 GiB 792 GiB 50 GiB 3.6 GiB 2.7
> TiB 23.66 0.35 88 up
> 3 ssd 3.49219 1.00000 3.5 TiB 812 GiB 776 GiB 33 GiB 3.3 GiB 2.7
> TiB 22.71 0.33 85 up
> 4 ssd 3.49219 1.00000 3.5 TiB 768 GiB 730 GiB 34 GiB 4.1 GiB 2.7
> TiB 21.47 0.31 83 up
> 6 ssd 3.49219 1.00000 3.5 TiB 765 GiB 731 GiB 31 GiB 3.3 GiB 2.7
> TiB 21.40 0.31 82 up
> 8 ssd 3.49219 1.00000 3.5 TiB 872 GiB 828 GiB 41 GiB 3.2 GiB 2.6
> TiB 24.40 0.36 85 up
> 10 ssd 3.49219 1.00000 3.5 TiB 789 GiB 743 GiB 42 GiB 3.3 GiB 2.7
> TiB 22.05 0.32 82 up
> 5 ssd 3.49219 1.00000 3.5 TiB 719 GiB 683 GiB 32 GiB 3.9 GiB 2.8
> TiB 20.12 0.29 78 up
> 7 ssd 3.49219 1.00000 3.5 TiB 741 GiB 698 GiB 39 GiB 3.8 GiB 2.8
> TiB 20.73 0.30 79 up
> 9 ssd 3.49219 1.00000 3.5 TiB 709 GiB 664 GiB 41 GiB 3.5 GiB 2.8
> TiB 19.82 0.29 78 up
> 11 ssd 3.49219 1.00000 3.5 TiB 858 GiB 834 GiB 22 GiB 2.4 GiB 2.7
> TiB 23.99 0.35 82 up
> 101 ssd 3.49219 1.00000 3.5 TiB 815 GiB 774 GiB 38 GiB 3.5 GiB 2.7
> TiB 22.80 0.33 80 up
> 103 ssd 3.49219 1.00000 3.5 TiB 827 GiB 783 GiB 40 GiB 3.3 GiB 2.7
> TiB 23.11 0.34 81 up
> 105 ssd 3.49219 1.00000 3.5 TiB 797 GiB 759 GiB 36 GiB 2.5 GiB 2.7
> TiB 22.30 0.33 81 up
> 107 ssd 3.49219 1.00000 3.5 TiB 840 GiB 788 GiB 50 GiB 2.8 GiB 2.7
> TiB 23.50 0.34 83 up
> 100 ssd 3.49219 1.00000 3.5 TiB 728 GiB 678 GiB 47 GiB 2.4 GiB 2.8
> TiB 20.36 0.30 78 up
> 102 ssd 3.49219 1.00000 3.5 TiB 764 GiB 750 GiB 12 GiB 2.2 GiB 2.7
> TiB 21.37 0.31 76 up
> 104 ssd 3.49219 1.00000 3.5 TiB 795 GiB 761 GiB 31 GiB 2.5 GiB 2.7
> TiB 22.22 0.33 78 up
> 106 ssd 3.49219 1.00000 3.5 TiB 730 GiB 665 GiB 62 GiB 2.8 GiB 2.8
> TiB 20.41 0.30 78 up
> 108 ssd 3.49219 1.00000 3.5 TiB 849 GiB 808 GiB 38 GiB 2.5 GiB 2.7
> TiB 23.73 0.35 92 up
> 109 ssd 3.49219 1.00000 3.5 TiB 798 GiB 754 GiB 41 GiB 2.7 GiB 2.7
> TiB 22.30 0.33 83 up
> 110 ssd 3.49219 1.00000 3.5 TiB 840 GiB 810 GiB 28 GiB 2.4 GiB 2.7
> TiB 23.49 0.34 85 up
> 111 ssd 3.49219 1.00000 3.5 TiB 788 GiB 741 GiB 45 GiB 2.5 GiB 2.7
> TiB 22.04 0.32 85 up
>
> PG's are badly distributed.
From what information do you draw that conclusion? You use about 22% on
all OSDs.
I suggest that you increase your PGs to at least 100 per OSD, that will
make distribution even better.
Wido
> ceph balancer status
> {
> "active": true,
> "plans": [],
> "mode": "upmap"
> }
>
> It is because of this?
> health: HEALTH_WARN
> Failed to send data to Zabbix
> 1 subtrees have overcommitted pool target_size_bytes
> 1 subtrees have overcommitted pool target_size_ratio
>
>
> Any ideas why its not working?
>
>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com