Hello,
On Tue, 24 Feb 2015 06:28:25 +0000 [email protected]
wrote:
> Hello everyone,
>
> In my rather heterogeneous setup ...
>
> -1 54.36 root default
> -2 42.44 room 2.162
> -4 6.09 host le09091
> 3 2.03 osd.3 up 1
> 1 2.03 osd.1 up 1
> 9 2.03 osd.9 up 1
> -6 36.35 host le10970
> 4 7.27 osd.4 up 1
> 5 7.27 osd.5 up 1
> 6 7.27 osd.6 up 1
> 7 7.27 osd.7 up 1
> 8 7.27 osd.8 up 1
> -3 11.92 room 2.166
> -5 5.83 host le09086
> 2 2.03 osd.2 up 1
> 0 2.03 osd.0 up 1
> 10 1.77 osd.10 up 1
> -7 6.09 host le08544
> 11 2.03 osd.11 up 1
> 12 2.03 osd.12 up 1
> 13 2.03 osd.13 up 1
>
> ... using size =3 for all pools,
> the OSDs are not filled according to the weights
> (which correspond to the disk sizes in TB).
>
> le09086 (osd)
> /dev/sdb1 1.8T 1.5T 323G 83% /var/lib/ceph/osd/ceph-10
> /dev/sdc1 2.1T 1.7T 408G 81% /var/lib/ceph/osd/ceph-0
> /dev/sdd1 2.1T 1.7T 344G 84% /var/lib/ceph/osd/ceph-2
> le09091 (osd)
> /dev/sda1 2.1T 1.6T 447G 79% /var/lib/ceph/osd/ceph-9
> /dev/sdc1 2.1T 1.8T 317G 85% /var/lib/ceph/osd/ceph-3
> /dev/sdb1 2.1T 1.7T 384G 82% /var/lib/ceph/osd/ceph-1
> le10970 (osd)
> /dev/sdd1 7.3T 1.4T 5.9T 19% /var/lib/ceph/osd/ceph-6
> /dev/sdf1 7.3T 1.5T 5.9T 21% /var/lib/ceph/osd/ceph-8
> /dev/sde1 7.3T 1.6T 5.7T 22% /var/lib/ceph/osd/ceph-7
> /dev/sdc1 7.3T 1.4T 6.0T 19% /var/lib/ceph/osd/ceph-5
> /dev/sdb1 7.3T 1.5T 5.8T 21% /var/lib/ceph/osd/ceph-4
> le08544 (osd)
> /dev/sdc1 2.1T 1.6T 443G 79% /var/lib/ceph/osd/ceph-13
> /dev/sdb1 2.1T 1.7T 339G 84% /var/lib/ceph/osd/ceph-12
> /dev/sda1 2.1T 1.7T 375G 82% /var/lib/ceph/osd/ceph-11
>
> Clearly, I would like le10970 to be selected more often!
>
> Increasing pg(p)_num from 256 to 512 of all the pools
> didn't help.
>
> Optimal tunables are used ...
>
> tunable choose_local_tries 0
> tunable choose_local_fallback_tries 0
> tunable choose_total_tries 50
> tunable chooseleaf_descend_once 1
> tunable chooseleaf_vary_r 1
>
> ... as well as the default algo (straw) and hash (0)
> for all buckets. The ruleset is pretty much standard,
> too ...
>
> rule replicated_ruleset {
> ruleset 0
> type replicated
> min_size 1
> max_size 10
> step take default
> step chooseleaf firstn 0 type host
> step emit
> }
>
> ... so I would assume it should select 3 hosts
> (originating from root) according to the weights.
>
> It is ceph version 0.87.
>
> Is it the room bucket, which circumvents
> a distribution according tot he weights?
>
That would be my guess, as a similar 4 node cluster here with a flat, host
only topology is having its data distributed as expected by the weights.
Christian
> Thanks and all the best,
>
> Frank
>
>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
--
Christian Balzer Network/Systems Engineer
[email protected] Global OnLine Japan/Fusion Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com