I don't see any very bad.
Try rename your racks from numbers to unique string, for example change
rack 1 {
to
rack rack1 {
and i.e.
On 09.09.2013, at 23:56, Gaylord Holder wrote:
> Thanks for your assistance.
>
> Crush map:
>
> # begin crush map
> tunable choose_local_tries 0
> tunable choose_local_fallback_tries 0
> tunable choose_total_tries 50
> tunable chooseleaf_descend_once 1
>
> # devices
> device 0 osd.0
> device 1 osd.1
> device 2 osd.2
> device 3 osd.3
> device 4 osd.4
> device 5 osd.5
> device 6 osd.6
> device 7 osd.7
> device 8 osd.8
> device 9 osd.9
> device 10 osd.10
> device 11 osd.11
>
> # types
> type 0 osd
> type 1 host
> type 2 rack
> type 3 row
> type 4 room
> type 5 datacenter
> type 6 root
>
> # buckets
> host celestia {
> id -2 # do not change unnecessarily
> # weight 7.600
> alg straw
> hash 0 # rjenkins1
> item osd.0 weight 1.900
> item osd.1 weight 1.900
> item osd.2 weight 1.900
> item osd.3 weight 1.900
> }
> rack 1 {
> id -3 # do not change unnecessarily
> # weight 7.600
> alg straw
> hash 0 # rjenkins1
> item celestia weight 7.600
> }
> host luna {
> id -4 # do not change unnecessarily
> # weight 7.600
> alg straw
> hash 0 # rjenkins1
> item osd.5 weight 1.900
> item osd.6 weight 1.900
> item osd.7 weight 1.900
> item osd.4 weight 1.900
> }
> rack 2 {
> id -5 # do not change unnecessarily
> # weight 7.600
> alg straw
> hash 0 # rjenkins1
> item luna weight 7.600
> }
> host twilight {
> id -6 # do not change unnecessarily
> # weight 7.600
> alg straw
> hash 0 # rjenkins1
> item osd.8 weight 1.900
> item osd.10 weight 1.900
> item osd.11 weight 1.900
> item osd.9 weight 1.900
> }
> rack 3 {
> id -7 # do not change unnecessarily
> # weight 7.600
> alg straw
> hash 0 # rjenkins1
> item twilight weight 7.600
> }
> root default {
> id -1 # do not change unnecessarily
> # weight 22.800
> alg straw
> hash 0 # rjenkins1
> item 1 weight 7.600
> item 2 weight 7.600
> item 3 weight 7.600
> }
>
> # rules
> rule data {
> ruleset 0
> type replicated
> min_size 1
> max_size 10
> step take default
> step chooseleaf firstn 0 type host
> step emit
> }
> rule metadata {
> ruleset 1
> type replicated
> min_size 1
> max_size 10
> step take default
> step chooseleaf firstn 0 type host
> step emit
> }
> rule rbd {
> ruleset 2
> type replicated
> min_size 1
> max_size 10
> step take default
> step chooseleaf firstn 0 type host
> step emit
> }
>
> # end crush map
>
> The full osds are 2 and 10.
>
> -Gaylord
>
> On 09/09/2013 03:49 PM, Timofey wrote:> Show crush map please
> >
> > 09.09.2013, в 21:32, Gaylord Holder <[email protected]> написал(а):
> >
> >> I'm starting to load up my ceph cluster.
> >>
> >> I currently have 12 2TB drives (10 up and in, 2 defined but down and out).
> >>
> >> rados df
> >>
> >> says I have 8TB free, but I have 2 nearly full OSDs.
> >>
> >> I don't understand how/why these two disks are filled while the others are
> >> relatively empty.
> >>
> >> How do I tell ceph to spread the data around more, and why isn't it
> >> already doing it?
> >>
> >> Thank you for helping me understand this system better.
> >>
> >> Cheers,
> >> -Gaylord
> >> _______________________________________________
> >> ceph-users mailing list
> >> [email protected]
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> On 09/09/2013 03:49 PM, Timofey wrote:
>> Show crush map please
>>
>> 09.09.2013, в 21:32, Gaylord Holder <[email protected]> написал(а):
>>
>>> I'm starting to load up my ceph cluster.
>>>
>>> I currently have 12 2TB drives (10 up and in, 2 defined but down and out).
>>>
>>> rados df
>>>
>>> says I have 8TB free, but I have 2 nearly full OSDs.
>>>
>>> I don't understand how/why these two disks are filled while the others are
>>> relatively empty.
>>>
>>> How do I tell ceph to spread the data around more, and why isn't it already
>>> doing it?
>>>
>>> Thank you for helping me understand this system better.
>>>
>>> Cheers,
>>> -Gaylord
>>> _______________________________________________
>>> ceph-users mailing list
>>> [email protected]
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com