On Tue, Jan 27, 2009 at 1:23 AM, Raghavendra G <[email protected]>wrote:

> Hi,
>
> On Tue, Jan 27, 2009 at 3:27 AM, Sean Davis <[email protected]> wrote:
>
>> If I am putting together several volumes of varying sizes using
>> distribute, what type of load balancing should I expect?  I understand
>> hashing and it sounds like if the disk fills, then it is not used, but can I
>> use ALU scheduler to cut things off before the disk becomes full to allow
>> for growth of directories and files?  How are people approaching this?
>
>
> Distribute,  does not have any schedulers. The hashing as of now is sort of
> static in the sense that if the disk becomes full, further creation of files
> which happen to be scheduled to that node fail. Future versions of
> distribute will reschedule the files to different nodes.
>
>

Thanks, Raghavendra.

So, it sounds like Distribute is problematic for any inhomogeneous file
system (where bricks are of different sizes) or for systems that are not
meant as "archival" (that is, write once, read many).  I understand that for
boatloads of small files, performance is improved over unify by using
distribute, but it sounds like unify is currently the better option for my
situation.

Is it worthwhile pointing out these details on the wiki somewhere?  The
website appears to suggest that unify/schedulers are "legacy" systems, which
implies that they are inferior to rather than an alternative to Distribute.
However, in my situation, it appears that Unify is the only viable solution.

Thanks for the help.

Sean
_______________________________________________
Gluster-users mailing list
[email protected]
http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users

Reply via email to