On August 7, 2003 02:45 pm, Ian Clarke wrote:
> The current bandwidth limiting mechanism seems to be causing serious
> problems.  It works by limiting the speed of transmission of data on a
> byte-per-byte basis.  Unfortunately this creates a situation where
> transfers of data occur more slowly, which means they take longer, which
> means that we have more concurrent transfers overall, which slows them
> down even further - and the cycle continues - with the entire Freenet
> network getting bogged down in a web of extremely slow transfers.
>
> The alternative is for a node to try to maximize the per-transfer
> connection speed by rejecting new requests when the upstream connection
> speed is maxed out.  Some claim that this is a terrible idea and will
> screw up routing because it will be impossible to get a node to accept a
> datarequest, but I disagree.

And I concur.   Its not an intuitive but I strongly suggest it will give much
better results than the current situation.

> Imagine you go to McDonalds and ask a server for some food, they take
> your order.  Now, you didn't know, but that server is actually serving
> 20 other people at the same time and consequently it takes you ages to
> get your food.  Wouldn't it be better if that server said "Sorry Sir,
> I'm really busy - please try another server".
>
> In short, by making a node try to service its existing transfers as
> quickly as it can, it gets them out of the way faster and can thus serve
> just as many requests as a node which accepts all requests but takes
> ages to serve each individual one.

Ed
_______________________________________________
devl mailing list
[EMAIL PROTECTED]
http://hawk.freenetproject.org:8080/cgi-bin/mailman/listinfo/devl

Reply via email to