On 10/11/11 03:24, Andrew Daviel wrote: > > When downloading a large file over a high-latency (e.g. long physical > distance) high-bandwidth link, the download time is dominated by the > round-trip time for TCP handshakes. > > In the past tools such as bbftp have mitigated this effect by using > multiple streams, but required both a special server and client. > > Using the "range" header in HTTP/1.1, it is possible to start multiple > simultaneous requests for different portions of a file using a > standard Apache server, and achieve a significant speedup. > I have a proof-of-principle Perl script using threads which was able > to download a medium-sized file from Europe to Vancouver in half the > normal time. > > I wondered it this was of interest as an enhanscement for wget. > > regards
I think setting a big SO_RCVBUF should also fix your issue, by using big window sizes, and it's cleaner. OTOH, you need support from the TCP stack, and that won't trick per-connection rate limits that may be limiting you in the single-connection case.
