Daniel Kinzler schrieb:
> Also, several http clients don't like files over 2GB - this is because the 
> large
> number of bytes in the Length field causes an integer overflow (2GB is the 31
> bit limit). wget likes to die with a segmentation fault on those. I found that
> curl works.

wget supports such big files. Perhaps you're running an old version?

IMHO the benefits of separated files are similar to the disadvantages. A
side side benefit if it would be that hashes would be splitted, too. If
you were unlucky, knowing that 'something' (perhaps just a bit) on the
150GB you downloaded is wrong, is not that helpful.
So having hashes for file sections on the big ones, even if not
'standard' would be an improvement.


_______________________________________________
Wikitech-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Reply via email to