[bug] wrong speed calculation in (--output-file) logfile
Hello. During download with wget I've redirected output into file with the following command: $ LC_ALL=C wget -o output 'ftp://mirror.yandex.ru/gentoo-distfiles/distfiles/OOo_3.0.0rc4_20080930_LinuxIntel_langpack_en-GB.tar.gz' I've set LC_ALL and LANG explicitly to be sure that this is not locale related problem. The output I saw in output file was: --2008-10-25 14:51:17-- ftp://mirror.yandex.ru/gentoo-distfiles/distfiles/OOo_3.0.0rc4_20080930_LinuxIntel_langpack_en-GB.tar.gz = `OOo_3.0.0rc4_20080930_LinuxIntel_langpack_en-GB.tar.gz.13' Resolving mirror.yandex.ru... 77.88.19.68 Connecting to mirror.yandex.ru|77.88.19.68|:21... connected. Logging in as anonymous ... Logged in! == SYST ... done.== PWD ... done. == TYPE I ... done. == CWD /gentoo-distfiles/distfiles ... done. == SIZE OOo_3.0.0rc4_20080930_LinuxIntel_langpack_en-GB.tar.gz ... 13633213 == PASV ... done.== RETR OOo_3.0.0rc4_20080930_LinuxIntel_langpack_en-GB.tar.gz ... done. Length: 13633213 (13M) 0K .. .. .. .. .. 0% 131K 1m41s 50K .. .. .. .. .. 0% 132K 1m40s 100K .. .. .. .. .. 1% 135K 99s 150K .. .. .. .. .. 1% 132K 99s 200K .. .. .. .. .. 1% 130K 99s 250K .. .. .. .. .. 2% 45.9K 2m9s 300K .. .. .. .. .. 2% 64.3M 1m50s [snip] 13250K .. .. .. .. .. 99% 131K 0s 13300K .. ...100% 134K=1m41s 2008-10-25 14:52:58 (132 KB/s) - `OOo_3.0.0rc4_20080930_LinuxIntel_langpack_en-GB.tar.gz.13' saved [13633213] Note the line above snip: 300K .. 2% 64.3M 1m50s This is impossible to download so much Mbytes as file is much less. I don't know why sometimes this number jumps, but in some cases it cause the following output at the end of download: 13300K .. ... 100% 26101G=1m45s Obviously I don't have possibility to download with such high (26101G=1m45s) speed. This is reproducible with wget 1.11.4. -- Peter.
Re: re-mirror + no-clobber
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 Jonathan Elsas wrote: ... I've issued the command wget -nc -r -l inf -H -D www.example.com,www2.example.com http://www.example.com but, I get the message: file 'www.example.com/index.html' already there; not retrieving. and the process exits. According to the man page files with .html suffix will be loaded off disk and parsed but this does not appear to be happening. Am I missing something? Yes. It has to download the files before they can be loaded from the disk and parsed. When it encounters a file at a given location, it doesn't have any way to know that that file corresponds to the one it's trying to download. Timestamping with -N may be more what you want, rather than -nc? I'm open to suggestions on clarifying the documentation. - -- Micah J. Cowan Programmer, musician, typesetting enthusiast, gamer. GNU Maintainer: wget, screen, teseq http://micah.cowan.name/ -BEGIN PGP SIGNATURE- Version: GnuPG v1.4.6 (GNU/Linux) Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org iD8DBQFJA7Ds7M8hyUobTrERAsONAJ0dqYh0av7rQ80F8JIcvxhZ1ee7fwCdFG+y AJJxMPVzHpmqAy7iGVRWmCU= =wwns -END PGP SIGNATURE-