Downloading multiple pages

2007-01-17 Thread graham hadgraft

I need some help using an application called interchange on my
webserver. I have a url www.somewebsite.com/questions/*.html I want to
index all files in that questions directory. I am using the following
wget command.

CODE

wget -r -l2  -A html -X cgi-bin -D www.somewebsite.co.uk/ -P
/home/httpd/vhosts/somewebsite.co.uk/catalogs/somewebsite/swish_site/
http://www.somewebsite.co.uk/questions/


This only index the index page of this folder. It wil not follow the
links on the page. What would be the appropriate command to use to
index all pages from that folder.


Possibly bug

2007-01-17 Thread Yuriy Padlyak

Hi,

Have been downloading slackware-11.0-install-dvd.iso, but It seems wget 
downloaded more then filesize and I found:


-445900K .. .. .. .. ..119%   
18.53 KB/s

in  wget-log.

Regards,
Yuriy Padlyak


Re: Possibly bug

2007-01-17 Thread M.
The file was probably being uploaded when you started downloading it, so
the HTTP server continued sending data even over the initially reported
filesize.

Just stop wget, and start it again with option -c to resume download.


MT

Le mercredi 17 janvier 2007 à 18:16 +0200, Yuriy Padlyak a écrit :
 Hi,
 
 Have been downloading slackware-11.0-install-dvd.iso, but It seems wget 
 downloaded more then filesize and I found:
 
 -445900K .. .. .. .. ..119%   
 18.53 KB/s
 in  wget-log.
 
 Regards,
 Yuriy Padlyak



Problems with progress on large files

2007-01-17 Thread Stas Boukarev
Wget aborts on huge files:

$ wget ftp://ftp.fsn.hu/testfiles/128T

get: progress.c:965: create_image: Assertion `p - bp-buffer =
bp-width' failed.
Aborted

But wget -q ftp://ftp.fsn.hu/testfiles/128T works fine.

I tried Wget 1.10+devel from svn trunk branch, Revision 2202.
Compiled with gcc 3.4.6 on 32-bit GNU/Linux system with glibc-2.3.6.

$ uname -a
Linux slack 2.6.19.2 #1 Thu Jan 11 11:40:47 MSK 2007 i686 pentium4 i386
GNU/Linux

-- 
With Best Regards, Stas.
All You Need Is Love!
Homo sum et nihil humani a me alienum puto.


Addition to problems with progress on large files

2007-01-17 Thread Stas Boukarev
I've noticed problem is related with ETA calculating
and on systems with high bandwidth it may not fail.

It better to test something like
$ wget --limit-rate=1024 ftp://ftp.fsn.hu/testfiles/128T

-- 
With Best Regards, Stas.
All You Need Is Love!
Homo sum et nihil humani a me alienum puto.


Re: Possibly bug

2007-01-17 Thread Steven M. Schweda
From: Yuriy Padlyak

 Have been downloading slackware-11.0-install-dvd.iso, but It seems wget
 downloaded more then filesize and I found: 
 
 -445900K .. .. .. .. ..119%
 18.53 KB/s 
 
 in  wget-log.

   As usual, it would help if you provided some basic information. 
Which wget version (wget -V)?  On which system type?  OS and version? 
Guesswork follows.

   Wget versions before 1.10 did not support large files, and a DVD
image could easily exceed 2GB.  Negative file sizes are a common symptom
when using a small-file program with large files.



   Steven M. Schweda   [EMAIL PROTECTED]
   382 South Warwick Street(+1) 651-699-9818
   Saint Paul  MN  55105-2547


Re: Downloading multiple pages

2007-01-17 Thread Steven M. Schweda
From: graham hadgraft

 I need some help using an application [...]

   You seem to need some help asking for help.

 wget -r -l2  -A html -X cgi-bin -D www.somewebsite.co.uk/ -P
 /home/httpd/vhosts/somewebsite.co.uk/catalogs/somewebsite/swish_site/
 http://www.somewebsite.co.uk/questions/
 
 This only index the index page of this folder. It wil not follow the
 links on the page. What would be the appropriate command to use to
 index all pages from that folder.

   Did it occur to you that it might matter which version of wget you're
using, and on which system type (and version)?  Or that it might be
difficult for someone else to guess what happens when no one else can
see the Web page which seems to be causing your trouble?  Does it
actually have links to other pages?



   Steven M. Schweda   [EMAIL PROTECTED]
   382 South Warwick Street(+1) 651-699-9818
   Saint Paul  MN  55105-2547