Problem recursive download

2003-10-16 Thread Sergey Vasilevsky
I use wget 1.8.2
Try recursive downdload www.map-by.info/index.html, but wget stop in first
page.
Why?
index.html have links to another page.

/usr/local/bin/wget -np -r -N -nH --referer=http://map-by.info  -P
/tmp/www.map-by.info -D map-by.info http://map-by.info
http://www.map-by.info
--10:09:25--  http://map-by.info/
   = `/p4/poisk/spider/resource/www.map-by.info/index.html'
Resolving proxy.open.by... done.
Connecting to proxy.open.by[193.232.92.3]:8080... connected.
Proxy request sent, awaiting response... 200 OK
Length: ignored [text/html]
Server file no newer than local file
`/p4/poisk/spider/resource/www.map-by.info/index.html' -- not retrieving.

--10:09:25--  http://www.map-by.info/
   = `/p4/poisk/spider/resource/www.map-by.info/index.html'
Connecting to proxy.open.by[193.232.92.3]:8080... connected.
Proxy request sent, awaiting response... 200 OK
Length: ignored [text/html]
Server file no newer than local file
`/p4/poisk/spider/resource/www.map-by.info/index.html' -- not retrieving.


FINISHED --10:09:26--
Downloaded: 0 bytes in 0 files



RE: Problem recursive download

2003-10-16 Thread Sergey Vasilevsky
I think wget strong verify link syntax:
a href=about_rus.html onMouseOver=img_on('main21');
onMouseOut=img_off('main21')
That link have incorrect symbol ';' not quoted in a 

 -Original Message-
 From: Sergey Vasilevsky [mailto:[EMAIL PROTECTED]
 Sent: Thursday, October 16, 2003 10:15 AM
 To: [EMAIL PROTECTED]
 Subject: Problem recursive download


 I use wget 1.8.2
 Try recursive downdload www.map-by.info/index.html, but wget stop in first
 page.
 Why?
 index.html have links to another page.

 /usr/local/bin/wget -np -r -N -nH --referer=http://map-by.info  -P
 /tmp/www.map-by.info -D map-by.info http://map-by.info
 http://www.map-by.info
 --10:09:25--  http://map-by.info/
= `/p4/poisk/spider/resource/www.map-by.info/index.html'
 Resolving proxy.open.by... done.
 Connecting to proxy.open.by[193.232.92.3]:8080... connected.
 Proxy request sent, awaiting response... 200 OK
 Length: ignored [text/html]
 Server file no newer than local file
 `/p4/poisk/spider/resource/www.map-by.info/index.html' -- not retrieving.

 --10:09:25--  http://www.map-by.info/
= `/p4/poisk/spider/resource/www.map-by.info/index.html'
 Connecting to proxy.open.by[193.232.92.3]:8080... connected.
 Proxy request sent, awaiting response... 200 OK
 Length: ignored [text/html]
 Server file no newer than local file
 `/p4/poisk/spider/resource/www.map-by.info/index.html' -- not retrieving.


 FINISHED --10:09:26--
 Downloaded: 0 bytes in 0 files





Re: Problem recursive download

2003-10-16 Thread Hrvoje Niksic
This seems to work in my copy of 1.8.2.  Perhaps you have something in
your .wgetrc that breaks things?



Re: Problem recursive download

2003-10-16 Thread Hrvoje Niksic
Sergey Vasilevsky [EMAIL PROTECTED] writes:

 I think wget strong verify link syntax:
 a href=about_rus.html onMouseOver=img_on('main21');
 onMouseOut=img_off('main21')
 That link have incorrect symbol ';' not quoted in a 

You are right.  However, this has been fixed in Wget 1.9-beta, which
will interpret the above as:

a href=about_rus.html onmouseover=img_on('main21') ;=; 
onmouseout=img_off('main21')

In other words, the HREF part will be correctly picked up by Wget.

Wget 1.9 will be released soon.  If you want to try it out, get it
from http://fly.srk.fer.hr/~hniksic/wget/wget-1.9-b5.tar.gz.