Re: bug in wget 1.8.1/1.8.2

2003-09-16 Thread Hrvoje Niksic
Dieter Drossmann [EMAIL PROTECTED] writes: I use a extra file with a long list of http entries. I included this file with the -i option. After 154 downloads I got an error message: Segmentation fault. With wget 1.7.1 everything works well. Is there a new limit of lines? No, there's no

Re: bug in wget - wget break on time msec=0

2003-09-13 Thread Hrvoje Niksic
Boehn, Gunnar von [EMAIL PROTECTED] writes: I think I found a bug in wget. You did. But I believe your subject line is slightly incorrect. Wget handles 0 length time intervals (see the assert message), but what it doesn't handle are negative amounts. And indeed: gettimeofday({1063461157,

RE: Bug in total byte count for large downloads

2003-08-26 Thread Herold Heiko
Wget 1.5.3 is ancient. You should be well advised to upgrade to the current stable version (1.8.2) or better the latest development version (1.9beta) even if wget is currently in develpment stasis due to lack of maintainer. You can find more information how to get the sources at

Re: Bug with specials characters : can't write output file

2003-02-08 Thread Kalin KOZHUHAROV
Hello! I have found the following bug with wget 1.8.1 (windows) : I try to download picture of CD audio from this URL : wget could get this picture from the web server, but can't write the output file : - http://www.aligastore.com/query.dll/img?gcdFab=8811803124type=0 =

Re: Bug in relative URL handling

2003-01-26 Thread Kalin KOZHUHAROV
Gary Hargrave wrote: --- Kalin KOZHUHAROV [EMAIL PROTECTED] wrote: Well, I am sure it is wrong URL, but took some time till I pinpoint it in RFC1808. Otherwise it would be very difficult to code URL parser. Ooops :-) It seems I was wrong... BTW, did you try to click in your browser on that

Re: Bug in relative URL handling

2003-01-24 Thread Gary Hargrave
--- Kalin KOZHUHAROV [EMAIL PROTECTED] wrote: Well, I am sure it is wrong URL, but took some time till I pinpoint it in RFC1808. Otherwise it would be very difficult to code URL parser. What you actually try to convince us is that you can omit the net-location (i.e. usually comes in the

Re: Bug in relative URL handling

2003-01-23 Thread Kalin KOZHUHAROV
I just realized, I didn't send this and some other post to the list, but directly to the replier... Gary Hargrave wrote: wget does not seem to handle relative links in web pages of the form http:page3.html According to my understanding of rfc1808 this is a valid URL. When recursively

Re: bug or limitation of wget used to access VMS servers

2003-01-08 Thread Max Bowsher
- Original Message - From: Ken Senior [EMAIL PROTECTED] There does not seem to be support to change disks when accessing a VMS server via wget. Is this a bug or just a limitation? Wget does plain old HTTP and FTP. I know nothing about VMS. Does it have some strange syntax for discs?

Re: Bug with user:pass in URL

2002-09-16 Thread Daniel Stenberg
On Tue, 17 Sep 2002, Nikolay Kuzmin wrote: There is a bug in wget1.8.2 when username or password contains symbol ''. I think you should change code in file src/url.c from I disagree. The name and password fields must never contain a letter, as it is a reserved letter in URL strings. If your

Re: bug?

2002-09-11 Thread Thomas Lussnig
Mats Andrén wrote: I found this problem when fetching files recursively: What if the filenames of linked files from a www-page contains the []-characters? They are treated as some kind of patterns, and not just the way they are. Clearly not desirable! Since wget just fetches the

Re: [BUG] assert test msecs

2002-08-04 Thread Colin 't Hart
I have run across this problem too. It is because with Linux 2.4.18 (and other versions??) in certain circumstances, gettimeofday() is broken and will jump backwards. See http://kt.zork.net/kernel-traffic/kt20020708_174.html#1. Is there any particular reason for this assert? If there is,

Re: [BUG] assert test msecs

2002-08-01 Thread Max Bowsher
Hartwig, Thomas wrote: I got a assert exit of wget in retr.c in the function calc_rate because msecs is 0 or lesser than 0 (in spare cases). I don't know how perhaps because I have a big line to the server or the wrong OS. To get worked with this I patched retr.c setting msecs = 1 if equal

Re: Bug with specific URLs

2002-06-21 Thread Kai Schaetzl
Your message of Thu, 20 Jun 2002 17:41:06 +0200: Short answer use quotes. Yeah, thanks. I thought it was the , but I wasn't aware that I could avoid this by quoting. Cannot write to `foto.php4?id=148087' (Invalid argument). And this is the question mark problem search the archives or

Re: Bug with specific URLs

2002-06-21 Thread Kai Schaetzl
Your message of Thu, 20 Jun 2002 15:49:52 +0200: I supposed people would read the index.html. Since this is becoming something of a faq I've now I've put a 00Readme.txt on the ftp server and a Readme.txt in the binary archives, we'll see if that helps. It should :-) Kai -- Kai Schätzl,

RE: Bug with wget ? I need help.

2002-06-21 Thread Herold Heiko
Try telnet www.sosi.cnrs.fr 80 if it connects type GET / HTTP/1.0 followed by two newlines. If you don't get the output of the webserver you probably have a routing problem or something else. Heiko -- -- PREVINET S.p.A.[EMAIL PROTECTED] -- Via Ferretto, 1ph

Re: Bug with wget ? I need help.

2002-06-21 Thread Hack Kampbjørn
Cédric Rosa wrote: Hello, First, scuse my english but I'm french. When I try with wget (v 1.8.1) to download an url which is behind a router, the software wait for ever even if I've specified a timeout. With ethereal, I've seen that there is no response from the server (ACK never

Re: Bug with wget ? I need help.

2002-06-21 Thread Cédric Rosa
thanks for your help :) I'm installing version 1.9 to check. I think this update may solve my problem. Cedric Rosa. - Original Message - From: Hack Kampbjørn [EMAIL PROTECTED] To: Cédric Rosa [EMAIL PROTECTED] Cc: [EMAIL PROTECTED] Sent: Friday, June 21, 2002 7:27 PM Subject: Re: Bug

Re: Bug ?

2002-06-02 Thread Hrvoje Niksic
I don't know why Wget dumps core on startup. Perhaps a gettext problem? I have seen reports of failure on startup on Solaris, and it strikes me that Wget could have picked up wrong or inconsistent gettext. Try unsetting the locale-related evnironment variables and seeing if Wget works then.

Re: bug report and patch, HTTPS recursive get

2002-05-17 Thread Kiyotaka Doumae
In message Re: bug report and patch, HTTPS recursive get, Ian Abbott wrote... Thanks again for the bug report and the proposed patch. I thought some of the scheme tests in recur.c were getting messy, so propose the following patch that uses a function to check for similar schemes. Thanks

Re: bug report and patch, HTTPS recursive get

2002-05-15 Thread Ian Abbott
On Wed, 15 May 2002 18:44:19 +0900, Kiyotaka Doumae [EMAIL PROTECTED] wrote: I found a bug of wget with HTTPS resursive get, and proposal a patch. Thanks for the bug report and the proposed patch. The current scheme comparison checks are getting messy, so I'll write a function to check schemes

Re: Bug report

2002-05-04 Thread Ian Abbott
On Fri, 3 May 2002 18:37:22 +0200, Emmanuel Jeandel [EMAIL PROTECTED] wrote: ejeandel@yoknapatawpha:~$ wget -r a:b Segmentation fault Patient: Doctor, it hurts when I do this Doctor: Well don't do that then! Seriously, this is already fixed in CVS.

Re: Bug#144242: is this progress bar problem [http://bugs.debian.org/144242] fixed in cvs (2002-04-09+10)?

2002-04-25 Thread Noel Koethe
On Don, 25 Apr 2002, Hrvoje Niksic wrote: Hello, Judging by the provided `strace' output, it seems that your problem is caused by the network or perhaps even by the remote server. Wget sleeps on `select', waiting for the connection to close. That should not happen -- if the connection is

Re: bug

2002-02-18 Thread Hrvoje Niksic
Peteris Krumins [EMAIL PROTECTED] writes: GNU Wget 1.8 get: progress.c:673: create_image: Assertion `p - bp-buffer = bp-width' failed. This problem has been fixed in Wget 1.8.1. Please upgrade.

Re: BUG https + index.html

2002-02-01 Thread csaba . raduly
On 01/02/2002 12:10:59 Mr.Fritz wrote: After the https/robots.txt bug, doing a recursive wget to an https-only server gives me this error: it searches for http://servername/index.html but there is no server on port 80, so wget receives a Connection refused error and quits. It should search for

Re: bug when processing META tag.

2002-01-31 Thread Hrvoje Niksic
An, Young Hun [EMAIL PROTECTED] writes: if HTML document contains code like this meta http-equiv=Refresh wget may be crushed. It has 'refresh' but does not have 'content'. Of course this is incorrect HTML. But I found some pages at web :) simply add check routine at 'tag_handle_meta'

Re: Bug report: 1) Small error 2) Improvement to Manual

2002-01-21 Thread Ian Abbott
On 17 Jan 2002 at 2:15, Hrvoje Niksic wrote: Michael Jennings [EMAIL PROTECTED] writes: WGet returns an error message when the .wgetrc file is terminated with an MS-DOS end-of-file mark (Control-Z). MS-DOS is the command-line language for all versions of Windows, so ignoring the

Re: Bug report: 1) Small error 2) Improvement to Manual

2002-01-21 Thread Thomas Lussnig
WGet returns an error message when the .wgetrc file is terminated with an MS-DOS end-of-file mark (Control-Z). MS-DOS is the command-line language for all versions of Windows, so ignoring the end-of-file mark would make sense. Ouch, I never thought of that. Wget opens files in binary mode and

Re: Bug report: 1) Small error 2) Improvement to Manual

2002-01-21 Thread Ian Abbott
On 21 Jan 2002 at 14:56, Thomas Lussnig wrote: Why not just open the wgetrc file in text mode using fopen(name, r) instead of rb? Does that introduce other problems? I think it has to do with comments because the defeinition is that starting with '#' the rest of the line is ignored. And

RE: Bug report: 1) Small error 2) Improvement to Manual

2002-01-17 Thread csaba . raduly
On 17/01/2002 07:34:05 Herold Heiko wrote: [proper order restored] -Original Message- From: Hrvoje Niksic [mailto:[EMAIL PROTECTED]] Sent: Thursday, January 17, 2002 2:15 AM To: Michael Jennings Cc: [EMAIL PROTECTED] Subject: Re: Bug report: 1) Small error 2) Improvement to Manual

Re: Bug report: 1) Small error 2) Improvement to Manual

2002-01-17 Thread Hrvoje Niksic
Herold Heiko [EMAIL PROTECTED] writes: My personal idea is: As a matter of fact no *windows* text editor I know of, even the supplied windows ones (notepad, wordpad) AFAIK will add the ^Z at the end of file.txt. Wget is a *windows* program (although running in console mode), not a *Dos*

Re: Bug report: 1) Small error 2) Improvement to Manual

2002-01-17 Thread Michael Jennings
- Obviously, this is completely your decision. You are right, only DOS editors make the mistake. (It should be noted that DOS is MS Windows only command line language. It isn't going away; even Microsoft supplies command line utilities with all versions of its OSs. Yes, Windows will probably

RE: Bug report: 1) Small error 2) Improvement to Manual

2002-01-17 Thread Herold Heiko
From: Michael Jennings [mailto:[EMAIL PROTECTED]] Obviously, this is completely your decision. You are right, only DOS editors make the mistake. (It should be noted that DOS is MS Windows only command line language. It isn't going away; even Microsoft supplies command line utilities with

Re: Bug? (not following 302's, or following them incorrectly)

2002-01-13 Thread Hrvoje Niksic
Brendan Ragan [EMAIL PROTECTED] writes: This is the problem i'm having with an older wget (1.5.3) when i enter the url 'http://www.tranceaddict.com/cgi-bin/songout.php?id=1217-dirty_dirtymonth=dec' it goes Connecting to www.tranceaddict.com:80... connected! HTTP request sent,

Re: Bug if current folder don't existe

2001-12-29 Thread Hrvoje Niksic
Jean-Edouard BABIN [EMAIL PROTECTED] writes: I found a little bug when we download from an deleted directory: [...] Thanks for the report. I wouldn't consider it a real bug. Downloading things into a deleted directory is bound to produce all kinds of problems. The diagnostic message could

Re: bug in wget 1.8

2001-12-17 Thread Hrvoje Niksic
Vladimir Volovich [EMAIL PROTECTED] writes: while downloading some file (via http) with wget 1.8, i got an error: assertion failed: p - bp-buffer = bp-width, file progress.c, line 673 Abort (core dumped) Thanks for the report. It's a known problem in 1.8, fixed by this patch. Index:

Re: Bug report

2001-12-13 Thread Hrvoje Niksic
Pavel Stepchenko [EMAIL PROTECTED] writes: Hello bug-wget, $ wget --version GNU Wget 1.8 $ wget ftp://password:[EMAIL PROTECTED]:12345/Dir%20One/This.Is.Long.Name.Of.The.Directory/* Warning: wildcards not supported in HTTP. Oooops! But this is FTP url, not HTTP! Are you using a

Re: bug with cookies being filed using proxy host, not real host

2001-12-13 Thread Hrvoje Niksic
[EMAIL PROTECTED] writes: I use a proxy server, and have a line in my .wgetrc that says something like: What version of Wget are you using? I believe this bug has been fixed in Wget 1.7.1 and later. By the way, your analysis is correct.

Re: bug in wget rate limit feature

2001-12-10 Thread Hrvoje Niksic
[EMAIL PROTECTED] writes: Today I downloaded the new wget release (1.8) (I'm a huge fan of the util btw ;p ) and have been trying out the rate-limit feature. [...] assertion p - bp-buffer = bp-width failed: file progress.c, line 673 Thanks for the report. The bug shows with downloads whose

Re: Bug: Wget 1.7 (Red Hat 7.2 dist wget-1.7-3)

2001-12-05 Thread Hrvoje Niksic
William H. Gilmore [EMAIL PROTECTED] writes: I have recently tripped across a bug with the version of wget shipped with RedHat 7.2. When I attempt to recursively retrieve a web tree starting with an html link that contains a base href, wget apparently limits all href to base href even if

Re: bug?

2001-11-22 Thread Daniel Stenberg
On Thu, 22 Nov 2001, Tomas Hjelmberg wrote: Sorry, but can't anybody say at least that I'm wrong when I state that the meta name=robots tag doesn't work? Has anyone got it to work under any circumstances? Yes: - State what you want to do. - Describe how you do it. - Describe what happens. -

RE: bug?

2001-11-22 Thread Tomas Hjelmberg
is unfourtnuately downloaded anyway... -Original Message- From: Daniel Stenberg [mailto:[EMAIL PROTECTED]] Sent: 22 November 2001 13:15 To: Tomas Hjelmberg Cc: Wget List Subject: Re: bug? On Thu, 22 Nov 2001, Tomas Hjelmberg wrote: Sorry, but can't anybody say at least that I'm wrong when I state

Re: bug?

2001-11-22 Thread Hrvoje Niksic
Tomas Hjelmberg [EMAIL PROTECTED] writes: Sorry, but can't anybody say at least that I'm wrong when I state that the meta name=robots tag doesn't work? Has anyone got it to work under any circumstances? These are two different questions. The answer to the first one is irrelevant, because

Re: bug?

2001-11-22 Thread Hrvoje Niksic
Tomas Hjelmberg [EMAIL PROTECTED] writes: I want to exclude /var/www/html/tomas.html from being indexed. It looks like: [...] meta name=robots content=noindex titleTomas/title [...] I invoke wget with: wget -r http://localhost And tomas.html is unfourtnuately downloaded

RE: bug?

2001-11-22 Thread Tomas Hjelmberg
Thanks! I see, but then, how to exclude from being downloaded per file-basis? -Original Message- From: Hrvoje Niksic [mailto:[EMAIL PROTECTED]] Sent: 22 November 2001 14:45 To: Wget List; Tomas Hjelmberg Subject: Re: bug? Tomas Hjelmberg [EMAIL PROTECTED] writes: I want to exclude

Re: bug?

2001-11-22 Thread Jens Roesner
Hi Tomas! I see, but then, how to exclude from being downloaded per file-basis? First, let me be a smartass: Go to http://www.acronymfinder.com and lokk up RTFM Then, proceed to the docs of wget. wget offers download restrictions on host, directory and file name. Search in the docs for -H -D

RE: bug?

2001-11-22 Thread Tomas Hjelmberg
Good idea, but how do you do that on a per file basis, if the files has nothing in common and it is thousands of them? -Original Message- From: Jens Roesner [mailto:[EMAIL PROTECTED]] Sent: 23 November 2001 04:09 To: Tomas Hjelmberg Cc: Wget List Subject: Re: bug? Hi Tomas! I see

RE: bug?

2001-11-22 Thread Tomas Hjelmberg
Isn't it a good idea to have an option to forbid wget to download files with the tag: meta name=robots content=noindex -Original Message- From: Jens Roesner [mailto:[EMAIL PROTECTED]] Sent: 23 November 2001 04:09 To: Tomas Hjelmberg Cc: Wget List Subject: Re: bug? Hi Tomas! I see

RE: bug?

2001-11-22 Thread Ian Abbott
On 22 Nov 2001, at 14:49, Tomas Hjelmberg wrote: Thanks! I see, but then, how to exclude from being downloaded per file-basis? Put the following in the /robots.txt on your website User-agent: * Disallow: /tomas.html See http://www.robotstxt.org/wc/exclusion-admin.html for more info.

Re: Bug when converting links in wget 1.6

2001-03-30 Thread Hrvoje Niksic
Dan Harkless [EMAIL PROTECTED] writes: --- src/recur.c Sun Dec 17 20:28:20 2000 +++ src/recur.c.new Sun Mar 25 20:25:12 2001 @@ -165,7 +165,18 @@ first_time = 0; } else +{ +u = newurl (); +err = parseurl (this_url, u, 0); +if (err == URLOK)

Re: Bug when converting links in wget 1.6

2001-03-30 Thread Dan Harkless
Hrvoje Niksic [EMAIL PROTECTED] writes: To be sure that *all* HTML files are handled, I think the addition needs to be triggered from within retrieve_url, say by calling a "register_html_file_for_conversion()". I think I'll provide such a fix tonight. Sounds good. Wonder if it should be

Re: Bug in locale handling when using glibc-2.2.1

2001-01-21 Thread Hrvoje Niksic
"Jan D." [EMAIL PROTECTED] writes: If I change main.c to do setlocale(LC_ALL, "") instead of setlocale(LC_MESSAGES, ""), the problem goes away. Not using LC_ALL is intentional. The problem with LC_ALL was that it affected other things (time and number representations, is* macros, etc.)

<    1   2