Dieter Drossmann [EMAIL PROTECTED] writes:
I use a extra file with a long list of http entries. I included this
file with the -i option. After 154 downloads I got an error
message: Segmentation fault.
With wget 1.7.1 everything works well.
Is there a new limit of lines?
No, there's no
Boehn, Gunnar von [EMAIL PROTECTED] writes:
I think I found a bug in wget.
You did. But I believe your subject line is slightly incorrect. Wget
handles 0 length time intervals (see the assert message), but what it
doesn't handle are negative amounts. And indeed:
gettimeofday({1063461157,
Wget 1.5.3 is ancient.
You should be well advised to upgrade to the current stable version (1.8.2)
or better the latest development version (1.9beta) even if wget is currently
in develpment stasis due to lack of maintainer.
You can find more information how to get the sources at
Hello!
I have found the following bug with wget 1.8.1 (windows) :
I try to download picture of CD audio from this URL :
wget could get this picture from the web server, but can't write the
output file :
-
http://www.aligastore.com/query.dll/img?gcdFab=8811803124type=0
=
Gary Hargrave wrote:
--- Kalin KOZHUHAROV [EMAIL PROTECTED] wrote:
Well, I am sure it is wrong URL, but took some time till I pinpoint
it in RFC1808. Otherwise it would be very difficult to code URL
parser.
Ooops :-) It seems I was wrong...
BTW, did you try to click in your browser on that
--- Kalin KOZHUHAROV [EMAIL PROTECTED] wrote:
Well, I am sure it is wrong URL, but took some time till I pinpoint it
in RFC1808. Otherwise it would be very difficult to code URL parser.
What you actually try to convince us is that you can omit the
net-location (i.e. usually comes in the
I just realized, I didn't send this and some other post to the list, but
directly to the replier...
Gary Hargrave wrote:
wget does not seem to handle relative links in web pages
of the form
http:page3.html
According to my understanding of rfc1808 this is a valid
URL. When recursively
- Original Message -
From: Ken Senior [EMAIL PROTECTED]
There does not seem to be support to change disks when accessing a VMS
server via wget. Is this a bug or just a limitation?
Wget does plain old HTTP and FTP. I know nothing about VMS. Does it have
some strange syntax for discs?
On Tue, 17 Sep 2002, Nikolay Kuzmin wrote:
There is a bug in wget1.8.2 when username or password contains symbol ''.
I think you should change code in file src/url.c from
I disagree. The name and password fields must never contain a letter, as it
is a reserved letter in URL strings. If your
Mats Andrén wrote:
I found this problem when fetching files recursively:
What if the filenames of linked files from a www-page contains the
[]-characters? They are treated as some kind of patterns, and not just
the way they are. Clearly not desirable! Since wget just fetches the
I have run across this problem too. It is because with Linux 2.4.18 (and
other
versions??) in certain circumstances, gettimeofday() is broken and will
jump
backwards. See http://kt.zork.net/kernel-traffic/kt20020708_174.html#1.
Is there any particular reason for this assert? If there is,
Hartwig, Thomas wrote:
I got a assert exit of wget in retr.c in the function calc_rate
because msecs is 0 or lesser than 0 (in spare cases).
I don't know how perhaps because I have a big line to the server or
the wrong OS. To get worked with this I patched retr.c setting
msecs = 1 if equal
Your message of Thu, 20 Jun 2002 17:41:06 +0200:
Short answer use quotes.
Yeah, thanks. I thought it was the , but I wasn't aware that I could
avoid this by quoting.
Cannot write to `foto.php4?id=148087' (Invalid argument).
And this is the question mark problem search the archives or
Your message of Thu, 20 Jun 2002 15:49:52 +0200:
I supposed people would read the index.html. Since this is becoming
something of a faq I've now I've put a 00Readme.txt on the ftp server and a
Readme.txt in the binary archives, we'll see if that helps.
It should :-)
Kai
--
Kai Schätzl,
Try telnet www.sosi.cnrs.fr 80
if it connects type GET / HTTP/1.0 followed by two newlines. If you don't
get the output of the webserver you probably have a routing problem or
something else.
Heiko
--
-- PREVINET S.p.A.[EMAIL PROTECTED]
-- Via Ferretto, 1ph
Cédric Rosa wrote:
Hello,
First, scuse my english but I'm french.
When I try with wget (v 1.8.1) to download an url which is behind a router,
the software wait for ever even if I've specified a timeout.
With ethereal, I've seen that there is no response from the server (ACK
never
thanks for your help :)
I'm installing version 1.9 to check. I think this update may solve my
problem.
Cedric Rosa.
- Original Message -
From: Hack Kampbjørn [EMAIL PROTECTED]
To: Cédric Rosa [EMAIL PROTECTED]
Cc: [EMAIL PROTECTED]
Sent: Friday, June 21, 2002 7:27 PM
Subject: Re: Bug
I don't know why Wget dumps core on startup. Perhaps a gettext
problem? I have seen reports of failure on startup on Solaris, and it
strikes me that Wget could have picked up wrong or inconsistent
gettext.
Try unsetting the locale-related evnironment variables and seeing if
Wget works then.
In message Re: bug report and patch, HTTPS recursive get,
Ian Abbott wrote...
Thanks again for the bug report and the proposed patch. I thought some
of the scheme tests in recur.c were getting messy, so propose the
following patch that uses a function to check for similar schemes.
Thanks
On Wed, 15 May 2002 18:44:19 +0900, Kiyotaka Doumae [EMAIL PROTECTED]
wrote:
I found a bug of wget with HTTPS resursive get, and proposal
a patch.
Thanks for the bug report and the proposed patch. The current scheme
comparison checks are getting messy, so I'll write a function to check
schemes
On Fri, 3 May 2002 18:37:22 +0200, Emmanuel Jeandel
[EMAIL PROTECTED] wrote:
ejeandel@yoknapatawpha:~$ wget -r a:b
Segmentation fault
Patient: Doctor, it hurts when I do this
Doctor: Well don't do that then!
Seriously, this is already fixed in CVS.
On Don, 25 Apr 2002, Hrvoje Niksic wrote:
Hello,
Judging by the provided `strace' output, it seems that your problem is
caused by the network or perhaps even by the remote server. Wget
sleeps on `select', waiting for the connection to close. That should
not happen -- if the connection is
Peteris Krumins [EMAIL PROTECTED] writes:
GNU Wget 1.8
get: progress.c:673: create_image: Assertion `p - bp-buffer = bp-width' failed.
This problem has been fixed in Wget 1.8.1. Please upgrade.
On 01/02/2002 12:10:59 Mr.Fritz wrote:
After the https/robots.txt bug, doing a recursive wget to an https-only
server
gives me this error: it searches for http://servername/index.html but
there
is no server on port 80, so wget receives a Connection refused error and
quits. It should search for
An, Young Hun [EMAIL PROTECTED] writes:
if HTML document contains code like this
meta http-equiv=Refresh
wget may be crushed. It has 'refresh' but
does not have 'content'. Of course this is
incorrect HTML. But I found some pages at web :)
simply add check routine at 'tag_handle_meta'
On 17 Jan 2002 at 2:15, Hrvoje Niksic wrote:
Michael Jennings [EMAIL PROTECTED] writes:
WGet returns an error message when the .wgetrc file is terminated
with an MS-DOS end-of-file mark (Control-Z). MS-DOS is the
command-line language for all versions of Windows, so ignoring the
WGet returns an error message when the .wgetrc file is terminated
with an MS-DOS end-of-file mark (Control-Z). MS-DOS is the
command-line language for all versions of Windows, so ignoring the
end-of-file mark would make sense.
Ouch, I never thought of that. Wget opens files in binary mode and
On 21 Jan 2002 at 14:56, Thomas Lussnig wrote:
Why not just open the wgetrc file in text mode using
fopen(name, r) instead of rb? Does that introduce other
problems?
I think it has to do with comments because the defeinition is that
starting with '#' the rest of the line
is ignored. And
On 17/01/2002 07:34:05 Herold Heiko wrote:
[proper order restored]
-Original Message-
From: Hrvoje Niksic [mailto:[EMAIL PROTECTED]]
Sent: Thursday, January 17, 2002 2:15 AM
To: Michael Jennings
Cc: [EMAIL PROTECTED]
Subject: Re: Bug report: 1) Small error 2) Improvement to Manual
Herold Heiko [EMAIL PROTECTED] writes:
My personal idea is:
As a matter of fact no *windows* text editor I know of, even the
supplied windows ones (notepad, wordpad) AFAIK will add the ^Z at the
end of file.txt. Wget is a *windows* program (although running in
console mode), not a *Dos*
-
Obviously, this is completely your decision. You are right, only DOS editors make the
mistake. (It should be noted that DOS is MS Windows only command line language. It
isn't going away; even Microsoft supplies command line utilities with all versions of
its OSs. Yes, Windows will probably
From: Michael Jennings [mailto:[EMAIL PROTECTED]]
Obviously, this is completely your decision. You are right,
only DOS editors make the mistake. (It should be noted that
DOS is MS Windows only command line language. It isn't going
away; even Microsoft supplies command line utilities with
Brendan Ragan [EMAIL PROTECTED] writes:
This is the problem i'm having with an older wget (1.5.3) when i
enter the url
'http://www.tranceaddict.com/cgi-bin/songout.php?id=1217-dirty_dirtymonth=dec'
it goes
Connecting to www.tranceaddict.com:80... connected!
HTTP request sent,
Jean-Edouard BABIN [EMAIL PROTECTED] writes:
I found a little bug when we download from an deleted directory:
[...]
Thanks for the report.
I wouldn't consider it a real bug. Downloading things into a deleted
directory is bound to produce all kinds of problems.
The diagnostic message could
Vladimir Volovich [EMAIL PROTECTED] writes:
while downloading some file (via http) with wget 1.8, i got an error:
assertion failed: p - bp-buffer = bp-width, file progress.c, line 673
Abort (core dumped)
Thanks for the report. It's a known problem in 1.8, fixed by this
patch.
Index:
Pavel Stepchenko [EMAIL PROTECTED] writes:
Hello bug-wget,
$ wget --version
GNU Wget 1.8
$ wget
ftp://password:[EMAIL PROTECTED]:12345/Dir%20One/This.Is.Long.Name.Of.The.Directory/*
Warning: wildcards not supported in HTTP.
Oooops! But this is FTP url, not HTTP!
Are you using a
[EMAIL PROTECTED] writes:
I use a proxy server, and have a line in my .wgetrc that says
something like:
What version of Wget are you using? I believe this bug has been fixed
in Wget 1.7.1 and later.
By the way, your analysis is correct.
[EMAIL PROTECTED] writes:
Today I downloaded the new wget release (1.8) (I'm a huge fan of the
util btw ;p ) and have been trying out the rate-limit feature.
[...]
assertion p - bp-buffer = bp-width failed: file progress.c,
line 673
Thanks for the report. The bug shows with downloads whose
William H. Gilmore [EMAIL PROTECTED] writes:
I have recently tripped across a bug with the version of wget shipped
with RedHat 7.2. When I attempt to recursively retrieve a web tree
starting with an html link that contains a base href, wget apparently
limits all href to base href even if
On Thu, 22 Nov 2001, Tomas Hjelmberg wrote:
Sorry, but can't anybody say at least that I'm wrong when I state that the
meta name=robots tag doesn't work?
Has anyone got it to work under any circumstances?
Yes:
- State what you want to do.
- Describe how you do it.
- Describe what happens.
-
is unfourtnuately downloaded anyway...
-Original Message-
From: Daniel Stenberg [mailto:[EMAIL PROTECTED]]
Sent: 22 November 2001 13:15
To: Tomas Hjelmberg
Cc: Wget List
Subject: Re: bug?
On Thu, 22 Nov 2001, Tomas Hjelmberg wrote:
Sorry, but can't anybody say at least that I'm wrong when I state
Tomas Hjelmberg [EMAIL PROTECTED] writes:
Sorry, but can't anybody say at least that I'm wrong when I state that the
meta name=robots tag doesn't work?
Has anyone got it to work under any circumstances?
These are two different questions.
The answer to the first one is irrelevant, because
Tomas Hjelmberg [EMAIL PROTECTED] writes:
I want to exclude /var/www/html/tomas.html from being indexed.
It looks like:
[...]
meta name=robots content=noindex
titleTomas/title
[...]
I invoke wget with:
wget -r http://localhost
And tomas.html is unfourtnuately downloaded
Thanks!
I see, but then, how to exclude from being downloaded per file-basis?
-Original Message-
From: Hrvoje Niksic [mailto:[EMAIL PROTECTED]]
Sent: 22 November 2001 14:45
To: Wget List; Tomas Hjelmberg
Subject: Re: bug?
Tomas Hjelmberg [EMAIL PROTECTED] writes:
I want to exclude
Hi Tomas!
I see, but then, how to exclude from being downloaded per file-basis?
First, let me be a smartass:
Go to
http://www.acronymfinder.com
and lokk up
RTFM
Then, proceed to the docs of wget.
wget offers download restrictions on
host, directory and file name.
Search in the docs for
-H
-D
Good idea, but how do you do that on a per file basis,
if the files has nothing in common and it is thousands of them?
-Original Message-
From: Jens Roesner [mailto:[EMAIL PROTECTED]]
Sent: 23 November 2001 04:09
To: Tomas Hjelmberg
Cc: Wget List
Subject: Re: bug?
Hi Tomas!
I see
Isn't it a good idea to have an option to forbid wget to download files with
the tag:
meta name=robots content=noindex
-Original Message-
From: Jens Roesner [mailto:[EMAIL PROTECTED]]
Sent: 23 November 2001 04:09
To: Tomas Hjelmberg
Cc: Wget List
Subject: Re: bug?
Hi Tomas!
I see
On 22 Nov 2001, at 14:49, Tomas Hjelmberg wrote:
Thanks!
I see, but then, how to exclude from being downloaded per file-basis?
Put the following in the /robots.txt on your website
User-agent: *
Disallow: /tomas.html
See http://www.robotstxt.org/wc/exclusion-admin.html for more
info.
Dan Harkless [EMAIL PROTECTED] writes:
--- src/recur.c Sun Dec 17 20:28:20 2000
+++ src/recur.c.new Sun Mar 25 20:25:12 2001
@@ -165,7 +165,18 @@
first_time = 0;
}
else
+{
+u = newurl ();
+err = parseurl (this_url, u, 0);
+if (err == URLOK)
Hrvoje Niksic [EMAIL PROTECTED] writes:
To be sure that *all* HTML files are handled, I think the addition
needs to be triggered from within retrieve_url, say by calling a
"register_html_file_for_conversion()". I think I'll provide such a
fix tonight.
Sounds good. Wonder if it should be
"Jan D." [EMAIL PROTECTED] writes:
If I change main.c to do setlocale(LC_ALL, "") instead of
setlocale(LC_MESSAGES, ""), the problem goes away.
Not using LC_ALL is intentional. The problem with LC_ALL was that it
affected other things (time and number representations, is* macros,
etc.)
101 - 151 of 151 matches
Mail list logo