Raúl, I'm not sure I understand what exactly you are trying to do. I
understand that you set page_requisites to on, but I don't see why you
feel you must activate recursion to download non-HTML files through
ftp.
Could you be more specific in how you invoke Wget, what you expect it
to do, and
DervishD [EMAIL PROTECTED] writes:
Yes. The timeout option is a shortcut for specifying all the
other options.
What I mean is: if I don't specify 'timeout' in my wgetrc it has the
default value ('900', AFAIK), but it is ignored?
I'm afraid I don't understand this question. Is what
[EMAIL PROTECTED] writes:
Well, to find out what was happening, I specified -d for the debug
output. The message was: debug support not compiled in
[...]
Is this an oversight or does it serve a purpose?
Heiko will know for sure, but it's most likely an oversight. The
Windows config.h.* files
Wget's default behavior is not to span hosts. However, a redirect
does not count as spanning because the redirected contents is
considered as a different name for the original contents, which *was*
requested.
Note that Wget will not follow the links from the redirected document
further into the
It seems that Apache's fnmatch.h is shadowing the one from libc.
Please remove the former and your build problems should go away.
[EMAIL PROTECTED] writes:
Will wget build me such a copy of the entire site? Full interlinked
and spiderable?
Yes, with several buts.
1. Your site should be written and interlinked in fairly discernable
HTML. No image rollovers linked only through JavaScript. No CSS
imports.
2.
To unsubscribe, send a message to [EMAIL PROTECTED].
[EMAIL PROTECTED] writes:
Will wget build me such a copy of the entire site? Full interlinked
and spiderable?
The command to make the copy would be something like
`wget --mirror --convert-links --html-extension URL'.
I started wget with
wget --mirror --convert-links --html-extension
DervishD [EMAIL PROTECTED] writes:
Right :)) If you want I can prepare the patch for you, containing
too a typo in the documentation.
I think I'll modify the Makefile. A patch that fixes (or points out)
the typo in the documentation would be appreciated, though.
BTW, in the documentation
Wget 1.9 has been uploaded to the GNU ftp site. I'll write an
announcement shortly.
Thanks to everyone who contributed to this release.
war [EMAIL PROTECTED] writes:
Using gcc-3.3.2:
[...]
Thanks for the report. I'll need a few more details, though.
What operating system is this? Does Wget 1.8.2 compile on the same
system?
Hmm. It seems that one of the enums is clashing with something on the
system -- I'm betting it's GETALL. I don't know why this never
happened with older versions.
Does it compile if you add the FTP_ prefix to the names of the
constants, both in ftp.h and ftp.c?
Frank Klemm [EMAIL PROTECTED] writes:
Wget don't work properly when the URL contains characters which are
not allowed in file names on the file system which is currently
used. These are often '\', '?', '*' and ':'.
Affected are at least:
- Windows and related OS
- Linux when using FAT or
Tony Lewis [EMAIL PROTECTED] writes:
Philip Mateescu wrote:
A warning message would be nice when for not so obvious reasons wget
doesn't behave as one would expect.
I don't know if there are other tags that could change wget's behavior
(like -r and meta name=robots do), but if they happen
Patrick Robinson [EMAIL PROTECTED] writes:
would someone be so kind to tell me the exact syntax to tell wget which
extensions it should reject, e.g. mp3 or mpg
-R mp3 should do the trick.
Gisle Vanem [EMAIL PROTECTED] writes:
#ifndef ENOTCONN
# define ENOTCONN X_ENOTCONN
#endif
Except you cannot make Winsock return X_ENOTCONN.
But we don't really care because we're in control of what gets stores
into errno after Winsock calls. So instead of:
errno = WSAGetLastError ();
??? ?? [EMAIL PROTECTED] writes:
I've seen pages that do that kind of redirections, but Wget seems
to follow them, for me. Do you have an example I could try?
[EMAIL PROTECTED]:~/ /usr/local/bin/wget -U
All.by -np -r -N -nH --header=Accept-Charset: cp1251, windows-1251, win,
Tony Lewis [EMAIL PROTECTED] writes:
Hrvoje Niksic wrote:
Incidentally, Wget is not the only browser that has a problem with
that. For me, Mozilla is simply showing the source of
http://www.minskshop.by/cgi-bin/shop.cgi?id=1cookie=set, because
the returned content-type is text/plain
Aaron S. Hawley [EMAIL PROTECTED] writes:
The HTML of those pages contains the meta-tag
meta name=robots content=noindex,nofollow /
and Wget listened, and only downloaded the first page.
Perhaps Wget should give a warning message that the file contained a
meta-robots tag, so that people
In case you're curious, I'm still waiting for a response from the GNU
people.
If I don't hear from them soon, I'll release 1.9 anyway and put it on
a private FTP site. That way there will be a release for Noel to
package for Debian and we can watch the fun as the bug reports start
pouring.
This seems to work in my copy of 1.8.2. Perhaps you have something in
your .wgetrc that breaks things?
Sergey Vasilevsky [EMAIL PROTECTED] writes:
I think wget strong verify link syntax:
a href=about_rus.html onMouseOver=img_on('main21');
onMouseOut=img_off('main21')
That link have incorrect symbol ';' not quoted in a
You are right. However, this has been fixed in Wget 1.9-beta, which
As the name implies, this should be 1.9 (with only version changed)
unless a show-stopper is discovered. Get it from:
http://fly.srk.fer.hr/~hniksic/wget/wget-1.9-rc1.tar.gz
Tony Lewis [EMAIL PROTECTED] writes:
Hrvoje Niksic wrote:
I'm about to release 1.9 today, unless it takes more time to upload it
to ftp.gnu.org.
If there's a serious problem you'd like fixed in 1.9, speak up now or
be silent until 1.9.1. :-)
I thought we were going to turn our
[ Moving discussion from wget-patches to wget. ]
Gisle Vanem [EMAIL PROTECTED] writes:
I'm pretty sure that other GNU applications -- that have also been
ported to Windows -- use errno. I wonder how they do it...
Lynx uses this:
#define SOCKET_ERRNO errno
#ifdef WINDOWS
#undef
Gisle Vanem [EMAIL PROTECTED] writes:
There is another possible approach. We already #define read and write
to call Winsock stuff. We could add some more magic so that they and
other Winsock invocations automatically set errno to last error value,
translating Windows errors to errno errors.
Unfortunately, I don't know what the problem is here. Perhaps some of
the Windows people can take over this one?
Gisle Vanem [EMAIL PROTECTED] writes:
Error in wget-1.9-b5.zip
wget cannot find the host. Turn on -d option and observe:
Location: http://www.yourworstenemy.com?tgpid=008drefid=393627 [following]
Closing fd 1952
--13:38:35-- http://www.yourworstenemy.com/?tgpid=008drefid=393627
Gisle Vanem [EMAIL PROTECTED] writes:
And anyway, aren't Winsock functions supposed to set errno? If they
don't, how does Wget report connection refused, to name but one
example? Wget relies on errno/strerror pretty heavily, and if that
were non-functional in such an obvious way, I'm sure
Tony Lewis [EMAIL PROTECTED] writes:
I'm trying to figure out how to do a POST followed by a GET.
If I do something like:
wget http://www.somesite.com/post.cgi --post-data 'a=1b=2'
http://www.somesite.com/getme.html -d
Well... `--post-data' currently affects all the URLs in the Wget run.
Sergey Vasilevsky [EMAIL PROTECTED] writes:
I use wget 1.8.2. When I try recursive download site site.com where
site.com/ first page redirect to site.com/xxx.html that have first
link in the page to site.com/ then Wget download only xxx.html and
stop. Other links from xxx.html not followed!
Sergey Vasilevsky [EMAIL PROTECTED] writes:
Have wget any rules to convert retrive url to store url? Or may be
in future?
For example:
Get - site.com/index.php?PHPSESSID=123124324
Filter - /PHPSESSID=[a-z0-9]+//i
Save as - site.com/index.php
The problem with this is that it would
I like these suggestions. How about the following: for 1.9, document
that `--post-data' expects one URL and that its behavior for multiple
specified URLs might change in a future version.
Then, for 1.10 we can implement one of the alternative behaviors.
You're right -- that code was broken. Thanks for the patch; I've now
applied it to CVS with the following ChangeLog entry:
2003-10-15 Philip Stadermann [EMAIL PROTECTED]
* ftp.c (ftp_retrieve_glob): Correctly loop through the list whose
elements might have been deleted.
Thanks for the report. I agree that the current code does not work
for many uses -- that's why IPv6 is still experimental. Mauro
Tortonesi is working on contributing IPv6 support that works better.
For the impending release, I think the workaround you posted makes
sense. Mauro, what do you
Simons, Rick [EMAIL PROTECTED] writes:
I'm having trouble with downloading a file across https using wget.
I can't figure out if it is something i'm doing wrong with wget
syntax, or if the httpd server isn't working like it should.
I don't know what's goingn wrong here. Your Wget syntax
Simons, Rick [EMAIL PROTECTED] writes:
Using 1.9 I get a different error ...
[...]
using 1.9b5
# ./wget https://filed1/InBox/FILE3 --http-user=user --http-passwd=pass
https://filed1/InBox/FILE3: Unsupported scheme.
That just means that you haven't compiled 1.9-b5 with SSL. Did you
compile
Simons, Rick [EMAIL PROTECTED] writes:
I believe 1.8 was an rpm install, but I could be mistaken. You are
right about the 1.9 install .. it was just a config/make/make
install on the tar I nabbed. How can I determine if I have SSL
includes on a RH9 box?
I think you need to install the
Herold Heiko [EMAIL PROTECTED] writes:
This beta includes portability tweaks and minor improvements. Please
test it on as many diverse platforms as possible, preferrably with
both gcc and non-gcc compilers. If all goes well, I'd like to release
1.9 perhaps as early as tomorrow.
Windows,
Herold Heiko [EMAIL PROTECTED] writes:
From: Hrvoje Niksic [mailto:[EMAIL PROTECTED]
Does it compile if you change #define HAVE_U_INT32_T 1 to #undef
HAVE_U_INT32_T in config.h.ms?
It does.
Windows msvc binary at http://xoomer.virgilio.it/hherold
Cool. BTW does MSVC have int32_t?
Herold Heiko [EMAIL PROTECTED] writes:
C:\Programmi\Microsoft Visual Studio\VC98\Include\Native.h has:
typedef long int32_t;
I see. We're currently not using signed 32-bit variables, so I guess
the point is moot.
Suhas Tembe [EMAIL PROTECTED] writes:
There are two steps involved:
1). Log in to the customer's web site. I was able to create the following link after
I looked at the form section in the source as explained to me earlier by Hrvoje.
wget
Suhas Tembe [EMAIL PROTECTED] writes:
I tried, but it doesn't seem to have worked. This what I did:
wget --save-cookies=cookies.txt
http://customer.website.com?UserAccount=USERAccessCode=PASSWORDLocale=English
(United States)TimeZone=(GMT-5:00) Eastern Standard Time (USA amp;
Jens Rösner [EMAIL PROTECTED] writes:
Well, I am by no means an expert, but I think that wget closes the
connection after the first retrieval. The SSL server realizes this
and decides that wget has no right to log in for the second
retrieval, eventhough the cookie is there. I think that is a
Suhas Tembe [EMAIL PROTECTED] writes:
Cookies.txt looks like this:
# HTTP cookie file.
# Generated by Wget on 2003-10-13 13:19:26.
# Edit at your own risk.
There is nothing after the 3rd line. So, it doesn't look like a
valid cookie file.
It's valid all right, but there are no cookies
Suhas Tembe [EMAIL PROTECTED] writes:
The other thing I noticed is that the first URL (to log in) does not
seem to work, because when I use that same URL in IE, it brings me
back to the login screen (see attached source of the login
page). I don't get logged-in.
Why are you using that URL if
Jens Rösner [EMAIL PROTECTED] writes:
Hi Hrvoje!
retrieval, eventhough the cookie is there. I think that is a
correct behaviour for a secure server, isn't it?
Why would it be correct?
Sorry, I seem to have been misled by my own (limited) experience:
From the few secure sites I use,
Gisle Vanem [EMAIL PROTECTED] writes:
It seems touch() is called on an open file and hence utime() is
either silently ignored or causing Access denied on Watcom.
[...]
wget -d -Otcpdump.tgz http://www.tcpdump.org/daily/tcpdump-2003.09.29.tar.gz
[touch] Should IMHO be called after the file is
Robert Poole [EMAIL PROTECTED] writes:
gcc -O2 -Wall -Wno-implicit -o wget cmpt.o connect.o convert.o
cookies.o ftp.o ftp-basic.o ftp-ls.o ftp-opie.o getopt.o hash.o
headers.o host.o html-parse.o html-url.o http.o init.o log.o main.o
gen-md5.o netrc.o progress.o rbuf.o recur.o res.o retr.o
Arkadiusz Miskiewicz [EMAIL PROTECTED] writes:
On Tuesday 07 of October 2003 12:08, Hrvoje Niksic wrote:
Thanks!
btw. looking into test4 I see that autoconf conception is used in
weird way. Normally aclocal.m4 is autogenerated by aclocal command
Normally only if you're also using Automake
Tony Lewis [EMAIL PROTECTED] writes:
Can you give us linguistically challenged Americans a phonetic
rendition of your name?
It's not easy to describe because of the phonems and concepts not
present in the English language. You'll probably regret having asked.
:-)
HUR-voh-yeh would be the
This beta includes portability tweaks and minor improvements. Please
test it on as many diverse platforms as possible, preferrably with
both gcc and non-gcc compilers. If all goes well, I'd like to release
1.9 perhaps as early as tomorrow.
Get it from:
I don't use an Amiga, nor do I have an idea what you mean by a
working Wget setup. Have you tried compiling from source?
Mauro Tortonesi [EMAIL PROTECTED] writes:
and i'm saying that for this task the ideal structure is
sockaddr_storage. notice that my code uses sockaddr_storage
(typedef'd as wget_sockaddr) only when dealing with socket
addresses, not for ip address caching.
Now I see. Thanks for clearing it
William J Poser [EMAIL PROTECTED] writes:
Using wget 1.8.1 on a GNU/Linux system I have made several
attempts to avoid following links such as www.clickXchange.com
using the --exclude-domains flag. This has no apparent effect:
wget tries to follow links to the excluded domains anyhow.
In
they can; Wget
will simply not harvest the links from such a page.
2003-10-10 Hrvoje Niksic [EMAIL PROTECTED]
* recur.c (retrieve_tree): Don't descend into documents that are
not expected to contain HTML, regardless of their content-type.
* html-url.c
Forrest Garnett [EMAIL PROTECTED] writes:
We see the enclosed errors attempting to compile the most recent
wget package with aix 4.3.3 and aix 5.1.
Thanks for the report. There are several problems with the
compilation.
For one, all the logprintf() lines are failing. This could come from
It seems that you're behind a firewall and need to use passive ftp.
Try the `--passive-ftp' flag, or specify `passive_ftp = on' in
~/.wgetrc.
Thanks for the patch, Herold. I've applied and also added similar
fixes for Borland's and Watcom's Makefiles. I've used the following
ChangeLog entry:
2003-10-09 Herold Heiko [EMAIL PROTECTED]
* windows/Makefile.watcom (OBJS): Ditto.
* windows/Makefile.src.bor: Ditto.
It's a bug. -O currently doesn't work everywhere in should. If you
just want to change the directory where Wget operates, the workaround
is to use `-P'. E.g.:
wget -N ftp://ftp.pld-linux.org/dists/ac/PLD/athlon/PLD/RPMS/packages.dir.mdd -P
[EMAIL PROTECTED] (Martin v. Löwis) writes:
Why do you think the scheme is narrow-minded?
Because 1.9-beta3 seems to be a problem.
VERSION = ('[.0-9]+-?b[0-9]+'
'|[.0-9]+-?dev[0-9]+'
'|[.0-9]+-?pre[0-9]+'
'|[.0-9]+-?rel[0-9]+'
'|[.0-9]+[a-z]?'
Mauro Tortonesi [EMAIL PROTECTED] writes:
so, i am asking you: what do you think of these changes?
Overall they look very good! Judging from the patch, a large piece of
the work part seems to be in an unexpected place: the FTP code.
Here are some remarks I got looking at the patch.
It
Mauro Tortonesi [EMAIL PROTECTED] writes:
I still don't understand the choice to use sockaddr and
sockaddr_storage in a application code.
They result in needless casts and (to me) uncomprehensible code.
well, using sockaddr_storage is the right way (TM) to write IPv6 enabled
code ;-)
Not
[EMAIL PROTECTED] (Martin v. Löwis) writes:
Hrvoje Niksic [EMAIL PROTECTED] writes:
VERSION = ('[.0-9]+-?b[0-9]+'
'|[.0-9]+-?dev[0-9]+'
'|[.0-9]+-?pre[0-9]+'
'|[.0-9]+-?rel[0-9]+'
'|[.0-9]+[a-z]?'
'|[0-9][0-9][0-9][0-9]-[0-9][0
Dan Jacobson [EMAIL PROTECTED] writes:
-q and -S are incompatible and should perhaps produce errors and be
noted thus in the docs.
They seem to work as I'd expect -- `-q' tells Wget to print *nothing*,
and that's what happens. The output Wget would have generated does
contain HTTP headers,
Karl Eichwalder [EMAIL PROTECTED] writes:
Hrvoje Niksic [EMAIL PROTECTED] writes:
As for the Polish translation, translations are normally handled
through the Translation Project. The TP robot is currently down, but
I assume it will be back up soon, and then we'll submit the POT file
Karl Eichwalder [EMAIL PROTECTED] writes:
Also, my Croatian translation of 1.9 doesn't seem to have made it
in. Is that expected?
Unfortunately, yes. Will you please resubmit it with the subject line
updated (IIRC, it's now):
TP-Robot wget-1.9-b3.hr.po
I'm not sure what b3 is, but
Karl Eichwalder [EMAIL PROTECTED] writes:
Hrvoje Niksic [EMAIL PROTECTED] writes:
I'm not sure what b3 is, but the version in the POT file was
supposed to be beta3. Was there a misunderstanding somewhere along
the line?
Yes, the robot does not like beta3 as part of the version
string. b3
Karl Eichwalder [EMAIL PROTECTED] writes:
Hrvoje Niksic [EMAIL PROTECTED] writes:
Ouch. Why does the robot care about version names at all?
It must know about the sequences; this is important for merging
issues. IIRC, we have at least these sequences supported by the
robot:
1.2
Tony Lewis [EMAIL PROTECTED] writes:
Hrvoje Niksic wrote:
Please be aware that Wget needs to know the size of the POST
data in advance. Therefore the argument to @code{--post-file}
must be a regular file; specifying a FIFO or something like
@file{/dev/stdin} won't work
Karl Eichwalder [EMAIL PROTECTED] writes:
I guess, you as the wget maintainer switched from something
supported to the unsupported betaX scheme and now we have
something to talk about ;)
I had no idea that something as usual as betaX was unsupported. In
fact, I believe that bX was added when
Stefan Eissing [EMAIL PROTECTED] writes:
Am Dienstag, 07.10.03, um 16:36 Uhr (Europe/Berlin) schrieb Hrvoje
Niksic:
What the current code does is: determine the file size, send
Content-Length, read the file in chunks (up to the promised size) and
send those chunks to the server
Martin, thanks for the patch and the detailed report. Note that it
might have made more sense to apply the patch to the latest CVS
version, which is somewhat different from 1.8.2.
I'm really not sure whether to add this patch. On the one hand, it's
nice to support as many architectures as
Tony Lewis [EMAIL PROTECTED] writes:
Hrvoje Niksic wrote:
I don't understand what you're proposing. Reading the whole file in
memory is too memory-intensive for large files (one could presumably
POST really huge files, CD images or whatever).
I was proposing that you read the file
Josh Brooks [EMAIL PROTECTED] writes:
I have noticed very unpredictable behavior from wget 1.8.2 -
specifically I have noticed two things:
a) sometimes it does not follow all of the links it should
b) sometimes wget will follow links to other sites and URLs - when the
command line used
Suhas Tembe [EMAIL PROTECTED] writes:
Thanks everyone for the replies so far..
The problem I am having is that the customer is using ASP Java
script. The URL stays the same as I click through the links.
URL staying the same is usually a sign of the use of frame, not of ASP
and
Suhas Tembe [EMAIL PROTECTED] writes:
this page contains a drop-down list of our customer's locations.
At present, I choose one location from the drop-down list click
submit to get the data, which is displayed in a report format. I
right-click then choose view source save source to a file.
Suhas Tembe [EMAIL PROTECTED] writes:
It does look a little complicated This is how it looks:
form action=InventoryStatus.asp method=post [...]
[...]
select name=cboSupplier
option value=4541-134289454A/option
option value=4542-134289 selected454B/option
/select
Those are the
To subscribe to this list, please send mail to
[EMAIL PROTECTED].
There is currently no way to disable following redirects. A patch to
do so has been submitted recently, but I didn't see a good reason why
one would need it, so I didn't add the option. Your mail is a good
argument, but I don't know how prevalent that behavior is.
What is it with servers that
Tony Lewis [EMAIL PROTECTED] writes:
wget
http://www.custsite.com/some/page.html --http-user=USER --http-passwd=PASS
If you supply your user ID and password via a web form, it will be
tricky (if not impossible) because wget doesn't POST forms (unless
someone added that option while I wasn't
Suhas Tembe [EMAIL PROTECTED] writes:
Hello Everyone,
I am new to this wget utility, so pardon my ignorance.. Here is a
brief explanation of what I am currently doing:
1). I go to our customer's website every day log in using a User Name Password.
2). I click on 3 links before I get to
Several bugs fixed since beta3, including a fatal one on Windows.
Includes a working Windows implementation of run_with_timeout.
Get it from:
http://fly.srk.fer.hr/~hniksic/wget/wget-1.9-beta4.tar.gz
Gisle Vanem [EMAIL PROTECTED] writes:
--- mswindows.c.org Mon Sep 29 11:46:06 2003
+++ mswindows.c Sun Oct 05 17:34:48 2003
@@ -306,7 +306,7 @@
DWORD set_sleep_mode (DWORD mode)
{
HMODULE mod = LoadLibrary (kernel32.dll);
- DWORD (*_SetThreadExecutionState) (DWORD) = NULL;
+
Gisle Vanem [EMAIL PROTECTED] writes:
Hrvoje Niksic [EMAIL PROTECTED] said:
I've committed this patch, with minor changes, such as moving the code
to mswindows.c. Since I don't have MSVC, someone else will need to
check that the code compiles. Please let me know how it goes.
It compiled
It's a feature. `-A zip' means `-A zip', not `-A zip,html'. Wget
downloads the HTML files only because it absolutely has to, in order
to recurse through them. After it finds the links in them, it deletes
them.
Thanks for the contribution. Note that a slightly more correct place
to send the patch is the [EMAIL PROTECTED] list, followed by
people with a keener interest in development.
Also, you should send at least a short explanation of what each patch
is supposed to do and why one should apply it.
Thanks for the patch, I've now applied it with the following ChangeLog
entry:
2003-10-03 Gisle Vanem [EMAIL PROTECTED]
* connect.c: And don't include them here.
* mswindows.h: Include winsock headers here.
However, I've postponed applying the part that changes `-d'. I agree
Jochen Roderburg [EMAIL PROTECTED] writes:
Zitat von Hrvoje Niksic [EMAIL PROTECTED]:
It's a feature. `-A zip' means `-A zip', not `-A zip,html'. Wget
downloads the HTML files only because it absolutely has to, in order
to recurse through them. After it finds the links in them, it deletes
Payal Rathod [EMAIL PROTECTED] writes:
On Wed, Oct 01, 2003 at 09:26:47PM +0200, Hrvoje Niksic wrote:
The way to do it with Wget would be something like:
wget --mirror --no-host-directories ftp://username:[EMAIL PROTECTED]
But if I run in thru' crontab, where will it store
This problem is not specific to timeouts, but to recursive download (-r).
When downloading recursively, Wget expects some of the specified
downloads to fail and does not propagate that failure to the code that
sets the exit status. This unfortunately includes the first download,
which should
The home page is back, but it says that the TP Robot is dead. I've
contacted Martin Loewis, perhaps he'll be able to provide more info.
Payal Rathod [EMAIL PROTECTED] writes:
On Thu, Oct 02, 2003 at 12:03:34PM +0200, Hrvoje Niksic wrote:
Payal Rathod [EMAIL PROTECTED] writes:
On Wed, Oct 01, 2003 at 09:26:47PM +0200, Hrvoje Niksic wrote:
The way to do it with Wget would be something like:
wget --mirror --no-host
Gisle Vanem [EMAIL PROTECTED] writes:
I've patched util.c to make run_with_timeout() work on Windows
(better than it does with alarm()!).
Cool, thanks! Note that, to save the honor of Unix, I've added
support for setitimer on systems that support it (virtually everything
these days), so
I've committed this patch, with minor changes, such as moving the code
to mswindows.c. Since I don't have MSVC, someone else will need to
check that the code compiles. Please let me know how it goes.
[ Added Cc to [EMAIL PROTECTED] ]
Tony Lewis [EMAIL PROTECTED] writes:
The following patch adds a command line option to save any links
that are not followed by wget. For example:
wget http://www.mysite.com --mirror --unfollowed-links=mysite.links
will result in mysite.links containing all
Does anyone know the current procedure for submitting the `.pot' file
to the GNU Translation Project? At the moment, the project home page
at http://www.iro.umontreal.ca/contrib/po/HTML/ appears dead.
Tony Lewis [EMAIL PROTECTED] writes:
Hrvoje Niksic wrote:
I'm curious: what is the use case for this? Why would you want to
save the unfollowed links to an external file?
I use this to determine what other websites a given website refers to.
For example:
wget
http
Tony Lewis [EMAIL PROTECTED] writes:
Would something like the following be what you had in mind?
301 http://www.mysite.com/
200 http://www.mysite.com/index.html
200 http://www.mysite.com/followed.html
401 http://www.mysite.com/needpw.html
--- http://www.othersite.com/notfollowed.html
Yes,
Payal Rathod [EMAIL PROTECTED] writes:
I have 5-7 user accounts in /home whose data is important. Every day at
12:00 I want to back their data to a differnt backup machine.
The remote machine has a ftp server.
Can I use wget for this? If yes, how do I proceed?
The way to do it with Wget
801 - 900 of 1457 matches
Mail list logo