On 2002-10-17 12:16 -0600, Daniel Webb wrote:
> Also, concerning the mailing list, I am not interested in using a kludgy
> web-based interface to an email archive. Where are the mbox download
> links?
Amen.
--
André Majorel <[EMAIL PROTECTED]>
http://www.teaser.fr/~amajorel/
I've just had a recursive wget do something unexpected : it
spanned hosts even though I didn't give the -H option. The command
was :
wget -r -l20 http://www.modcan.com/page2.html
http://www.modcan.com/pg2_main.html contains a link to
www.paypal.com, and that link was followed.
That was Wget 1
On 2002-09-20 08:15 +0200, Dominic Chambers wrote:
> I am using wget 1.82 on Win2K SP2, and wget froze on the fifth
1.8.2.
> downloaded file 'prn.html' using the command line:
>
> wget -r -l0 -A htm,html,png,gif,jpg,jpeg --no-parent
> http://java.sun.com/products/jlf/at/book
>
> About twenty
On 2002-07-26 01:59 +0200, Hrvoje Niksic wrote:
> Only the bare minimum of characters should be encoded. The ones that
> come to mind are '/' (illegal), '~' (rm -r ~foo dangerous), '*' and
> '?' (used in wildcards), control characters 0-31 (controls), and chars
> 128-159 (non-printable).
While
I'm trying to snarf a web site that is served over FTP. wget -r
doesn't work probably because Wget doesn't parse HTML documents
retrieved with FTP (which is reasonable).
Is there a sort of --follow-html option to force Wget to parse
HTML documents served over FTP and follow the links, as if they
On 2002-06-29 21:09 -0400, Dang P. Tran wrote:
> I use the -i option to download files from an url list. The
> server I use have a password that change often. When I have a
> large list if the password change while I'm downloading and give
> 401 error, I want wget stop to prevent hammering the si
On 2002-05-14 13:01 -0400, Kevin Murphy wrote:
> However, I am trying to suck a particular site which relies excessively
> on javascript'ed links, e.g. via window.open, sometimes wrapped in
> function calls.
>
> I realize that in general this an intractable problem, but is anybody
> aware of
On 2002-04-19 16:46 +0100, Ian Abbott wrote:
> > It means that `www.arsdigita.de' cannot set
> > the cookie for `arsdigita.de'. To make *that* work, you'd have to
> > maintain a database of domains that use ".co.xxx" convention, as
> > opposed to those that use just ".xxx".
>
> Could you assume
On 2002-04-19 11:21 +0200, Hrvoje Niksic wrote:
> There are now fewer spams than there were (I know because I get the
> ones that get caught in the net), but we're not quite there yet. We
> will be, though.
In case this is of any use to you, these procmail recipes block
at least 3/4 of the asia
On 2002-04-14 05:00 +0200, Hrvoje Niksic wrote:
>The moderators are informed about each message that awaits
>moderation; that alert would contain a URL they can visit and
>approve or reject the mail, at their discretion.
The web interface is not necessary. Listar, for instance, just
On 2002-04-10 01:14 +0200, Hrvoje Niksic wrote:
> Andre Majorel <[EMAIL PROTECTED]> writes:
>
> > If find it very annoying when a downloader plays yoyo with the
> > remaining time. IMHO, remaining time is by nature a long term thing
> > and short term jitter should
On 2002-04-09 20:51 +0200, Hrvoje Niksic wrote:
> The one remaining problem is the ETA. Based on the current speed, it
> changes value wildly. Of course, over time it is generally
> decreasing, but one can hardly follow it. I removed the flushing by
> making sure that it's not shown more than
On 2002-04-03 08:02 -0800, Tony Lewis wrote:
> > > Yes, that allows me to specify _A_ referrer, like www.aol.com.
> > > When I'm trying to help my users mirror their old angelfire pages
> > > or something like that, very often the link has to come from the
> > > same directory. I'd like to see so
On 2002-04-03 08:50 -0500, Dan Mahoney, System Admin wrote:
> > > 1) referrer faking (i.e., wget automatically supplies a referrer
> > > based on the, well, referring page)
> >
> > It is the --referer option, see (wget)HTTP Options, from the Info
> > documentation.
>
> Yes, that allows me to spe
On 2002-03-22 04:08 +0100, Hrvoje Niksic wrote:
> > May I suggest that you set a filter that prevents postings to the
> > list unless the poster is a subscriber. That filter should forward
> > the mail to the admins to allow them the pass the mail through if
> > suitable.
>
> Do you volunteer to
On 2002-03-19 00:06 +0100, Fabrice Bauzac wrote:
> SMIL files resemble to XHTML files. Here is a sample:
>
> Content-type: application/smil
>
>
, heh ? Looks like XML to me. I could be wrong but I don't
think Wget groks this for
On 2002-03-14 23:33 +0800, steven wrote:
> Thank you, but ... You may mistaked something. I want to
> download those wav files anywhere. For those html files,
> however, I only want to grab ones under the starting directory
> 'start/'. That means what I need is anything under the
> directory of
On 2002-03-05 11:41 +0100, Philipp Thomas wrote:
> When requesting a URL like http://tmp.logix.cz/slash.xp , wget shortens
> this to http://tmp.logix.cz/slash.xp/. All Browsers I tested (Opera 6b1,
> Mozilla 0.9.8, Konqueror 2.9.2) pass this URL as given.
>
> So the question is, why wget (1.
On 2002-02-08 16:09 -, Ian Abbott wrote:
> This does not really apply to computer usage where typically "kilo"
> has been overloaded to mean 2^10 (=1024) because it happens to be
> close enough to its more correct meaning. That's why "K" is often
> used to mean 2^10 to avoid confusion with "k
On 2002-02-08 08:54 +0100, Hrvoje Niksic wrote:
> Wget currently uses "KB" as abbreviation for "kilobyte". In a Debian
> bug report someone suggested that "kB" should be used because it is
> "more correct". The reporter however failed to cite the reference for
> this, and a search of the web ha
On 2002-01-29 22:02 +0100, Hrvoje Niksic wrote:
> But that was just an example. The actual reasoning for allowing
> non-subscriber posting boils down to three reasons:
>
> 1. I believe it is the right thing to do. I personally hate allegedly
>"supportive" mailing lists that require me to s
On 2002-01-29 09:56 -0500, Alan Eldridge wrote:
> In particular, does wget parse and follow links in an HTML document,
> when that document is retrieved (using -r) via the FTP protocol? If
> not, why not?
>
> FTP presents an interesting case because because there are two
> independent concepts o
On 2002-01-28 14:33 -0500, Thomas Reinke wrote:
> Is anyone else not finding the noise ratio (i.e. spam)
> a bit high here?
A bit *low* you mean ? You bet.
> I sympathize with the effort required
> to lightly moderate, but might I recommend that
> _something_ be done to rid us all of this spam?
On 2002-01-25 14:01 +0100, Jens Röder wrote:
> for wget I would suggest a switch that allows to send the output directly
> to stdout. It would be easier to use it in pipes.
Does
wget ... 2>&1 | command
solve your problem ?
--
André Majorel http://www.teaser.fr/~amajorel/>
std::disclaimer (
On 2002-01-21 18:53 +0100, Hrvoje Niksic wrote:
> "Ian Abbott" <[EMAIL PROTECTED]> writes:
>
> > Why not just open the wgetrc file in text mode using fopen(name,
> > "r") instead of "rb"? Does that introduce other problems?
>
> Not that I'm aware of. The reason we use "rb" now is the fact that
On 2002-01-15 11:59 -, Ian Abbott wrote:
> This is an initial proposal for naming the files and directories
> that Wget creates, based on the URLs of the retrieved documents.
[massive snippage]
Ian, can your proposal be summarised as :
- in the internal representation of filenames, don't es
On 2001-12-27 01:53 +0100, Svante Signell wrote:
> Thank you for your prompt reply. The -R option is usable for a _few_
> files. I was more thinking of a file:
>
> cat excludefile.txt
> XFree86-Servers-3.3.6-102.src.rpm
> XFree86-Servers-3.3.6-42.src.rpm
> compat-egcs-6.2-1.1.2.16.src.rpm
> comp
On 2001-12-16 19:02 +0100, Hrvoje Niksic wrote:
> Andre Majorel <[EMAIL PROTECTED]> writes:
>
> > On 2001-12-15 07:37 +0100, Hrvoje Niksic wrote:
> >
> >> Is there a good fallback value of RAND_MAX for systems that don't
> >> bother
On 2001-12-15 07:37 +0100, Hrvoje Niksic wrote:
> Is there a good fallback value of RAND_MAX for systems that don't
> bother to define it?
The standard (SUS2) says :
The value of the {RAND_MAX} macro will be at least 32767.
--
André Majorel
Work: <[EMAIL PROTECTED]>
Home: <[EMAIL PROTECTED]
On 2001-12-07 15:10 +0100, Hrvoje Niksic wrote:
> But: a character being unsafe for URL doesn't mean that the same
> character must be unsafe for the file name. Wget currently contains
> the two, and that's a bug. I'll try to fix that bug by adding another
> bitflag to the table, e.g. "F" which
On 2001-12-03 21:55 +0100, Hrvoje Niksic wrote:
> Bugfixes since 1.8-beta2. Please test it from clean compilation on
> Unix (Windows and MacOS are known not to compile without modifications
> when SSL is used.)
>
> Get it from:
>
> ftp://gnjilux.srk.fer.hr/pub/unix/util/wget/.betas/wget-1.8
On 2001-12-03 19:16 +0100, Hrvoje Niksic wrote:
> > I find describing HP-UX 10 as a modern OS mildly amusing. :-)
>
> How old is it? I used to work on HPUX 9, and I'm not old by most
> definitions of the word.
Around 1995.
> > I completely disagree with your perception that snprintf() is to b
On 2001-12-01 23:30 +0100, Hrvoje Niksic wrote:
> Here is the next 1.8 beta. Please test it if you can -- try compiling
> it on your granma's Ultrix box, run it on your niece's flashy web
> site, see if cookies work, etc.
>
> Get it from:
>
> ftp://gnjilux.srk.fer.hr/pub/unix/util/wget/.bet
On 2001-12-03 18:30 +0100, Hrvoje Niksic wrote:
> Andre Majorel <[EMAIL PROTECTED]> writes:
>
> > gcc -I. -I.-DHAVE_CONFIG_H -DSYSTEM_WGETRC=\"/usr/local/etc/wgetrc\"
>-DLOCALEDIR=\"/usr/local/share/locale\" -O2 -Wall -Wno-implicit -c
On 2001-12-01 23:30 +0100, Hrvoje Niksic wrote:
> Here is the next 1.8 beta. Please test it if you can -- try compiling
> it on your granma's Ultrix box, run it on your niece's flashy web
> site, see if cookies work, etc.
>
> Get it from:
>
> ftp://gnjilux.srk.fer.hr/pub/unix/util/wget/.bet
On 2001-09-13 11:16 +0200, Jan Hnila wrote:
> Hello!
>
> To be able to use the referer switch, you must have a new version of
> wget - I'm not sure, if 1.6 is enough, 1.7 certainly is enough and 1.5.3
> is not enough.(Get more info's from http://wget.sunsite.dk)
>
> The switch is --referer=URL
On 2001-06-15 16:30 -0500, Bazuka wrote:
> So what would be the advantage of multithreading this application ?
Multithreading might be an advantage when retrieving files from
several hosts because gethostbyname() is blocking (and often
takes a while to complete).
--
André Majorel <[EMAIL PROTE
Executive summary: complete success.
On NCR MP-RAS, Wget 1.7.1-pre1 configured and compiled fine, and
passed a few simple tests. The -lnsl/-lsocket and MAP_FAILED
problems seen with previous versions did not occur.
No SSL library is installed on the system. ./configure
--with-ssl detected that c
On 2001-06-08 17:57 -0400, Parsons, Donald wrote:
> Previous versions up to 1.6 compiled fine.
>
> cd src && make CC='gcc' CPPFLAGS='' DEFS='-DHAVE_CONFIG_H
>-DSYSTEM_WGETRC=\"/usr/etc/wgetrc\" -DLOCA
> LEDIR=\"/usr/share/locale\"' CFLAGS='-O2 -fomit-frame-pointer -march=pentium
>-mcpu=pentium
On 2001-06-06 07:29 +0200, Jan Prikryl wrote:
> It seems that -lsocket is not found as it requires -lnsl for
> linking. -lnsl is not detected as it does not contain
> `gethostbyname()' function.
>
> Would the attacheed patch to configure.in solve the problem? Please
> not that the patch tries to
On 2001-06-06 12:47 +0200, Jan Prikryl wrote:
> > Jan Prikryl <[EMAIL PROTECTED]> writes:
> >
> > > It seems that -lsocket is not found as it requires -lnsl for
> > > linking. -lnsl is not detected as it does not contain
> > > `gethostbyname()' function.
> >
> > That's weird. What does libnsl c
On 2001-06-02 20:50 +0200, Andre Majorel wrote:
> On 2001-06-02 17:30 +0200, Hrvoje Niksic wrote:
> > > - The empty LIBS problem remains (add -lsocket -lnsl).
> >
> > Do you have a config.log for this? Wget's configure tries hard to
> > determine whether `-l
On 2001-06-02 17:30 +0200, Hrvoje Niksic wrote:
> Andre Majorel <[EMAIL PROTECTED]> writes:
>
> > - The CC "-cX" problem in wget 1.6 is fixed. Good. :-)
>
> You mean "-Xc"?
Sorry, my tyop.
> I decided that "strict ANSI" mode is the wro
On 2001-05-26 13:35 +0200, Hrvoje Niksic wrote:
> The pre-release is available at:
>
> ftp://gnjilux.srk.fer.hr/pub/unix/util/wget/.betas/wget-1.7-pre1.tar.gz
Tried it on NCR MP-RAS 3.0 :
- The CC "-cX" problem in wget 1.6 is fixed. Good. :-)
- The empty LIBS problem remains (add -lsocket
On 2001-05-26 11:10 +0200, Hrvoje Niksic wrote:
> Andre Majorel <[EMAIL PROTECTED]> writes:
>
> > Compiling Wget 1.6 on an SVR4 derivative (NCR MP-RAS 3.0), I got
> > this strange error:
>
> I think the problem is that Wget 1.6 tried to force "strict ANSI mo
Compiling Wget 1.6 on an SVR4 derivative (NCR MP-RAS 3.0), I got
this strange error:
# make
CONFIG_FILES= CONFIG_HEADERS=src/config.h ./config.status
creating src/config.h
src/config.h is unchanged
generating po/POTFILES from ./po/POTFILES.in
creating po/Makefile
cd
On 2001-05-06 23:15 +0200, Jonas Jensen wrote:
> Did you test it? On my Windows machine I have a file named [];,= on my
> desktop right now, it causes no problems at all. When you rename a file to
> an illegal name in Explorer, you get the error message:
>
> A filename cannot contain any of the
On 2001-03-20 00:25 +0100, Hrvoje Niksic wrote:
> "Eddy Thilleman" <[EMAIL PROTECTED]> writes:
>
> > Wget sends its output to standard error. Why is that?
>
> "It seemed like a good idea."
>
> The rationale behind it is that Wget's "output" is not real output,
> more a progress indication thing
I'm mirroring a very large tree locally. As the tree is larger
than the local filesystem, I periodically stop wget, save what
I've downloaded on CD-ROM, truncate the saved files to 0 and
then start wget -N -r again to get more files.
Unfortunately, wget checks not only the mtime but also the size
On 2001-02-18 01:08 -0500, Mordechai T. Abzug wrote:
>
> Sometimes, I run wget in background to download a file that will take
> hours or days to complete. It would be handy to have an option for
> wget to send me mail when it's done, so I can fire and forget.
Perhaps this will do the trick :
50 matches
Mail list logo