Re: How do I update working dir to wget-1.5.3?

2001-01-17 Thread Thor A. Lange

 "Dan" == Dan Harkless [EMAIL PROTECTED] writes:

Dan Cool.  Hey, speaking of webby CVS stuff, do you guys by any
Dan chance have cvsweb installed on sunsite.dk?  I find it very
Dan handy for browsing CVS source.  For example:

We have now made an installation of cvsweb. It is accessible from:
http://SunSITE.dk/cvsweb

/Thor

-- 
Thor Langemailto:[EMAIL PROTECTED]
Department of Computer Sciencehttp://SunSITE.dk/~lange/
Aalborg University



Re: How do I update working dir to wget-1.5.3?

2001-01-17 Thread Hrvoje Niksic

[EMAIL PROTECTED] (Thor A. Lange) writes:

  "Dan" == Dan Harkless [EMAIL PROTECTED] writes:
 
 Dan Cool.  Hey, speaking of webby CVS stuff, do you guys by any
 Dan chance have cvsweb installed on sunsite.dk?  I find it very
 Dan handy for browsing CVS source.  For example:
 
 We have now made an installation of cvsweb. It is accessible from:
 http://SunSITE.dk/cvsweb

Unspeakably cool!  Thanks for this.

I just noticed one bug: the `diff' cvsweb executes does not appear to
be GNU diff, so most of the diff options (colored diff, unidiff,
side-by-side) fail to work.  Please make sure to have GNU diff
somewhere on the machine, and that cvsweb's PATH points to that one.



Feature Idea

2001-01-17 Thread Sam Tregar

Hey guys - I've got an idea for a new feature I'd like to add to wget.
I'd like a way to specify a program to be run that can filter the URLs
just before they are fetched.  I'd like this so that I could use wget to
do recursive retrievals against Google's web cache.  This would be useful
for restoring deleted web sites, reading sites under heavy load, etc.
Something like this was my first shot:

   wget -r "http://www.google.com/search?q=cache:www.tregar.com/"

That works fine for the first page but the page that comes back contains
links that refer to www.tregar.com, not Google's cache.  My solution,
given the proposed feature, would be something like:

   wget -r --url-filter=google.pl \
  http://www.google.com/search?q=cache:www.tregar.com/"

Where google.pl would be something like (assuming the url comes in
through STDIN and goes out through STDOUT and minus error checking):

   #!/usr/bin/perl
   while(STDIN) {
s!^http://!!;
print "http://www.google.com/search?q=cache:$_\n";

Another possible implementation would be to include a regex engine in wget
and allow the user to specify the filter as a regex.  This obviously makes
for less powerful filters but might be more UNIXy.

Reactions?

-sam