[±¤ °í]ÀÌ»ç°èȹÀÌ ÀÖÀ¸½Å °í°´À» À§ÇÑ ¸ÂÃã¼­ºñ½º~!

2002-04-03 Thread good2002
Title: Untitled Document





  

  

  

  

 
  

  

  


  

  

  
 
  

 
  

 
   
  
   

 
   
  
   

 
   
   

 
   
  
   

  

   
±ÍÇÏÀÇ ½Â¶ô ¾øÀÌ È«º¸¼º ÀüÀÚ ¿ìÆíÀ» º¸³»°Ô µÈ Á¡ Á¤ÁßÈ÷ »ç°ú 
µå¸³´Ï´Ù.
   ¢º ÀúÈñ ȸ»ç´Â Á¤º¸Åë½Å¸ÁÀÌ¿ëÃËÁø¹ý ±ÔÁ¤À» ÁؼöÇÏ¿© ±¤°í ¸ÞÀÏÀÓÀ» Ç¥½ÃÇÏ¿´À¸¸ç, 

¼ö½Å °ÅºÎ ÀåÄ¡¸¦ ÇÊÈ÷ ¸¶·ÃÇÏ°í ÀÖ½À´Ï´Ù.
   ¢º ±ÍÇÏÀÇ ÀüÀÚ ¿ìÆí ÁÖ¼Ò´Â ÀÎÅÍ³Ý »óÀÇ °ø°³µÈ Àå¼Ò¿¡¼­ ½ÀµæÇÏ¿´À¸¸ç, 

ÀúÈñ ȸ»ç´Â ±ÍÇÏÀÇ ÀüÀÚ¿ìÆí ÁÖ¼Ò ¿Ü ¾î¶°ÇÑ °³ÀÎ Á¤º¸µµ °¡Áö°í ÀÖÁö ¾ÊÀ¸¹Ç·Î ¾È½ÉÇϽñ⠹ٶø´Ï´Ù.
  ¾ÕÀ¸·Î ÀúÈñ°¡ ¹ß¼ÛÇÏ´Â ¸ÞÀÏÀ» ¼ö½Å°ÅºÎÇÏ·Á¸é 
¾Æ·¡¿¡ ¼ö½Å°ÅºÎ ¹öÆ°À» ´­·¯ÁÖ¼¼¿ä.
  

  


  

  



  

  

  

  




O »çÀü Çã¶ô¾øÀÌ ¸ÞÀÏÀ» º¸³»°Ô µÈÁ¡ »ç°úµå¸³´Ï´Ù.O ¸ÞÀÏ ¼ö½ÅÀ» ¿øÇϽÃÁö ¾ÊÀ¸½Ã¸é  ´ÙÀ½À» Ŭ¸¯ÇØ ÁÖ¼¼¿ä

wget 1.8: FTP recursion through proxy servers does not work anymore

2002-04-03 Thread Vladimir Volovich

Hi!

additional note: wget 1.7.1 works fine WRT FTP recursion through proxy
servers; that was the last version before the change from
recursive_retrieve to retrieve_tree was made (according to ChangeLog).

So, indeed the change from recursive_retrieve to retrieve_tree
introduced a bug.

Best,
v.

== original message ==
in wget 1.8, FTP recursion through proxy servers does not work anymore:
when i run

wget --execute "ftp_proxy = http://some.proxy:3128"; -m ftp://some.host/dir/

wget only retrieves the file ftp://some.host/dir/index.html
and stops.

i found the following ChangeLog entries:

* recur.c (recursive_retrieve): Enable FTP recursion through proxy
servers.

(at that point recussion started working), and

* recur.c (url_queue_new): New function.
(url_queue_delete): Ditto.
(url_enqueue): Ditto.
(url_dequeue): Ditto.
(retrieve_tree): New function, replacement for recursive_retrieve.
(descend_url_p): New function.
(register_redirection): New function.

at this point (when retrieve_tree replaced recursive_retrieve) it
probably was broken.

Best,
v.




option -nh (was: Re: option changed: -nh -> -nH)

2002-04-03 Thread Noel Koethe

On Mit, 03 Apr 2002, Jens Rösner wrote:

> I already complained that many old scripts now break and suggested 
> that entering -nh at the command line would 
> either be completely ignored or the user would be 
> informed and wget executed nevertheless.
> Apparently this was not regarded as useful.

:( OK. So this bug wont be fixed.:(
thx.

-- 
Noèl Köthe



Re: option changed: -nh -> -nH

2002-04-03 Thread Noel Koethe

On Mit, 03 Apr 2002, Jens Rösner wrote:

Hello Jens,

> -nh
> and
> -nH
> are totally different.

Thanks for your explanation!

-- 
Noèl Köthe



--non-verbose disables connection error messages

2002-04-03 Thread Kevin Rodgers

The wget 1.8.1 man pages says

  --non-verbose
  Non-verbose output---turn off verbose without being completely quiet
  (use -q for that), which means that error messages and basic
  information still get printed.

but it also prevents messages such as the following from being logged:

Connecting to localhost[127.0.0.1]:80... failed: Connection refused.

I think the calls to logprintf() in src/connect.c:connect_to_one()
should be changed from (LOG_VERBOSE, ...) to (LOG_NONVERBOSE, ...).

-- 
Kevin




wget 1.8: FTP recursion through proxy servers does not work anymore

2002-04-03 Thread Vladimir Volovich

Hi!

in wget 1.8, FTP recursion through proxy servers does not work anymore:
when i run

wget --execute "ftp_proxy = http://some.proxy:3128"; -m ftp://some.host/dir/

wget only retrieves the file ftp://some.host/dir/index.html
and stops.

i found the following ChangeLog entries:

* recur.c (recursive_retrieve): Enable FTP recursion through proxy
servers.

(at that point recussion started working), and

* recur.c (url_queue_new): New function.
(url_queue_delete): Ditto.
(url_enqueue): Ditto.
(url_dequeue): Ditto.
(retrieve_tree): New function, replacement for recursive_retrieve.
(descend_url_p): New function.
(register_redirection): New function.

at this point (when retrieve_tree replaced recursive_retrieve) it
probably was broken.

Best,
v.




Re: Referrer Faking and other nifty features

2002-04-03 Thread Ian Abbott

On 3 Apr 2002 at 16:10, Andre Majorel wrote:

> On 2002-04-03 08:50 -0500, Dan Mahoney, System Admin wrote:
> 
> > > > 1) referrer faking (i.e., wget automatically supplies a referrer
> > > > based on the, well, referring page)
> > >
> > > It is the --referer option, see (wget)HTTP Options, from the Info
> > > documentation.
> > 
> > Yes, that allows me to specify _A_ referrer, like www.aol.com.  When I'm
> > trying to help my users mirror their old angelfire pages or something like
> > that, very often the link has to come from the same directory.  I'd like
> > to see something where when wget follows a link to another page, or
> > another image, it automatically supplies the URL of the page it followed
> > to get there.  Is there a way to do this?
> 
> Somebody already asked for this and AFAICT, there's no way to do
> that.

Wget is supposed to pass on the referring page when following
links, but Wget 1.8 had a little bug that stopped this happening.
The bug is fixed in Wget 1.8.1.

ISTR that the somebody that Andre refers to wanted to be able to
specify different referring URLs for each URL specified on the
command-line.



Re: Referrer Faking and other nifty features

2002-04-03 Thread Andre Majorel

On 2002-04-03 08:02 -0800, Tony Lewis wrote:

> > > Yes, that allows me to specify _A_ referrer, like www.aol.com.
> > > When I'm trying to help my users mirror their old angelfire pages
> > > or something like that, very often the link has to come from the
> > > same directory. I'd like to see something where when wget follows
> > > a link to another page, or another image, it automatically
> > > supplies the URL of the page it followed to get there.  Is there a
> > > way to do this?
> >
> > Somebody already asked for this and AFAICT, there's no way to do
> > that
> 
> Not only is it possible, it is the behavior (at least in wget 1.8.1).
> If you run with -d, you will see that every GET after the first one
> includes the appropriate referer.

Right, for some reason I thought it used the provided referer
throughout. Thanks for the correction !

-- 
André Majorel http://www.teaser.fr/~amajorel/>
std::disclaimer ("Not speaking for my employer");



suspected bug in WGET 1.8.1

2002-04-03 Thread Matt Jackson

I'm using the NT port of WGET 1.8.1.

FTP retrieval of files works fine, retrieval of directory listings fails.
The problem happens under certain conditions when connecting to OS2 FTP
servers.

For example, if the "current directory" on the FTP server at login time is
"e:/abc", the command "wget ftp://userid:password@ipaddr/g:\def\test.doc";
works fine to retrieve the file, but the command "wget
ftp://userid:password@ipaddr/g:\def\"; fails to retrieve the directory
listing.

For what it's worth, "g:\def/", "g:/def\" and "g:/def/" also fail.


Matt Jackson
(919) 254-4547
[EMAIL PROTECTED]




Re: Referrer Faking and other nifty features

2002-04-03 Thread Tony Lewis

Andre Majorel wrote:

> > Yes, that allows me to specify _A_ referrer, like www.aol.com.  When I'm
> > trying to help my users mirror their old angelfire pages or something
like
> > that, very often the link has to come from the same directory.  I'd like
> > to see something where when wget follows a link to another page, or
> > another image, it automatically supplies the URL of the page it followed
> > to get there.  Is there a way to do this?
>
> Somebody already asked for this and AFAICT, there's no way to do
> that

Not only is it possible, it is the behavior (at least in wget 1.8.1). If you
run with -d, you will see that every GET after the first one includes the
appropriate referer.

If I execute: wget -d -r http://www.exelana.com --referer=http://www.aol.com

The first request is reported as:
GET / HTTP/1.0
User-Agent: Wget/1.8.1
Host: www.exelana.com
Accept: */*
Connection: Keep-Alive
Referer: http://www.aol.com

But, the third request is:
GET /left.html HTTP/1.0
User-Agent: Wget/1.8.1
Host: www.exelana.com
Accept: */*
Connection: Keep-Alive
Referer: http://www.exelana.com/

The second request is for robots.txt and uses the referer from the command
line.

Tony




Re: cuj.com file retrieving fails -why?

2002-04-03 Thread Ian Abbott

On 3 Apr 2002 at 17:09, Markus Werle wrote:

> Ian Abbott wrote:
> 
> > On 3 Apr 2002 at 14:56, Markus Werle wrote:
> > I've just built Wget 1.7 on Linux and it seemed to download your
> > problem file okay. So I don't know what your problem is either!
> 
> Ah! The kind of problem I like most!
> Did You have a special .wgetrc?

Nothing special.

$HOME/.wgetrc :

robots = off

system wgetrc :

# Comments stripped out
passive_ftp = on
waitretry = 10




Re: Referrer Faking and other nifty features

2002-04-03 Thread Daniel Stenberg

On Wed, 3 Apr 2002, Andre Majorel wrote:

> > No, I mean downloading multiple files from the SAME uri in parallel,
> > instead of downloading files one-by-one-by-one (thus saving time on a
> > fast pipe).
>
> This doesn't make sense to me. When downloading from a single server, the
> bottleneck is generally either the server or the link ; in either case,
> there's nothing to win by attempting several simultaneous transfers. Unless
> there are several servers at the same IP and the bottleneck is the server,
> not the link ?

That would also be violating the RFCs in regard on to how to behave as a Good
User Agent (tm). Never use more than at most 2 connections to the same host
from a single client.

What does spead up things though, is persistant connections. One might also
argue that pipelining can do a little too.

-- 
  Daniel Stenberg - http://daniel.haxx.se - +46-705-44 31 77
   ech`echo xiun|tr nu oc|sed 'sx\([sx]\)\([xoi]\)xo un\2\1 is xg'`ol




Re: cuj.com file retrieving fails -why?

2002-04-03 Thread Markus Werle

Ian Abbott wrote:

> On 3 Apr 2002 at 14:56, Markus Werle wrote:
>
> > Jens Rösner wrote:
> > > So, I do not know what your problem is, but is neither wget#s nor cuj's
> > > fault, AFAICT.
> >
> > :-(
>
> I've just built Wget 1.7 on Linux and it seemed to download your
> problem file okay. So I don't know what your problem is either!

Ah! The kind of problem I like most!
Did You have a special .wgetrc?


regards,


Markus




Re: Referrer Faking and other nifty features

2002-04-03 Thread Andre Majorel

On 2002-04-03 08:50 -0500, Dan Mahoney, System Admin wrote:

> > > 1) referrer faking (i.e., wget automatically supplies a referrer
> > > based on the, well, referring page)
> >
> > It is the --referer option, see (wget)HTTP Options, from the Info
> > documentation.
> 
> Yes, that allows me to specify _A_ referrer, like www.aol.com.  When I'm
> trying to help my users mirror their old angelfire pages or something like
> that, very often the link has to come from the same directory.  I'd like
> to see something where when wget follows a link to another page, or
> another image, it automatically supplies the URL of the page it followed
> to get there.  Is there a way to do this?

Somebody already asked for this and AFAICT, there's no way to do
that.

> > > 3) Multi-threading.
> >
> > I suppose you mean downloading several URIs in parallel.  No, wget
> > doesn't support that.  Sometimes, however, one may start several wget
> > in parallel, thanks to the shell (the & operator on Bourne shells).
> 
> No, I mean downloading multiple files from the SAME uri in parallel,
> instead of downloading files one-by-one-by-one (thus saving time on a fast
> pipe).

This doesn't make sense to me. When downloading from a single
server, the bottleneck is generally either the server or the link
; in either case, there's nothing to win by attempting several
simultaneous transfers. Unless there are several servers at the
same IP and the bottleneck is the server, not the link ?

-- 
André Majorel http://www.teaser.fr/~amajorel/>
std::disclaimer ("Not speaking for my employer");



Re: cuj.com file retrieving fails -why?

2002-04-03 Thread Ian Abbott

On 3 Apr 2002 at 14:56, Markus Werle wrote:

> Jens Rösner wrote:
> > So, I do not know what your problem is, but is neither wget#s nor cuj's
> > fault, AFAICT.
> 
> :-(

I've just built Wget 1.7 on Linux and it seemed to download your
problem file okay. So I don't know what your problem is either!



Re: Referrer Faking and other nifty features

2002-04-03 Thread Dan Mahoney, System Admin

On Wed, 3 Apr 2002, fabrice bauzac wrote:

> Good morning,
>
> Please note that I am a wget user, there may be errors.
>
> On Tue, Apr 02, 2002 at 11:50:03PM -0500, Dan Mahoney, System Admin
> wrote:
>
> > 1) referrer faking (i.e., wget automatically supplies a referrer
> > based on the, well, referring page)
>
> It is the --referer option, see (wget)HTTP Options, from the Info
> documentation.

Yes, that allows me to specify _A_ referrer, like www.aol.com.  When I'm
trying to help my users mirror their old angelfire pages or something like
that, very often the link has to come from the same directory.  I'd like
to see something where when wget follows a link to another page, or
another image, it automatically supplies the URL of the page it followed
to get there.  Is there a way to do this?

> > 2) Teh regex support like in the "gold" package that I can no longer
> > find.
>
> No; however you may use shell globs, see (wget)Accept/Reject Options.
>
> > 3) Multi-threading.
>
> I suppose you mean downloading several URIs in parallel.  No, wget
> doesn't support that.  Sometimes, however, one may start several wget
> in parallel, thanks to the shell (the & operator on Bourne shells).

No, I mean downloading multiple files from the SAME uri in parallel,
instead of downloading files one-by-one-by-one (thus saving time on a fast
pipe).

>
> > Also, I have in the past encountered a difficulty with the ~ being
> > escaped the wrong way, has this been fixed?  I know at one point one
> > site suggested you modify url.c to "fix" this.
>
> AFAIK, I have never had that problem; maybe it has been fixed.

I remember the problem now.  I was trying to mirror
homepages.go.com/~something and for whatever reason, wget would follow a
link to homepages.go.com/~somethingelse and parse it out to
homepages.go.com/%7esomethingelse, which for some reason the webserver
DIDN'T like, and because the tilde character IS passable in a url, I was
able to use the windows version, which didn't have this behavior.

>
> > Finally, is there a way to utilize the persistent cookie file that
> > lynx generates to "feed" wget?
>
> There is the --load-cookies=FILE option, see (wget)HTTP Options.
>
> How to read the Info documentation: type "info wget" from a shell.
> The "?" key may help you.  Use "gHTTP Options" to go to node "HTTP
> Options".

Hrmm, one other thought, is there support for sftp in wget?

-Dan Mahoney

--

"If you aren't going to try something, then we might as well just be
friends."

"We can't have that now, can we?"

-SK & Dan Mahoney,  December 9, 1998

Dan Mahoney
Techie,  Sysadmin,  WebGeek
Gushi on efnet/undernet IRC
ICQ: 13735144   AIM: LarpGM
Web: http://prime.gushi.org
finger [EMAIL PROTECTED]
for pgp public key and tel#
---





Re: cuj.com file retrieving fails -why?

2002-04-03 Thread Markus Werle

Jens Rösner wrote:

> Hallo Markus!
>
> This is not a bug (I reckon) and should therefore have been sent to
> the normal wget list.

Sorry.

> Using both wget 1.7.1 and 1.8.1 on Windows the file is
> downloaded with
>
> wget -d -U "Mozilla/5.0 (compatible; Konqueror/2.2.1; Linux)" -r
> http://www.cuj.com/images/resource/experts/alexandr.gif

Does not work for me on linux

> as well as with
>
> wget http://www.cuj.com/images/resource/experts/alexandr.gif

dito.

> So, I do not know what your problem is, but is neither wget#s nor cuj's
> fault, AFAICT.

:-(


Markus




Re: cuj.com file retrieving fails -why?

2002-04-03 Thread Jens Rösner

Hallo Markus!

This is not a bug (I reckon) and should therefore have been sent to 
the normal wget list.

Using both wget 1.7.1 and 1.8.1 on Windows the file is 
downloaded with 

wget -d -U "Mozilla/5.0 (compatible; Konqueror/2.2.1; Linux)" -r
http://www.cuj.com/images/resource/experts/alexandr.gif

as well as with

wget http://www.cuj.com/images/resource/experts/alexandr.gif

So, I do not know what your problem is, but is neither wget#s nor cuj's
fault, AFAICT.

CU
Jens


> 
> This problem is independent on whether a proxy is used or not:
> The download hangs, though I can read the content using konqueror.
> So what do cuj people do to inhibit automatic download and how can
> I circumvent it?
> 
> 
> wget --proxy=off -d -U "Mozilla/5.0 (compatible; Konqueror/2.2.1; Linux)"
> -r http://www.cuj.com/images/resource/experts/alexandr.gif
> DEBUG output created by Wget 1.7 on linux.
> 
> parseurl ("http://www.cuj.com/images/resource/experts/alexandr.gif";) ->
> host www.cuj.com -> opath images/resource/experts/alexandr.gif -> dir
> images/resource/experts -> file alexandr.gif -> ndir
> images/resource/experts
> newpath: /images/resource/experts/alexandr.gif
> Checking for www.cuj.com in host_name_address_map.
> Checking for www.cuj.com in host_slave_master_map.
> First time I hear about www.cuj.com by that name; looking it up.
> Caching www.cuj.com <-> 66.35.216.85
> Checking again for www.cuj.com in host_slave_master_map.
> --14:32:35--  http://www.cuj.com/images/resource/experts/alexandr.gif
>=> `www.cuj.com/images/resource/experts/alexandr.gif'
> Connecting to www.cuj.com:80... Found www.cuj.com in
> host_name_address_map: 66.35.216.85
> Created fd 3.
> connected!
> ---request begin---
> GET /images/resource/experts/alexandr.gif HTTP/1.0
> User-Agent: Mozilla/5.0 (compatible; Konqueror/2.2.1; Linux)
> Host: www.cuj.com
> Accept: */*
> Connection: Keep-Alive
> 
> ---request end---
> HTTP request sent, awaiting response...
> 
> nothing happens 
> 
> 
> Markus
> 

-- 
GMX - Die Kommunikationsplattform im Internet.
http://www.gmx.net




cuj.com file retrieving fails -why?

2002-04-03 Thread Markus Werle

Hi!

This problem is independent on whether a proxy is used or not:
The download hangs, though I can read the content using konqueror.
So what do cuj people do to inhibit automatic download and how can
I circumvent it?


wget --proxy=off -d -U "Mozilla/5.0 (compatible; Konqueror/2.2.1; Linux)" -r 
http://www.cuj.com/images/resource/experts/alexandr.gif
DEBUG output created by Wget 1.7 on linux.

parseurl ("http://www.cuj.com/images/resource/experts/alexandr.gif";) -> host 
www.cuj.com -> opath images/resource/experts/alexandr.gif -> dir
images/resource/experts -> file alexandr.gif -> ndir images/resource/experts
newpath: /images/resource/experts/alexandr.gif
Checking for www.cuj.com in host_name_address_map.
Checking for www.cuj.com in host_slave_master_map.
First time I hear about www.cuj.com by that name; looking it up.
Caching www.cuj.com <-> 66.35.216.85
Checking again for www.cuj.com in host_slave_master_map.
--14:32:35--  http://www.cuj.com/images/resource/experts/alexandr.gif
   => `www.cuj.com/images/resource/experts/alexandr.gif'
Connecting to www.cuj.com:80... Found www.cuj.com in host_name_address_map: 
66.35.216.85
Created fd 3.
connected!
---request begin---
GET /images/resource/experts/alexandr.gif HTTP/1.0
User-Agent: Mozilla/5.0 (compatible; Konqueror/2.2.1; Linux)
Host: www.cuj.com
Accept: */*
Connection: Keep-Alive

---request end---
HTTP request sent, awaiting response...

nothing happens 


Markus




Re: option changed: -nh -> -nH

2002-04-03 Thread Jens Rösner

Hi Noèl!

-nh
and
-nH
are totally different.

from wget 1.7.1 (I think the last version to offer both):
`-nh'
`--no-host-lookup'
 Disable the time-consuming DNS lookup of almost all hosts (*note
 Host Checking::).

`-nH'
`--no-host-directories'
 Disable generation of host-prefixed directories.  By default,
 invoking Wget with `-r http://fly.srk.fer.hr/' will create a
 structure of directories beginning with `fly.srk.fer.hr/'.  This
 option disables such behavior.

For wget 1.8.x -nh became the default behavior.
Switching back to host-look-up is not possible.
I already complained that many old scripts now break and suggested 
that entering -nh at the command line would 
either be completely ignored or the user would be 
informed and wget executed nevertheless.
Apparently this was not regarded as useful.

CU
Jens


> 
> The option --no-host-directories
> changed from -nh to -nH (v1.8.1).
> 
> Is there a reason for this?
> It breaks a lot of scripts when upgrading,
> I think.
> 
> Could this be changed back to -nh?
> 
> Thank you.
> 
> -- 
>   Noèl Köthe
> 

-- 
GMX - Die Kommunikationsplattform im Internet.
http://www.gmx.net




option changed: -nh -> -nH

2002-04-03 Thread Noel Koethe

Hello,

The option --no-host-directories
changed from -nh to -nH (v1.8.1).

Is there a reason for this?
It breaks a lot of scripts when upgrading,
I think.

Could this be changed back to -nh?

Thank you.

-- 
Noèl Köthe



Re: new platform compile

2002-04-03 Thread Ian Abbott

On 2 Apr 2002 at 22:34, Phillip Morelock wrote:

> Just thought I would let you know that I have compiled wget successfully 
> on mac os x, in case you want to update your architectures or whatever 
> (it was mentioned in the documentation to send this if i succeeded).

Thanks. Actually, the MACHINES file in the current CVS version
already mentions it!

> But it's not working.   i made symlinks in every place i could think 
> of.  I mean, the configure script says it can't find a bunch of stuff 
> it's expecting for ssl, but maybe it works...make and make install don't 
> complain.

If configure couldn't find it, the executable won't be built with
SSL support. At the moment, the configure scripts only check for
the OpenSSL package. Maybe Apple supply their own SSL package. If
that is the case, you'll have to wait for someone to support it in
Wget or install the OpenSSL package into /usr/local (so it
doesn't clash with Apple's SSL package) before configuring Wget.

> Also I noted that the install documentation is a little out of whack WRT 
> the --info-dir and --mandir params.  It puts man1 directly under 
> /usr/share in my case (i supplied /usr/share because INSTALL said it 
> puts it into ${--mandir}/man . --infodir works properly as documented, 
> but the INSTALL document appears to say differently for --mandir (at 
> least on my system).

If you want /usr/share/man/man1/wget.1 and
/usr/share/info/wget.info you should use --mandir=/usr/share/man
--infodir=/usr/share/info. The default for --mandir is $PREFIX/man
and the default for --infodir is $PREFIX/info, where $PREFIX can be
set with the --prefix=$PREFIX option and defaults to /usr/local. I
think that is what the INSTALL document says in fact, if you read
it a bit more closely.