Re: Cookie options

2003-11-05 Thread Daniel Stenberg
On Tue, 4 Nov 2003, Hrvoje Niksic wrote:

 Interestingly enough, curl has a `--junk-session-cookies', which indicates
 that it keeps them by default (?).  Daniel, are you still listening?  :-)

I am, and that is correct!

I never thought that single-shot curl invokes could be seen as single
sessions and therefore I think removing (or not saving) session cookies is
plain wrong and will only make things harder and more confusing to users. But
then, I believe the average usage of wget and curl differ a bit.

Many users that use curl to script actions and that use cookies, operate with
multiple curl command lines.

The option named --junk-session-cookies is a rather new baby in curl land, as
previously curl never ignored session cookies at load-time.

Also, curl doesn't have any option to send out-of-date cookies such as has
been proposed here as an option for wget, and I've never got a request from a
single soul to add it to curl. I'd say it is an indication that such an option
would not be widely used. (Given that curl has been used for scripting POSTs,
logins and cookie stuff for years already.)

-- 
 -=- Daniel Stenberg -=- http://daniel.haxx.se -=-
  ech`echo xiun|tr nu oc|sed 'sx\([sx]\)\([xoi]\)xo un\2\1 is xg'`ol


RE: About termination of wget and -T option

2003-11-05 Thread Luigi Stefano Sona (lsona)
Hrvoje,
This is different from what I observed:
- I execute a cgi by wget. The cgi should return a file of about 7MB.
- Data start coming, and after 15 minutes, while the file is not
complete, wget starts a new trial.

Luigi 

-Original Message-
From: Hrvoje Niksic [mailto:[EMAIL PROTECTED] 
Sent: Tuesday, November 04, 2003 9:35 PM
To: Luigi Stefano Sona (lsona)
Cc: [EMAIL PROTECTED]
Subject: Re: About termination of wget and -T option


Luigi Stefano Sona (lsona) [EMAIL PROTECTED] writes:

 About -T, you confirm that anyway, the timeout is for the start of
 the answer, not for the finish ?

Almost.  In fact, the timeout applies whenever the download stalls,
at any point when Wget waits for data to arrive.  You can think of it
this way: Wget reads data in a loop like this one:

while data_pending:
  read_chunk_from_network
  write_chunk_to_disk

If a read_chunk_from_network step takes more than 15min, the
download is interrupted (and retried).  But the whole download can
take as long as it takes.



Re: Cookie options

2003-11-05 Thread Hrvoje Niksic
Daniel Stenberg [EMAIL PROTECTED] writes:

 I never thought that single-shot curl invokes could be seen as
 single sessions and therefore I think removing (or not saving)
 session cookies is plain wrong and will only make things harder and
 more confusing to users.

Good point, although I somewhat disagree about the plain wrong part.
From what I gather, the `cookies.txt' file to which Wget and curl dump
their cookies is meant to hold permanent cookies.  Putting session
cookies inside might confuse other browsers.

That notwithstanding, another problem is that, since session cookies
have no expiry time, saving them by default would make them
effectively immortal, which sounds quite wrong to me.  I may be
missing curl-specific details, though.

 The option named --junk-session-cookies is a rather new baby in curl
 land, as previously curl never ignored session cookies at load-time.

Hmm.  Wget will not ignore session cookies at load-time (any more),
regardless of options.  `--keep-session-cookies' causes it not to
discard them at save time.

 Also, curl doesn't have any option to send out-of-date cookies such
 as has been proposed here as an option for wget, and I've never got
 a request from a single soul to add it to curl. I'd say it is an
 indication that such an option would not be widely used. (Given that
 curl has been used for scripting POSTs, logins and cookie stuff for
 years already.)

Thanks for the insight.  If someone *really* wants to send stale
cookies, he can still hand-hack the cookies.txt file.


Re: About termination of wget and -T option

2003-11-05 Thread Hrvoje Niksic
Luigi Stefano Sona (lsona) [EMAIL PROTECTED] writes:

 Hrvoje,
 This is different from what I observed:
 - I execute a cgi by wget. The cgi should return a file of about 7MB.
 - Data start coming, and after 15 minutes, while the file is not
 complete, wget starts a new trial.

I don't doubt your observation, but it might well be the result of a
different problem.  The debug output might help narrow it down.


Re: FTP time-stamping and time zones

2003-11-05 Thread Hrvoje Niksic
Jochen Hayek [EMAIL PROTECTED] writes:

 But as FTP directory listings do not include the time zone, the
 files live in, (I think) wget just assumes the local time zone to be
 identical to the remote one.  Am I right with this?

 As long as this concept is being kept to entirely,
 this is probably the best, that can be done.

You're right.  Wget assumes local time zone and sets the local time
stamp.  Even if the time stamp is not correct (because of time zone
difference), it will be *the same* incorrect time stamp each time
around and mirroring will work.

 But it leads to a situation, that wget creates files with time
 stamps much older, than the files actually are.

Or newer -- yes.  Until MLST/MDTM are supported, there is no good cure
for this (that I can think of).

 I wouldn't mind making wget believe (maybe through setting the
 environment variable TZ) it actually lives in New-England,
 although it lives here in Europe.

That might work.  Wget uses the library functions for time
conversions, and they should respect TZ.


Re: Cookie options

2003-11-05 Thread Daniel Stenberg
On Wed, 5 Nov 2003, Hrvoje Niksic wrote:

 Good point, although I somewhat disagree about the plain wrong part. From
 what I gather, the `cookies.txt' file to which Wget and curl dump their
 cookies is meant to hold permanent cookies.  Putting session cookies inside
 might confuse other browsers.

Not necessarily, since they are stored without any expire time (using zero)
and ordinary non-session cookies always have a number different than zero in
that field.

But sure, they might still confuse a browser that reads the file. I've never
considered that a problem and I've never got any problems with that behavior
reported. I believe very few people actually try that stunt.

 That notwithstanding, another problem is that, since session cookies have no
 expiry time, saving them by default would make them effectively immortal,
 which sounds quite wrong to me.  I may be missing curl-specific details,
 though.

Yes, it makes them immortal and lets the user decide when the session ends
(by using the correct option). There is no one but the user who can tell when
one session ends and another starts. In most cases, people will make a fresh
cookie file each time they start a new session and then just keep session
cookies around until they're done for this time. (Based on what I've done
myself and seen others do with automated curl operations using cookies.)

 Hmm.  Wget will not ignore session cookies at load-time (any more),
 regardless of options.  `--keep-session-cookies' causes it not to discard
 them at save time.

Right, but then again people who use wget are more likely (I'm guessing here)
to use one single wget command line and consider that to be the full session.
In curl's case, when using cookies you will very often enter multiple command
lines to sort out a single session's business.

-- 
 -=- Daniel Stenberg -=- http://daniel.haxx.se -=-
  ech`echo xiun|tr nu oc|sed 'sx\([sx]\)\([xoi]\)xo un\2\1 is xg'`ol


retrieving only index.html

2003-11-05 Thread Thomas Wagner
Hi,

we have a problem downloading files using wget version 1.9 without
explicitly providing the file names. With --accept we have specified the files we want
to receive. Version 1.7 works as expected. Version 1.9 is retrieving the index-file 
only. 
We have searched the internet for any clues. But no luck. We want to get all files that
have the extensions zip and ini. Version 1.7 provides the index file but also the 
other 
files. The shell script is listed below. 
Thanks for your attention

#!/bin/sh

VIRPATH=/usrhome/ftp/pub/antivir
CHKPATH=/usrhome/ftp/antivir.new

SERVERS=ftp.nai.com ftpeur.nai.com ftpde.nai.com
URL=pub/antivirus/datfiles/4.x/
 
# autoreorg:
TMPLOG=/tmp/$P.log.$$
http_proxy=x.x.x.x:x
ftp_proxy=x.x.x.x:
export ftp_proxy 
export http_proxy
 

for server in $SERVERS ; do
   
echo +++ Pruefe ftp://$server/$URL;
rm -f index.html
./wget \
--no-directories \
--no-parent \
--recursive \
--level=1 \
--no-clobber \
--cache=off \
--proxy=on \
--accept=.zip,.ini \
ftp://$server/$URL
rc=$?
[ $rc = 0 ]  break
done


Output using wget version 1.7 
This is actualiy the second call. Just to make the output shorter.
All files have been downloaded with the first call.

+++ Testing ftp://ftp.nai.com/pub/antivirus/datfiles/4.x/
--12:36:47--  ftp://ftp.nai.com/pub/antivirus/datfiles/4.x/
= `index.html' 
Proxy request sent, awaiting response... 200 OK 
Length: unspecified [text/html]
   
  
0K .. @  11.18 KB/s   
12:36:50 (11.18 KB/s) - `index.html' saved [2748]
File `dat-4301.zip' already there, will not retrieve.
File `delta.ini' already there, will not retrieve. 
File `update.ini' already there, will not retrieve.
FINISHED --12:36:50--
Downloaded: 2,748 bytes in 1 files   

Output using wget version 1.9. The script ist the same.

+++ Testing ftp://ftp.nai.com/pub/antivirus/datfiles/4.x/ 
--12:51:35--  ftp://ftp.nai.com/pub/antivirus/datfiles/4.x/  
 = `index.html' 
Resolving wgetproxy... x.x.x.x 
Proxy request sent, awaiting response... 200 OK 
Length: unspecified [text/html] 
[  =] 2,748 12.20K/s 
12:51:37 (12.20 KB/s) - `index.html' saved [2748] 
FINISHED --12:51:37--  
Downloaded: 2,748 bytes in 1 files

Mit freundlichen Gren / Best regards 

 
Thomas Wagner 

KORDOBA GmbH  Co. KG 
Tel.:  +49 69 797 - 6539
Fax:  +49 69 797 - 6597
[EMAIL PROTECTED] 
http://www.kordoba.de 
 



GNU TLS vs. OpenSSL

2003-11-05 Thread Hrvoje Niksic
Does anyone know what the situation is regarding the use of GNU TLS
vs. OpenSSL?  Last time I checked (several years ago), GNU TLS was
still in beta; also, it was incompatible with OpenSSL.

Is GNU TLS usable now?  Should Wget try to use it in preference to
OpenSSL and thus get rid of the GPL exception?


RE: About termination of wget and -T option

2003-11-05 Thread Luigi Stefano Sona (lsona)
This is the end of debug output...


===
...
 6250K .. .. .. .. ..   12.11
KB/s
 6300K .. .. .. .. ..   11.36
KB/s
 6350K .. .. .. .. ..   10.96
KB/s
 6400K .. .. .. .. ..8.33
KB/s
 6450K .. .. .. .. ..2.02
KB/s
 6500K .. .. .. .. ..5.90
KB/s
 6550K .. .. .. .. ..7.76
KB/s
 6600K .. .. .. .. ..7.03
KB/s
 6650K .. .. .. .. . 4.56
KB/s

Closing fd 5
15:05:13 (6.74 KB/s) - `pages/OM2_report_21705.txt' saved [6852406]


===


The file received is not complete.

Luigi


-Original Message-
From: Hrvoje Niksic [mailto:[EMAIL PROTECTED] 
Sent: Wednesday, November 05, 2003 12:36 PM
To: Luigi Stefano Sona (lsona)
Cc: [EMAIL PROTECTED]
Subject: Re: About termination of wget and -T option


Luigi Stefano Sona (lsona) [EMAIL PROTECTED] writes:

 Hrvoje,
 This is different from what I observed:
 - I execute a cgi by wget. The cgi should return a file of about 7MB.
 - Data start coming, and after 15 minutes, while the file is not
 complete, wget starts a new trial.

I don't doubt your observation, but it might well be the result of a
different problem.  The debug output might help narrow it down.


Re: About termination of wget and -T option

2003-11-05 Thread Hrvoje Niksic
Luigi Stefano Sona (lsona) [EMAIL PROTECTED] writes:

 ---request end---
 HTTP request sent, awaiting response... HTTP/1.1 200 OK
 Date: Wed, 05 Nov 2003 13:48:56 GMT
 Server: Apache/1.3.26 (Unix) PHP/4.2.2 mod_perl/1.27
 Content-Type: text/plain
 Via: 1.1 Application and Content Networking System Software 5.0.5
 Connection: Close
[...]
  6650K .. .. .. .. . 4.56
 KB/s

 Closing fd 5
 15:05:13 (6.74 KB/s) - `pages/OM2_report_21705.txt' saved [6852406]

As far as I can tell:

* Either your server is not sending Content-Length for some reason
  (such as the file being dynamically generated), or a proxy server is
  stripping it.

* The download is interrupted without a clear indication that this has
  happened.  Wget's network read indicates end-of-transfer.

* Since Wget doesn't know the length of the file, it cannot know that
  end-of-transfer is an error and considers the file fully retrieved.

A question: are you absolutely certain that the file is not truncated
on the server?  Can you download it using a web browser?


Re: gettext and charsets

2003-11-05 Thread Karl Eichwalder
Hrvoje Niksic [EMAIL PROTECTED] writes:

 Perhaps you could try posting to [EMAIL PROTECTED]?  Failing
 that, you might want to try at the address of the Free Translation
 Project and/or the Norwegian national team near you.

[EMAIL PROTECTED] is the preferred list for reporting gettext
bugs.

Implementation issues can be discussed at
[EMAIL PROTECTED] (also read by Bruno Haible).


RE: About termination of wget and -T option

2003-11-05 Thread Luigi Stefano Sona (lsona)
The file is dinamically generated.

Most of times the download goes well, but in specific hours it doesn't
complete.

I've tried from Netscape directly, and even in this case the download is
not completed sometimes.

Thanks
Luigi

-Original Message-
From: Hrvoje Niksic [mailto:[EMAIL PROTECTED] 
Sent: Wednesday, November 05, 2003 3:40 PM
To: Luigi Stefano Sona (lsona)
Cc: [EMAIL PROTECTED]
Subject: Re: About termination of wget and -T option


Luigi Stefano Sona (lsona) [EMAIL PROTECTED] writes:

 ---request end---
 HTTP request sent, awaiting response... HTTP/1.1 200 OK
 Date: Wed, 05 Nov 2003 13:48:56 GMT
 Server: Apache/1.3.26 (Unix) PHP/4.2.2 mod_perl/1.27
 Content-Type: text/plain
 Via: 1.1 Application and Content Networking System Software 5.0.5
 Connection: Close
[...]
  6650K .. .. .. .. . 4.56
 KB/s

 Closing fd 5
 15:05:13 (6.74 KB/s) - `pages/OM2_report_21705.txt' saved [6852406]

As far as I can tell:

* Either your server is not sending Content-Length for some reason
  (such as the file being dynamically generated), or a proxy server is
  stripping it.

* The download is interrupted without a clear indication that this has
  happened.  Wget's network read indicates end-of-transfer.

* Since Wget doesn't know the length of the file, it cannot know that
  end-of-transfer is an error and considers the file fully retrieved.

A question: are you absolutely certain that the file is not truncated
on the server?  Can you download it using a web browser?


Hello, new feature added!

2003-11-05 Thread Dennis Smit
Heya people!

I am new to the list so hello.

I sended a patch to add the --preserve-permissions option to wget this
option can be used to preserve permissions when retrieving dirlistings
from a remote site. The new standard behavior is not copying permissions
this is because users often don't want the permissions a server had set.



RE: About termination of wget and -T option

2003-11-05 Thread Luigi Stefano Sona (lsona)
DEBUG output created by Wget 1.8.2 on solaris2.8.

--14:48:31--  http://xxx
   = `pages/OM2_report_21705.txt'
Resolving yyy... done.
Caching  = NN.MM.NN.MM
Connecting to yyy[NN.MM.NN.MM]:80... connected.
Created socket 5.
Releasing nn (new refcount 1).
---request begin---
GET xxx HTTP/1.0^M
User-Agent: Wget/1.8.2^M
Host: x^M
Accept: */*^M
Connection: Keep-Alive^M
Authorization: Basic x=^M
^M
---request end---
HTTP request sent, awaiting response... HTTP/1.1 200 OK
Date: Wed, 05 Nov 2003 13:48:56 GMT
Server: Apache/1.3.26 (Unix) PHP/4.2.2 mod_perl/1.27
Content-Type: text/plain
Via: 1.1 Application and Content Networking System Software 5.0.5
Connection: Close


Length: unspecified [text/plain]

0K .. .. .. .. ..3.69
KB/s
   50K .. .. .. .. ..   39.68
KB/s
  100K .. .. .. .. ..   21.19
KB/s
  150K .. .. .. .. ..   26.58
KB/s
  200K .. .. .. .. ..   31.67
KB/s
  250K .. .. .. .. ..   33.78
KB/s
..
 6250K .. .. .. .. ..   12.11
KB/s
 6300K .. .. .. .. ..   11.36
KB/s
 6350K .. .. .. .. ..   10.96
KB/s
 6400K .. .. .. .. ..8.33
KB/s
 6450K .. .. .. .. ..2.02
KB/s
 6500K .. .. .. .. ..5.90
KB/s
 6550K .. .. .. .. ..7.76
KB/s
 6600K .. .. .. .. ..7.03
KB/s
 6650K .. .. .. .. . 4.56
KB/s

Closing fd 5
15:05:13 (6.74 KB/s) - `pages/OM2_report_21705.txt' saved [6852406]


-Original Message-
From: Hrvoje Niksic [mailto:[EMAIL PROTECTED] 
Sent: Wednesday, November 05, 2003 3:17 PM
To: Luigi Stefano Sona (lsona)
Cc: [EMAIL PROTECTED]
Subject: Re: About termination of wget and -T option


Luigi Stefano Sona (lsona) [EMAIL PROTECTED] writes:

 This is the end of debug output...

Why not the whole output?  Some crucial portions are missing.  If you
have confidential information in the output, feel free to delete it
first.

From seeing this portion, it seems that there is no timeout, and that
Wget believes that the whole file has been retrieved.  Additionally,
the server does not appear to be sending the `Content-Type' header, so
Wget cannot tell that the file is incomplete.


Re: About termination of wget and -T option

2003-11-05 Thread Hrvoje Niksic
Luigi Stefano Sona (lsona) [EMAIL PROTECTED] writes:

 The file is dinamically generated.

 Most of times the download goes well, but in specific hours it doesn't
 complete.

 I've tried from Netscape directly, and even in this case the
 download is not completed sometimes.

This could be a bug in the dynamic page that generates the file, or
possibly a problem with the proxy server or with your network
configuration.

Wget is designed to deal with this issue, but if the sevrer doesn't
provide Content-Length, its hands are tied.  If you can convince the
author of the dynamic page to send Content-Length, Wget will detect
the problem and retry.  Otherwise, there is simply no indication that
anything is wrong, and the browser cannot help you.


Re: The libtool situation

2003-11-05 Thread Daniel Stenberg
On Wed, 5 Nov 2003, Hrvoje Niksic wrote:

 Is it me, or is there something really strange going on with libtool? I
 tried updating Wget's libtool, but libtool 1.5 is *still* unavailable from
 the GNU FTP site and its mirrors.  It points to a BACK-RSN type of file
 that explains how they've been compromised, and so on.

It must be something that has happened recently, as the 1.5 version used to be
around on all the GNU mirrors. In case it helps, I happened to find my own
downloaded package lying around and I can offer it for download from here:

http://www.contactor.se/~dast/libtool-1.5.tar.gz

-- 
 -=- Daniel Stenberg -=- http://daniel.haxx.se -=-
  ech`echo xiun|tr nu oc|sed 'sx\([sx]\)\([xoi]\)xo un\2\1 is xg'`ol


Windows msvc 1.9-stable bugfix release

2003-11-05 Thread Herold Heiko
Binary for current 1.9 cvs branch at http://xoomer.virgilio.it/hherold/

Heiko

-- 
-- PREVINET S.p.A. www.previnet.it
-- Heiko Herold [EMAIL PROTECTED]
-- +39-041-5907073 ph
-- +39-041-5907472 fax


RE: GNU TLS vs. OpenSSL

2003-11-05 Thread Aaron S. Hawley
From Various Licenses and Comments about Them
http://www.gnu.ctssn.com/licenses/license-list.html

The OpenSSL license.
The license of OpenSSL is a conjunction of two licenses, One of them
being the license of SSLeay. You must follow both. The combination results
in a copyleft free software license that is incompatible with the GNU GPL.
It also has an advertising clause like the original BSD license and the
Apache license.

We recommend using GNUTLS instead of OpenSSL in software you write.
However, there is no reason not to use OpenSSL and applications that work
with OpenSSL.

shouldn't the above say say non-copyleft?
/a

On Wed, 5 Nov 2003, Post, Mark K wrote:

 I'm a little confused.  OpenSSL is licensed pretty much the same as Apache.
 What's the GPL issue with that style of license?


 Mark Post

 -Original Message-
 From: Hrvoje Niksic [mailto:[EMAIL PROTECTED]
 Sent: Wednesday, November 05, 2003 8:27 AM
 To: [EMAIL PROTECTED]
 Subject: GNU TLS vs. OpenSSL


 Does anyone know what the situation is regarding the use of GNU TLS
 vs. OpenSSL?  Last time I checked (several years ago), GNU TLS was
 still in beta; also, it was incompatible with OpenSSL.

 Is GNU TLS usable now?  Should Wget try to use it in preference to
 OpenSSL and thus get rid of the GPL exception?


Re: The libtool situation

2003-11-05 Thread Hrvoje Niksic
Daniel Stenberg [EMAIL PROTECTED] writes:

 On Wed, 5 Nov 2003, Hrvoje Niksic wrote:

 Is it me, or is there something really strange going on with
 libtool? I tried updating Wget's libtool, but libtool 1.5 is
 *still* unavailable from the GNU FTP site and its mirrors.  It
 points to a BACK-RSN type of file that explains how they've been
 compromised, and so on.

 It must be something that has happened recently, as the 1.5 version
 used to be around on all the GNU mirrors.

I'm not sure this is that recent, unless there was a new compromise I
never heard about.  As for the mirrors, well, they're mirrors -- they
just picked up the new file, which happens to be a symlink.  (Which is
just as well because they would have happily picked up the compromised
version too.)

 In case it helps, I happened to find my own downloaded package lying
 around and I can offer it for download from here:
[...]

Thanks, but I got it off Debian.  Although I suppose I could pick it
up and compare it with what I got just in case.



WGET Unable to Get Through Proxy Server with HTTPS

2003-11-05 Thread Alesna, Sancho
I'm unable to get through the proxy server when downloading via WGET HTTPS.  The 
following is the message.  Can you help?
 
[] C:\WGET\wget-1.9\srcwget.exe -dS --proxy-user=censored --proxy-p
asswd=censored  https:// 
:@momentumw1.boh.com/WebDownloads/bankahttps://censored id:censored 
pwd@momentumw1.boh.com/WebDownloads/banka
_an/andrept.txt030919105214
wget: debug support not compiled in.
--13:21:50--  https://banka_an:[EMAIL PROTECTED]/WebDownloads/banka_
an/andrept.txt030919105214
   = `andrept.txt030919105214.1'
Resolving censored proxy.bankofamerica.com... censored ip 
Connecting to censored proxy[censored ip]:8080... connected.
Proxy request sent, awaiting response...
 1 HTTP/1.0 502 internal error - server connection terminated
 2 Date: Wed, 05 Nov 2003 21:21:50 GMT
 3 Via: HTTPS/1.1 nup-ctc (Traffic-Server/4.0.9 [c sEf ])
 4 Cache-Control: no-store
 5 Content-Type: text/html
 6 Content-Language: en
 7 Content-Length: 368
13:21:50 ERROR 502: internal error - server connection terminated.
 
Thank you.
 
 
 

 


Re: GNU TLS vs. OpenSSL

2003-11-05 Thread Hrvoje Niksic
Post, Mark K [EMAIL PROTECTED] writes:

 I'm a little confused.  OpenSSL is licensed pretty much the same as
 Apache.  What's the GPL issue with that style of license?

I don't speak for the FSF, but from reading the license, it imposes
additional constraints, specifically the advertising clause.  That
makes it incompatible with the GPL, despite of it clearly being free
software.  There might be other incompatibilities I'm not aware of.

The exception, phrased by Eben Moglen and approved by RMS, allows
distributors such as Debian to legally distribute SSL-enabled Wget
binaries.