Re: (Fwd) Automatic posting to forms

2002-03-08 Thread Tony Lewis
Daniel Stenberg responded to my original suggestion: With this information, any time that wget encounters a form whose action is /cgi-bin/auth.cgi, it will enqueue the submission of the form using the values provided for the fields id and pw. Now, why would wget do this? There are many

Re: wget parsing JavaScript

2002-03-26 Thread Tony Lewis
Csaba Ráduly wrote: I see that wget handles SCRIPT with tag_find_urls, i.e. it tries to parse whatever it's inside. Why was this implemented ? JavaScript is most used to construct links programmatically. wget is likely to find bogus URLs until it can properly parse JavaScript. wget is

Re: wget parsing JavaScript

2002-03-26 Thread Tony Lewis
I wrote: wget is parsing the attributes within the script tag, i.e., script src=url. It does not examine the content between script and /script. and Ian Abbott responded: I think it does, actually, but that is mostly harmless. You're right. What I meant was that it does not examine the

Re: wget parsing JavaScript

2002-03-27 Thread Tony Lewis
Ian Abbott wrote: For example, a recursive retrieval on a page like this: html body script a href=foo.htmlfoo/a /script /body /html will retrieve foo.html, regardless of the script.../script tags. We seem to be talking about two completely different things, Ian. A

Re: Referrer Faking and other nifty features

2002-04-03 Thread Tony Lewis
Andre Majorel wrote: Yes, that allows me to specify _A_ referrer, like www.aol.com. When I'm trying to help my users mirror their old angelfire pages or something like that, very often the link has to come from the same directory. I'd like to see something where when wget follows a

Re: Current download speed in progress bar

2002-04-09 Thread Tony Lewis
Hrvoje Niksic wrote: The one remaining problem is the ETA. Based on the current speed, it changes value wildly. Of course, over time it is generally decreasing, but one can hardly follow it. I removed the flushing by making sure that it's not shown more than once per second, but this

Re: Current download speed in progress bar

2002-04-09 Thread Tony Lewis
Hrvoje Niksic wrote: I'll grab the other part and explain what curl does. It shows a current speed based on the past five seconds, Does it mean that the speed doesn't change for five seconds, or that you always show the *current* speed, but relative to the last five seconds? I may be

Re: HTTP 1.1

2002-04-12 Thread Tony Lewis
Hrvoje Niksic wrote: Is there any way to make Wget use HTTP/1.1 ? Unfortunately, no. In looking at the debug output, it appears to me that wget is really sending HTTP/1.1 headers, but claiming that they are HTTP/1.0 headers. For example, the Host header was not defined in RFC 1945, but wget

Re: apache irritations

2002-04-22 Thread Tony Lewis
Maciej W. Rozycki wrote: Hmm, it's too fragile in my opinion. What if a new version of Apache defines a new format? I think all of the expressions proposed thus far are too fragile. Consider the following URL: http://www.google.com/search?num=100q=%2Bwget+-GNU The regular expression needs

Re: apache irritations

2002-04-22 Thread Tony Lewis
Maciej W. Rozycki wrote: I'm not sure what you are referring to. We are discussing a common problem with static pages generated by default by Apache as index.html objects for server's filesystem directories providing no default page. Really? The original posting from Jamie Zawinski said:

Re: wget does not honour content-length http header [http://bugs.debian.org/143736]

2002-04-25 Thread Tony Lewis
Hrvoje Niksic wrote: If your point is that Wget should print a warning when it can *prove* that the Content-Length data it received was faulty, as in the case of having received more data, I agree. We're already printing a similar warning when Last-Modified is invalid, for example. I'm

Re: Virus mails

2002-04-27 Thread Tony Lewis
Brix Lichtenberg wrote: But I'm still getting three or more virus mails with attachments 100k+ daily from the wget lists and they're blocking my mailbox (dial-up). And getting those dumb system warnings accompanying them doesn't make it better. Isn't there really no way to stop that (at

Re: (Extended) Reading commandline option values from files or file descriptors (for wget v1.8.1)

2002-04-29 Thread Tony Lewis
Herold Heiko wrote: It would be better imho if the options itself are modified, in that case the variable option wouldn't be necessary, supposing we keep the and :, this could be --@http-passwd=passwd.txt --:proxy-passwd=0 It seems to me that a convention like this should be adopted (or

Re: Virus messages .....

2002-05-06 Thread Tony Lewis
Frank Helk wrote: Free (web based) scanning is available at http://www.antivirus.com. Select Free tools in the top menu and then Scan Your PC, Free from the list. You'll not even have to register to use it. Please. It may not be so simple. Klez uses anti-anti-virus techniques to prevent

Re: newbie doubts

2002-12-04 Thread Tony Lewis
Nandita Shenvi wrote: I have not copied the whole script but just the last few lines.The variable $all_links[3] has an URL: http://bolinux39.europe.nokia.com/database2/MIDI100/GS001/01FINALC.MID. the link follows a file, which I require. I remove the http:// before calling the wget, but i

Re: ralated links in javascripts script

2002-12-16 Thread Tony Lewis
cyprien wrote: I want to mirror my homesite, everything works fine expect one : my site is a photo site based on php scipts : gallery (http://gallery.sourceforge.net) it have also some javascripts script... [snip] what can i do to have that (on mirror site) : You cannot because wget does

Re: image tags not read

2003-01-04 Thread Tony Lewis
Johannes Berg wrote: Maybe this isn't really a bug in wget but rather in the file, but since this is standard as exported from MS Word I'd like to see wget recognize the images and download them. Microsoft Word claims to create a valid HTML file. In fact, what it creates can only reliably be

Re: conditional url encoding

2003-02-22 Thread Tony Lewis
Ryan Underwood wrote: It seems that some servers are broken and in order to fetch files with certain filenames, some characters that are normally encoded in HTTP sequences must be sent through unencoded. For example, I had a server the other day that I was fetching files from at the URL:

Re: Static Mirror of DB-Driven Site

2003-03-17 Thread Tony Lewis
Dan Mahoney, System Admin wrote: Assume I have a site that I want to create a static mirror of. Normally this site is database driven, but I figure if I spider the entire site, and map all the GET URLS to static urls I can have a full mirror. Has anyone known of this being successfully

Re: Cannot get wildcards to work ??

2003-03-28 Thread Tony Lewis
Dick Penny wrote: I have just successfully used WGET on a single file download. I even figured out how to specify a destination. But, I cannot seem to get wildcards to work. Help please: wget -o log.txt -P c:/Documents and Settings/Administrator/My Documents/CME_data/bt

Re: Comment handling

2003-05-31 Thread Tony Lewis
George Prekas wrote: I have found a bug in Wget version 1.8.2 concerning comment handling ( !-- comment -- ). Take a look at the following illegal HTML code: HTML BODY a href=test1.htmltest1.html/a !-- a href=test2.htmltest2.html/a !-- /BODY /HTML Now, save the above snippet as

Re: Comment handling

2003-06-01 Thread Tony Lewis
George Prekas wrote: You are probably right. I have pointed this because I have seen pages that use as a separator !-- with lots of dashes and althrough Internet Explorer shows the page, wget can not download it correctly. What do think about finishing the comment at the ? After

Re: Comment handling

2003-06-03 Thread Tony Lewis
Georg Bauhaus wrote: I don't think so. Actually the rules for SGML comments are somewhat different. Georg, I think we're talking about apples and oranges here. I'm talking about what is legitimate in a comment in an SGML document. I think you're talking about what is legitimate as a comment

Re: Comment handling

2003-06-05 Thread Tony Lewis
Aaron S. Hawley wrote: why not just have the default wget behavior follow comments explicitly (i've lost track whether wget does that or needs to be ammended) /and/ have an option that goes /beyond/ quirky comments and is just --ignore-comments ? :) The issue we've been discussing is what to

Re: Comment handling

2003-06-05 Thread Tony Lewis
Aaron S. Hawley wrote: i'm just saying what's going to happen when someone posts to this list: My Web Pages have [insert obscure comment format] for comments and Wget is considering them to (not) be comments. Can you change the [insert Wget comment mode] comment mode to (not) recognize my

wget is smarter than Internet Explorer!

2003-06-13 Thread Tony Lewis
I tried to retreive a URL with Internet Explorer and it continued to retrieve the URL forever. I tried to grab that same URL with wget, which tried twice and then reported redirection cycle detected. Perhaps we should send the wget code to someone in Redmond. Tony

Re: wget problem

2003-07-03 Thread Tony Lewis
Rajesh wrote: Wget is not mirroring the web site properly. For eg it is not copying symbolic links from the main web server.The target directories do exist on the mirror server. wget can only mirror what can be seen from the web. Symbolic links will be treated as hard references (assuming

Re: wget problem

2003-07-03 Thread Tony Lewis
Rajesh wrote: Thanks for your reply. I have tried using the command wget --user-agent=Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1), but it didn't work. Adding the user agent helps some people -- I think most often with web servers from the evil empire. I have one more question. In

Re: Reminder: wget has no maintainer

2003-08-14 Thread Tony Lewis
Daniel Stenberg wrote: The GNU project is looking for a new maintainer for wget, as the current one wishes to step down. I think that means we need someone who: 1) is proficient in C 2) knows Internet protocols 3) is willing to learn the intricacies of wget 4) has the time to go through

Re: Option to save unfollowed links

2003-10-01 Thread Tony Lewis
Hrvoje Niksic wrote: I'm curious: what is the use case for this? Why would you want to save the unfollowed links to an external file? I use this to determine what other websites a given website refers to. For example: wget

Re: Web page source using wget?

2003-10-06 Thread Tony Lewis
Suhas Tembe wrote: 1). I go to our customer's website every day log in using a User Name Password. [snip] 4). I save the source to a file subsequently perform various tasks on that file. What I would like to do is automate this process of obtaining the source of a page using wget. Is

Re: Using chunked transfer for HTTP requests?

2003-10-07 Thread Tony Lewis
Hrvoje Niksic wrote: Please be aware that Wget needs to know the size of the POST data in advance. Therefore the argument to @code{--post-file} must be a regular file; specifying a FIFO or something like @file{/dev/stdin} won't work. There's nothing that says you have to

Re: Using chunked transfer for HTTP requests?

2003-10-07 Thread Tony Lewis
Hrvoje Niksic wrote: I don't understand what you're proposing. Reading the whole file in memory is too memory-intensive for large files (one could presumably POST really huge files, CD images or whatever). I was proposing that you read the file to determine the length, but that was on the

Re: Using chunked transfer for HTTP requests?

2003-10-07 Thread Tony Lewis
Hrvoje Niksic wrote: That would work for short streaming, but would be pretty bad in the mkisofs example. One would expect Wget to be able to stream the data to the server, and that's just not possible if the size needs to be known in advance, which HTTP/1.0 requires. One might expect it,

How do you pronounce Hrvoje?

2003-10-11 Thread Tony Lewis
I've been on this list for a couple of years now and I've always wondered how our illustrious leader pronounces his name. Can you give us linguistically challenged Americans a phonetic rendition of your name? Tony Lewis (toe knee loo iss)

Re: How do you pronounce Hrvoje?

2003-10-12 Thread Tony Lewis
Hrvoje and I have had an off-list dialogue about this subject. We've settled on HUR-voy-eh as the closest phonetic rendition of his name for English speakers. It helps to remember that the r is rolled. Tony

POST followed by GET

2003-10-14 Thread Tony Lewis
I'm trying to figure out how to do a POST followed by a GET. If I do something like: wget http://www.somesite.com/post.cgi --post-data 'a=1b=2' http://www.somesite.com/getme.html -d I get the following behavior: POST /post.cgi HTTP/1.0 snip [POST data: a=1b=2] snip POST /getme.html HTTP/1.0

Re: POST followed by GET

2003-10-14 Thread Tony Lewis
Hrvoje Niksic wrote: Maybe the right thing would be for `--post-data' to only apply to the URL it precedes, as in: wget --post-data=foo URL1 --post-data=bar URL2 URL3 snip But I'm not at all sure that it's even possible to do this and keep using getopt! I'll start by saying that I

Re: POST followed by GET

2003-10-14 Thread Tony Lewis
Hrvoje Niksic wrote: I like these suggestions. How about the following: for 1.9, document that `--post-data' expects one URL and that its behavior for multiple specified URLs might change in a future version. Then, for 1.10 we can implement one of the alternative behaviors. That works for

Re: Wget 1.9 about to be released

2003-10-16 Thread Tony Lewis
Hrvoje Niksic wrote: I'm about to release 1.9 today, unless it takes more time to upload it to ftp.gnu.org. If there's a serious problem you'd like fixed in 1.9, speak up now or be silent until 1.9.1. :-) I thought we were going to turn our attention to 1.10. :-)

Re: Wget 1.8.2 bug

2003-10-17 Thread Tony Lewis
Hrvoje Niksic wrote: Incidentally, Wget is not the only browser that has a problem with that. For me, Mozilla is simply showing the source of http://www.minskshop.by/cgi-bin/shop.cgi?id=1cookie=set, because the returned content-type is text/plain. On the other hand, Internet Explorer will

Re: wget downloading a single page when it should recurse

2003-10-17 Thread Tony Lewis
Philip Mateescu wrote: A warning message would be nice when for not so obvious reasons wget doesn't behave as one would expect. I don't know if there are other tags that could change wget's behavior (like -r and meta name=robots do), but if they happen it would be useful to have a message.

Re: The patch list

2003-11-04 Thread Tony Lewis
Hrvoje Niksic wrote: I'm curious... is anyone using the patch list to track development? I'm posting all my changes to that list, and sometimes it feels a lot like talking to myself. :-) I read the introductory stuff to see what's changed, but I never extract the patches from the messages.

Re: Does HTTP allow this?

2003-11-10 Thread Tony Lewis
Hrvoje Niksic wrote: Assume that Wget has retrieved a document from the host A, which hasn't closed the connection in accordance with Wget's keep-alive request. Then Wget needs to connect to host B, which is really the same as A because the provider uses DNS-based virtual hosts. Is it OK

Re: Does HTTP allow this?

2003-11-10 Thread Tony Lewis
Hrvoje Niksic wrote: The thing is, I don't want to bloat Wget with obscure options to turn off even more obscure (and *very* rarely needed) optimizations. Wget has enough command-line options as it is. If there are cases where the optimization doesn't work, I'd rather omit it completely.

Re: feature request: --second-guess-the-dns

2003-11-18 Thread Tony Lewis
Hrvoje Niksic wrote: Have you seen the rest of the discussion? Would it do for you if Wget correctly handled something like: wget --header='Host: jidanni.org' http://216.46.192.85/ I think that is an elegant solution. Tony

Re: can you authenticate to a http proxy with a username that contains a space?

2003-11-25 Thread Tony Lewis
antonio taylor wrote: http://fisrtname lastname:[EMAIL PROTECTED] Have you tried http://fisrtname%20lastname:[EMAIL PROTECTED] ?

Re: a problem on wgetting a PNG image

2003-11-27 Thread Tony Lewis
[EMAIL PROTECTED] wrote: I am not sure if this is a bug, but it's really out of my expectation. Here is the way to reproduce the problem. 1. Put the URL http://ichart.yahoo.com/b?s=CSCO into the browser and then drag out the image. It should be a file with .png extension. So I believe this

Re: question

2003-12-03 Thread Tony Lewis
Danny Linkov wrote: I'd like to download recursively the content of a web directory WITHOUT AN INDEX file. What shows up in your web browser if you enter the directory (such as http://www.somesite.com/dir/)? The most common responses are: * some HTML file selected by the server (often

Re: wget Suggestion: ability to scan ports BESIDE #80, (like 443) Anyway Thanks for WGET!

2003-12-07 Thread Tony Lewis
- Original Message - From: [EMAIL PROTECTED] To: [EMAIL PROTECTED] Sent: Sunday, December 07, 2003 8:04 AM Subject: wget Suggestion: ability to scan ports BESIDE #80, (like 443) Anyway Thanks for WGET! What's wrong with wget https://www.somesite.com ?

Re: fork_to_background() on Windows

2003-12-21 Thread Tony Lewis
Gisle Vanem wrote: I've searched google and the only way AFAICS to get redirection in a GUI app to work is to create 3 pipes. Then use a thread (or run_with_timeout with infinite timeout) to read/write the console handles to put/get data into/from the parent's I/O handles. I don't fully

Re: IPv6 support of wget v 1.9.1

2003-12-25 Thread Tony Lewis
Kazu Yamamoto wrote: Thank you for supporting IPv6 in wget v 1.9.1. Unfortunately, wget v 1.9.1 does not work well, at least, on NetBSD. NetBSD does not allow to use IPv4-mapped IPv6 addresses for security reasons. To know the background of this, please refer to:

Re: IPv6 support of wget v 1.9.1

2003-12-25 Thread Tony Lewis
YOSHIFUJI Hideaki wrote: NetBSD etc. is NOT RFC compliant here, however, it would be better if one supports wider platforms / configurations. My patch is quick hack'ed, but I believe that it should work for NetBSD and FreeBSD 5. Please consider applying it. It's not my call as to whether

Re: need help

2003-12-30 Thread Tony Lewis
Anurag Jain wrote: downloading a bin big file(268MB) using wget command on our solrise box using wget http url/bin filename which located on some webserver it start downloading it and after 42% it give a msg no disk space available and it get stopped. although i check on sever lot more

Re: IPv6 support of wget v 1.9.1

2003-12-31 Thread Tony Lewis
Kazu Yamamoto wrote: Since I have experiences to modify IPv4 only programs, including FTP and HTTP, to IPv6-IPv4 one, I know this problem. Yes, some part of wget *would* remain protocol dependent. Kazu, it's been said that a picture is worth a thousand words. Perhaps in this case, a patch

Re: passing a login and password

2004-01-06 Thread Tony Lewis
robi sen wrote: Hi I have a client who basically needs to regularly grab content from part of their website and mirror or it and or save it so they can disseminate it as HTML on a CD. The website though is written in ColdFusion as requires application level authentication which is just form

Re: Startup delay on Windows

2004-02-08 Thread Tony Lewis
running older OS versions to confirm that.) For example, on my Windows XP machine, I have to following variables: HOMEDRIVE=C: HOMEPATH=\Documents and Settings\Tony Lewis so my home directory is C:\Documents and Settings\Tony Lewis HTH, Tony

Re: not downloading at all, help

2004-02-12 Thread Tony Lewis
Juhana Sadeharju wrote: I placed use_proxy = off to .wgetrc (which file I did not have earlier) and to ~/wget/etc/wgetrc (which file I had), and tried wget --proxy=off http://www.maqamworld.com and it still does not work. Could there be some system wgetrc files somewhere? I have compiled

Re: Problem Accessing FTP Site Where Password Contains @

2004-03-09 Thread Tony Lewis
[EMAIL PROTECTED] wrote: I came across a problem accessing an FTP site where the password contained a @ sign. The password was [EMAIL PROTECTED] So I tried the following: wget -np --server-response -H --tries=1 -c --wait=60 --retry-connrefused -R * ftp://guest:[EMAIL

Re: trouble with encoded filename

2004-04-08 Thread Tony Lewis
[EMAIL PROTECTED] wrote: Well, I found out a little bit more about the real reason for the problem. Opera has a very convenient option called Encode International Web Addresses with UTF-8. When I had this option checked, it could retrieve the file without problems. Without this option

Re: OpenVMS URL

2004-05-26 Thread Tony Lewis
How do you enter the path in your web browser? - Original Message - From: Bufford, Benjamin (AGRE) [EMAIL PROTECTED] To: [EMAIL PROTECTED] Sent: Wednesday, May 26, 2004 7:32 AM Subject: OpenVMS URL I am trying to use wget to retrieve a file from an OpenVMS server but have been unable

Re: OpenVMS URL

2004-05-26 Thread Tony Lewis
Then your problem isn't with wget. Once you figure out how to access the file in a web browser, use the same URL in wget. Tony - Original Message - From: Bufford, Benjamin (AGRE) [EMAIL PROTECTED] To: Tony Lewis [EMAIL PROTECTED]; [EMAIL PROTECTED] Sent: Wednesday, May 26, 2004 8:41 AM

Re: OpenVMS URL

2004-05-27 Thread Tony Lewis
Hrvoje Niksic wrote: Wget could always support a URL parameter, such as: wget 'ftp://server/dir1/dir2/file;disk=foo' Assuming, you can detect a VMS connection, why not simply ftp://server/foo:[dir1.dir2]? Tony

Re: file name problem

2004-06-01 Thread Tony Lewis
henry luo wrote: i find a problem at GNU Wget 1.9.1, but i dont know it is a new function or a bug; the old version(1.8.2) download a link ,for example: wget 'http://www.expekt.com/odds/eventsodds.jsp?range=100sortby=dateactive= bettingbetcategoryId=SOC%25'

Re: Escaping semicolons

2004-06-24 Thread Tony Lewis
Phil Endecott wrote: I am using wget to build a downloadable zip file for offline viewing of a CGI-intensive web site that I am building. Essentially it works, but I am encountering difficulties with semicolons. I use semicolons in CGI URIs to separate parameters. (Ampersand is more often

Re: Escaping semicolons

2004-06-27 Thread Tony Lewis
, wget encodes it to create a valid name. Tony Lewis wrote: I use semicolons in CGI URIs to separate parameters. (Ampersand is more often used for this, but semicolon is also allowed and has the advantage that there is no need to escape it in HTML.) There is no need to escape ampersands

Re: Escaping semicolons (actually Ampersands)

2004-06-28 Thread Tony Lewis
Phil Endecott wrote: Tony The stuff between the quotes following HREF is not HTML; it Tony is a URL. Hence, it must follow URL rules not HTML rules. No, it's both a URL and HTML. It must follow both rules. Please see the page that I cited in my previous message:

Re: question on wget via http proxy

2004-07-12 Thread Tony Lewis
Malte Schünemann wrote: Since wget is able to obtain directoy listings / retrieve data from there is should be possible to also upload data Then it would be wput. :-) What is so special about wget that it is able to perform this task? You can learn a LOT about how wget is communicating with

Re: retrieve a whole website with image embedded in css

2004-07-13 Thread Tony Lewis
Ploc wrote: The result is a website very different from the original one as it lacks backgrounds. Can you please confirm if what I think is true (or not), if it is registered as a bug, and if there is a date planning to correct it. It is true. wget only retrieves objects that appear in the

Re: retrieve a whole website with image embedded in css

2004-07-13 Thread Tony Lewis
Ploc wrote: Is it already registered as a bug or in a whishlist ? It's not a bug. This feature has been on the wishlist for a long time. Tony

Re: Stratus VOS support

2004-07-28 Thread Tony Lewis
Stratus VOS supportJonathan Grubb wrote: Any thoughts of adding support for Stratus VOS file structures? Your question is a little too vague -- even for me (I used to work for Stratus and actually know what VOS is :-) What file structures are you needing supported that wget does not currently

Re: Stratus VOS support

2004-07-28 Thread Tony Lewis
Jonathan Grubb wrote: Um. I'm using wget on Win2000 to ftp to a VOS machine. I'm finding that the usual '' sign for directories isn't supported by wget and that '/' doesn't work either, I think because the ftp server itself is expecting ''. The problem may be that Win 2000 grabs the before

Re: wput mailing list

2004-08-29 Thread Tony Lewis
Justin Gombos wrote: Since I feel that computers serve man, not the reverse, so I don't intend to change my file organization to be web page centric. Looking around the web, I was quite surprized to find that I'm the only one with this problem. I was very relieved to find that there was a wput -

RE: date based retrieval

2004-12-19 Thread Tony Lewis
Anthony Caetano wrote: I am looking for a way to stay current without mirroring an entire site. [snip] Does anyone else see a use for this? Yes. Here's my non-wget solution. I truncate all the files in the directories that I don't want, but maintain the date/time accessed and modified. The

RE: Metric units

2004-12-23 Thread Tony Lewis
John J Foerch wrote: It seems that the system of using the metric prefixes for numbers 2^n is a simple accident of history. Any thoughts on this? I would say that the practice of using powers of 10 for K and M is a response to people who cannot think in binary. Tony

RE: Metric units

2004-12-23 Thread Tony Lewis
Carlos Villegas snidely wrote: I would say that the original poster understands what he is saying, and you clearly don't... I'll put my computer science degree up against your business administration and accounting degree any day. A kilobyte has always been 1024 bytes and the choice was not

RE: Metric units

2004-12-23 Thread Tony Lewis
Mark Post wrote: While we're at it, why don't we just round off the value of pi to be 3.0 Do you live in Indiana? Actually, Dr. Edwin Goodwin wanted to round off pi to any of several values including 3.2. http://www.agecon.purdue.edu/crd/Localgov/Second%20Level%20pages/Indiana_Pi_ Story.htm

RE: new string module

2005-01-05 Thread Tony Lewis
Mauro Tortonesi wrote: Alle 18:28, mercoled 5 gennaio 2005, Draen Kacar ha scritto: Jan Minar wrote: What's wrong with mbrtowc(3) and friends? The mysterious solution is probably to use wprintf(3) instead printf(3). Couple of questions on #c on freenode would give you that answer.

RE: wget: question about tag

2005-02-02 Thread Tony Lewis
Normand Savard wrote: I have a question about wget. Is is possible to download other attribute value other than the harcoded ones? No, at least not in the existing versions of wget. I have not heard that anyone is working on such an enhancement.

RE: one bug?

2005-03-04 Thread Tony Lewis
Jesus Legido wrote: I'm getting a file from https://mfi-assets.ecb.int/dla/EA/ea_all_050303.txt: The problem is not with wget. The file on the server starts with 0xFF 0xFE. Put the following into an HTML file (say temp.html) on your hard drive, open it in your web browser, right click on

RE: Curb maximum size of headers

2005-03-17 Thread Tony Lewis
Hrvoje Niksic wrote: I don't see how and why a web site would generate headers (not bodies, to be sure) larger than 64k. To be honest, I'm less concerned about the 64K header limit than I am about limiting a header line to 4096 bytes. I don't know any sites that send back header lines that

RE: help!!!

2005-03-21 Thread Tony Lewis
The --post-data option was added in version 1.9. You need to upgrade your version of wget. Tony -Original Message- From: Richard Emanilov [mailto:[EMAIL PROTECTED] Sent: Monday, March 21, 2005 8:49 AM To: Tony Lewis; [EMAIL PROTECTED] Cc: wget@sunsite.dk Subject: RE: help!!! wget

RE: File rejection is not working

2005-04-06 Thread Tony Lewis
Jens Rösner wrote: AFAIK, RegExp for (HTML?) file rejection was requested a few times, but is not implemented at the moment. It seems all the examples people are sending are just attempting to get a match that is not case sensitive. A switch to ignore case in the file name match would be a

RE: newbie question

2005-04-14 Thread Tony Lewis
Alan Thomas wrote: I am having trouble getting the files I want using a wildcard specifier... There are no options on the command line for what you're attempting to do. Neither wget nor the server you're contacting understand *.pdf in a URI. In the case of wget, it is designed to read web

RE: SSL options

2005-04-21 Thread Tony Lewis
Hrvoje Niksic wrote: The question is what should we do for 1.10? Document the unreadable names and cryptic values, and have to support them until eternity? My vote is to change them to more reasonable syntax (as you suggested earlier in the note) for 1.10 and include the new syntax in the

RE: links conversion; non-existent index.html

2005-05-01 Thread Tony Lewis
Andrzej wrote: Two problems: There is no index.html under this link: http://znik.wbc.lublin.pl/Mineraly/Ftp/UpLoad/ [snip] it creates a non existing link: http://znik.wbc.lublin.pl/Mineraly/Ftp/UpLoad/index.html When you specify a directory, it is up to the web server to determine what

RE: Is it just that the -m (mirror) option an impossible task [Was: wget 1.91 skips most files]

2005-05-28 Thread Tony Lewis
Maurice Volaski wrote: wget's -m option seems to be able to ignore most of the files it should download from a site. Is this simply because wget can download only the files it can see? That is, if the web server's directory indexing option is off and a page on the site is present on the

RE: Removing thousand separators from file size output

2005-06-24 Thread Tony Lewis
Hrvoje Niksic wrote: In fact, I know of no application that accepts numbers as Wget prints them. Microsoft Calculator does. Tony

Name or service not known error

2005-06-27 Thread Tony Lewis
I got a "Name or service not known" error from wget 1.10 running on Linux. When I installed an earlier version of wget, it worked just fine.It also works just fine on version 1.10 running on Windows. Any ideas? Here's the output on Linux: wget --versionGNU Wget 1.9-beta1 wget

RE: Invalid directory names created by wget

2005-07-08 Thread Tony Lewis
Larry Jones wrote: Of course it's directly accessible -- you just have to quote it to keep the shell from processing the parentheses: cd 'title.Die-Struck+(Gold+on+Gold)+Lapel+Pins' You can also make the individual characters into literals: cd

RE: wget a file with long path on Windows XP

2005-07-21 Thread Tony Lewis
PoWah Wong wrote: The login page is: http://safari.informit.com/?FPI=uicode= How to figure out the login command? These two commands do not work: wget --save-cookies cookies.txt http://safari.informit.com/?FPI= [snip] wget --save-cookies cookies.txt

RE: connect to server/request multiple pages

2005-07-21 Thread Tony Lewis
Pat Malatack wrote: is there a way to stay connected, because it seems to me that this takes a decent amount of time that could be minimized The following command will do what you want: wget "google.com/news" "google.com/froogle" Tony

RE: Wget patches for .files

2005-08-19 Thread Tony Lewis
Mauro Tortonesi wrote: this is a very interesting point, but the patch you mentioned above uses the LIST -a FTP command, which AFAIK is not supported by all FTP servers. As I recall, that's why the patch was not accepted. However, it would be useful if there were some command line option to

RE: wget can't handle large files

2005-10-18 Thread Tony Lewis
Eberhard Wolff wrote: Apparently wget can't handle large file. [snip] wget --version GNU Wget 1.8.2 This bug was fixed in version 1.10 of wget. You should obtain a copy of the latest version, 1.10.2. Tony

RE: Error connecting to target server

2005-11-11 Thread Tony Lewis
[EMAIL PROTECTED] wrote: Thanks for your reply. Only ping works for bbc.com and not wget. When I issue the command wget www.bbc.com, it successfully downloads the following file: !DOCTYPE HTML PUBLIC -//W3C//DTD HTML 3.2//EN HTML HEAD META HTTP-EQUIV=Refresh content=0;

RE: spaces in pathnames using --directory-prefix=prefix

2005-11-30 Thread Tony Lewis
Jonathan DeGumbia wrote: I'm trying to use the --directory-prefix=prefix option for wget on a Windows system. My prefix has spaces in the path directories. Wget appears to terminate the path at the first space encountered. In other words if my prefix is: c:/my prefix/ then wget copies

RE: wget 1.10.x fixed recursive ftp download over proxy

2006-01-09 Thread Tony Lewis
I believe the following simplified code would have the same effect: if ((opt.recursive || opt.page_requisites || opt.use_proxy) url_scheme (*t) != SCHEME_FTP) status = retrieve_tree (*t);else status = retrieve_url (*t, filename, redirected_URL, NULL, dt); Tony From: [EMAIL PROTECTED]

RE: wget option (idea for recursive ftp/globbing)

2006-03-02 Thread Tony Lewis
Mauro Tortonesi wrote: i would like to read other users' opinion before deciding which course of action to take, though. Other users have suggested adding a command line option for -a two or three times in the past: - 2002-11-24: Steve Friedl [EMAIL PROTECTED] submitted a patch - 2002-12-24:

RE: Bug in ETA code on x64

2006-03-28 Thread Tony Lewis
Hrvoje Niksic wrote: The cast to int looks like someone was trying to remove a warning and botched operator precedence in the process. I can't see any good reason to use , here. Why not write the line as: eta_hrs = eta / 3600; eta %= 3600; This makes it much less likely that someone

RE: regex support RFC

2006-03-30 Thread Tony Lewis
How many keywords do we need to provide maximum flexibility on the components of the URI? (I'm thinking we need five.) Consider http://www.example.com/path/to/script.cgi?foo=bar --filter=uri:regex could match against any part of the URI --filter=domain:regex could match against www.example.com

  1   2   >