Re: Save request headers
On Mon, Mar 17, 2008 at 10:24 PM, Julian Burgess [EMAIL PROTECTED] wrote: Hello, is there any way in wget I can save the request headers to the file, at the moment I'm downloading a fairly large number of files and sometimes it would be really helpful to have the original url from which it was downloaded. Thanks C:\wget --help | grep header ... --save-headers save the HTTP headers to file. ... C:\ wget --save-headers --proxy=off http://localhost/ --2008-03-17 22:29:49-- http://localhost/ ... Saving to: `index.html' The contents of index.html: HTTP/1.1 200 OK Date: Mon, 17 Mar 2008 15:29:49 GMT Server: Apache/2.2.8 (Win32) mod_view/2.2 mod_python/3.3.1 Python/2.5.1 Connection: close Content-Type: text/html !DOCTYPE html PUBLIC -//W3C//DTD XHTML 1.0 Transitional//EN DTD/xhtml1-transitional.dtd htmlhead ... So there is a way, but I am not sure that it is the way you want :-). Maybe a better way is to run wget in the background so that it produce a wget-log that can be used to trace the URLs or 'tee' the output of wget to a file. --- Charles
Re: Save request headers
On Mon, Mar 17, 2008 at 10:24 PM, Julian Burgess [EMAIL PROTECTED] wrote: Hello, is there any way in wget I can save the request headers to the file, at the moment I'm downloading a fairly large number of files and sometimes it would be really helpful to have the original url from which it was downloaded. Thanks C:\wget --help | grep header ... --save-headers save the HTTP headers to file. ... C:\ wget --save-headers --proxy=off http://localhost/ --2008-03-17 22:29:49-- http://localhost/ ... Saving to: `index.html' The contents of index.html: HTTP/1.1 200 OK Date: Mon, 17 Mar 2008 15:29:49 GMT Server: Apache/2.2.8 (Win32) mod_view/2.2 mod_python/3.3.1 Python/2.5.1 Connection: close Content-Type: text/html !DOCTYPE html PUBLIC -//W3C//DTD XHTML 1.0 Transitional//EN DTD/xhtml1-transitional.dtd htmlhead ... So there is a way, but I am not sure that it is the way you want :-). Maybe a better way is to run wget in the background so that it produce a wget-log that can be used to trace the URLs or 'tee' the output of wget to a file. --- Charles
Re: Save request headers
--save-header only seems to save the headers returned by the webserver, I need to save the http header I originally sent, or some other way which will show the full url of the page I downloaded. On 17/03/2008, Charles [EMAIL PROTECTED] wrote: On Mon, Mar 17, 2008 at 10:24 PM, Julian Burgess [EMAIL PROTECTED] wrote: Hello, is there any way in wget I can save the request headers to the file, at the moment I'm downloading a fairly large number of files and sometimes it would be really helpful to have the original url from which it was downloaded. Thanks C:\wget --help | grep header ... --save-headers save the HTTP headers to file. ... C:\ wget --save-headers --proxy=off http://localhost/ --2008-03-17 22:29:49-- http://localhost/ ... Saving to: `index.html' The contents of index.html: HTTP/1.1 200 OK Date: Mon, 17 Mar 2008 15:29:49 GMT Server: Apache/2.2.8 (Win32) mod_view/2.2 mod_python/3.3.1 Python/2.5.1 Connection: close Content-Type: text/html !DOCTYPE html PUBLIC -//W3C//DTD XHTML 1.0 Transitional//EN DTD/xhtml1-transitional.dtd htmlhead ... So there is a way, but I am not sure that it is the way you want :-). Maybe a better way is to run wget in the background so that it produce a wget-log that can be used to trace the URLs or 'tee' the output of wget to a file. --- Charles