Hi Brian,
maybe this helps:
--html-extension
If a file of type application/xhtml+xml or text/html is downloaded and
the URL does not end with the regexp \.[Hh][Tt][Mm][Ll]?, this option
will cause the suffix .html to be appended to the local filename. This
is useful, for instance, when you're mirroring a remote site that uses
.asp pages, but you want the mirrored pages to be viewable on your stock
Apache server. Another good use for this is when you're downloading
CGI-generated materials. A URL like http://site.com/article.cgi?25 will
be saved as article.cgi?25.html.
At least to me it seems that wget than should download everything. Not
though that it will redownload all kinds of mangeled URLs (like this
one) when wget is told to redownload the file or when wget reencounters
it as link.
Else you could append a ? to the URL which should be stripped on the
server side anyway.
Hope that helps
Matthias
Brian Keck schrieb:
Hello,
If you do
wget http://www.ifixit.com/Guide/First-Look/iPhone3G
then you get an HTML file called iPhone3G.
But if you do
wget -p http://www.ifixit.com/Guide/First-Look/iPhone3G
then you get a directory called iPhone3G.
This makes sense if you look at the links in the HTML file, like
/Guide/First-Look/iPhone3G/images/3jYKHyIVrAHnG4Br-standard.jpg
But of course I want both. Is there a way of getting wget -p to do
something clever, like renaming the HTML file? I've looked through
wget(1) & /usr/share/doc/wget & the comments in the 1.10.2 source
without seeing anything relevant.
Thanks,
Brian Keck