You're looking for:
  --page-requisites    get all images, etc. needed to display HTML page.

wget URL --page-requisites

should give you what you need.
-----Original Message-----
From: [email protected]
[mailto:[email protected]] On Behalf Of Garry
Sent: Thursday, April 19, 2012 2:45 AM
To: [email protected]
Subject: [Bug-wget] Feature question ...

Hi,

not exactly a bug question/report, but something I was trying to get done
with wget but have either overlooked in the docs, misread or it's not plain
not possible at the moment ...

I'm trying to mirror a full web page, but with some restrictions ... I need
a single page (either the full path, or - if it's the main page in a
directory - just that index page) to be downloaded, with all contained media
(at least images, css, js-includes etc.), even if that media/files are not
stored on that server. As I need the information for archival purposes, I do
not want a full tree of directories rebuilt, as wget would normally do. All
the files should be downloaded and stored in some unique file names in the
same directory as the page file, and of course the html page should be
re-coded as to use the relative path to those renamed files.

Can this be done with wget? Or if not, is there a different program
(Linux) that will do this?

Tnx, Garry


Reply via email to