Peter Rundle <[EMAIL PROTECTED]> writes: > I'm looking for some recommendations for a *simple* Linux based tool > to spider a web site and pull the content back into plain html files, > images, js, css etc.
Others have suggested wget, which works very well. You might also consider 'puf': Package: puf Priority: optional Section: universe/web Description: Parallel URL fetcher puf is a download tool for UNIX-like systems. You may use it to download single files or to mirror entire servers. It is similar to GNU wget (and has a partly compatible command line), but has the ability to do many downloads in parallel. This is very interesting, if you have a high-bandwidth internet connection. This works quite well when, as it notes, presented with sufficient bandwidth (and server resources) to have multiple links fetched at once. Regards, Daniel -- SLUG - Sydney Linux User's Group Mailing List - http://slug.org.au/ Subscription info and FAQs: http://slug.org.au/faq/mailinglists.html