I know I have a similar need. But wget doesn't work for me, because the
links I need to follow are CGI, and wget doesn't follow cgi. I have
tried several things, but so far, I've not had any luck getting things
to work.

-- 
<URL: http://wiki.tcl.tk/ > Indescribable,uncontainable,all
powerful,untameable
Even if explicitly stated to the contrary, nothing in this posting
should be construed as representing my employer's opinions.
<URL: mailto:[EMAIL PROTECTED] > <URL: http://www.purl.org/NET/lvirden/
>
 


-----Original Message-----
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Konstantin
Khomoutov
Sent: Wednesday, March 29, 2006 10:30 AM
To: [email protected]
Subject: Help! How capture the HTLM pages on the PC?


On Wed, Mar 29, 2006 at 03:56:30PM +0200, Enzo Telatin wrote:

> Some body can help me?
> I would like to capture some HTML pages (with there links) from the 
> Web on the PC and thereby edit/prepare/format this pages with an HTML 
> editor for pluker desktop? A lot of space is lost if you capture 
> directly this pages with the pluker desktop and is not so easy to read

> this pages on the PDA.
I'm not quite sure about what *exactly* do you mean saying "lot of
space", so let me guess.

1) Have you played with the "maximum depth" parameter in your Plucker
spider? It's possibly that it's set too high and you just suck too many
linked pages. Try to set it to 2 (means "this page and all pages that it
has links to").

2) wget is your best friend when fetching sites. On Unices you have to
get it either installed or available. Good ports to Win32 exist.
google.com?q=wget+win32 To get you up quickly, this encantation usually
works: $ wget -nd -nc -np -nH -k -r -l 2 http://www.site.com/root.html
which means:
* don't create URL-path-named intermediate directories;
* don't clobber existing files (on multiple runs);
* don't go up in the URL-path if a link leads there;
* don't go to another hosts;
* convert links in fetched files so that they reference each other;
* go recursively...
* ...with level 2 (this page and all pages it references).

Hope it helps.

_______________________________________________
plucker-list mailing list
[email protected]
http://lists.rubberchicken.org/mailman/listinfo/plucker-list
_______________________________________________
plucker-list mailing list
[email protected]
http://lists.rubberchicken.org/mailman/listinfo/plucker-list

Reply via email to