from man wget:
In one case you'll need to add a couple more options.
If document is a "<FRAMESET>" page, the "one more hop"
that -p gives you won't be enough---you'll get the
"<FRAME>" pages that are referenced, but you won't get
their requisites. Therefore, in this case you'll need
to add -r -l1 to the commandline. The -r -l1 will
recurse from the "<FRAMESET>" page to to the "<FRAME>"
pages, and the -p will get their requisites. If
you're already using a recursion level of 1 or more,
you'll need to up it by one. In the future, -p may be
made smarter so that it'll do "two more hops" in the
case of a "<FRAMESET>" page.
wget handles frames just fine.
Seth
----- Original Message -----
From: Ralph Zeller <[EMAIL PROTECTED]>
To: <[EMAIL PROTECTED]>
Sent: Tuesday, October 02, 2001 11:37 AM
Subject: [EUG-LUG:3047] Recursive web-sucking
> I sometimes like to suck down an entire web-site using wget. However,
> wget doesn't follow links in frames. Is there another tool that can
> recursively web-suck including following links in frames? Thanks.
>
>