On Thu, Feb 06, 2003 at 06:07:02PM +0100, Andrea Censi wrote:
> Which is the best tool to create an offline copy of a cocoon live site?
> 
> It should spider through the pages, follow every link and gather both html and 
> images/css/pdf. Then it should rearrange internal links from absolute to 
> relative ("http://site/page"; -> page.html,  "/" -> "index.html").
> The result is to be loaded on a low-spec [ = no cocoon :( ] webserver.
> 
> I don't consider the batch use from command line to be a viable alternative, 
> because:
>  - There are different views of single xml files.

Don't the different views have different URLs?  If so, and if you link to
those different URLs, then they will each have a file written.

>  - I don't want to explicitly change the internal URL format used by the site 
> (/ ... /page/ with a final slash)

I think the crawler will convert links to a directory, eg 'foo/' to
'foo/index.html'.

>  - (not sure) Would it work with dynamic SVG->gif?

Yes. I'd say, give it a try.  Works fine rendering Forrest sites.
Alternatively, you could try spidering tools like 'wget'.

--Jeff


---------------------------------------------------------------------
Please check that your question  has not already been answered in the
FAQ before posting.     <http://xml.apache.org/cocoon/faq/index.html>

To unsubscribe, e-mail:     <[EMAIL PROTECTED]>
For additional commands, e-mail:   <[EMAIL PROTECTED]>

Reply via email to