Ack! Fingers slipped! Here's what I meant to write: > I did something like this once, to aggregate a list of html > pages together. Each page had a [next] link at the bottom, > and my pipeline basically followed these links. I think you > want something similar. I don't have the code still, but from > memory, something like this: >
<map:match pattern="page/*"> <map:generate type="html" src="http://www.foo.com/?page="{1}"/> <map:transform src="check-if-found-or-keep-looking.xsl"> <map:parameter name="current-page" value="{1}"/> </map:transform> <map:transform type="xinclude"/> <map:serialize type="xml"/> </map:match> Call the pipeline first with "page/1". The XSL check-if-found-or-keep-looking.xsl should check if the record is in the page: if so it should return the contents of the page (i.e. an identity transform). If not found, it should return only an xinclude statement. So the entire output document would be just the single element: <xi:include xmlns:xi="I-don't-remember-the-xinclude-namespace-uri :-)" href="cocoon:/page/{$current-page+1}"/> Then the xinclude transformer handles the rest. If the record was found then the xinclude transformer does nothing, and the page is serialized. If the record was not found then the xinclude transformer will make the recursive sitemap call. Cheers Con --------------------------------------------------------------------- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED]