On 18 February 2014 20:41, Antoine Musso <[email protected]> wrote:

> Le 18/02/2014 08:18, Philip Neustrom a écrit :
> > The last details on their technical infrastructure indicated that Snowden
> > used "web crawler" (love the quotes) software to obtain information from
> > their internal wiki:
> >
> >
> http://www.nytimes.com/2014/02/09/us/snowden-used-low-cost-tool-to-best-nsa.html?hp
>
> Hello,
>
> From a usability point of view, I am wondering why he had to rely on a
> web crawler to export the whole Wiki as HTML.  For those wondering, you
> could use:
>
>  https://www.mediawiki.org/wiki/Extension:DumpHTML  maintenance script
> which generate a HTML version of your wiki for archival purposes
>
>  https://www.mediawiki.org/wiki/Extension:Pdf_Book to assemble articles
> in a single document export as PDF.
>
>
> MediaWiki has Special:Export, but it only export Wikitext.
>
> cheers,
>

Because all the articles are just speculation?
_______________________________________________
Wikitech-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Reply via email to