> On Feb 8, 2020, at 1:07 AM, Ayaskant Swain <[email protected]> wrote: > > Thanks Arlo for replying. > > Can you please give me some referenc elink to the native parser pf Mediawiki > that you have suggested? A native parser will always be the easiest way to > cater our need. We want to convert the pages pf our Mediawiki (1.17.5) to > either pdf or html pages . All the attachments (images), comments should also > come as part of the output file.
When you visit https://<host>/wiki/TestPage, MediaWiki has already parsed the content to HTML for you. I was suggesting you scrape those pages using wget, Scrapy, HTTrack, or some other tool. It's also possible this extension works for you, https://www.mediawiki.org/wiki/Extension:DumpHTML _______________________________________________ Wikitext-l mailing list [email protected] https://lists.wikimedia.org/mailman/listinfo/wikitext-l
