You can use http://meritbadge.org/wiki/index.php/Special:Export, just add *all* pages (API call/Special:AllPages/similar solution) to the textbox, uncheck "Include only the current revision, not full history" to have full history, save the file prepared. Then you should buy a new server, put the dump on the server and call https://www.mediawiki.org/wiki/Manual:ImportDump.php. You will have new wiki with old content. If you want to go for hosting, you should use Special:Import as the reverse solution.
Hope that helps. Best, Martin pá 18. 5. 2018 v 20:27 odesílatel Bart Humphries <bart.humphr...@gmail.com> napsal: > We have a very old wiki which has basically never been updated for the past > decade and which was proving stubbornly resistant to updating several years > ago. And now the owner of the server has drifted away, but we do still > have control over the domain name itself. The best way that we can think > of to update everything is to scrape all of the pages/file, add them to a > brand new updated wiki on a new server, then point the domain to that new > server. Yes, user accounts will be broken, but we feel that this is the > most feasible solution unless someone else has another idea. > > However, there's a lot of pages on meritbadge.org -- which is the wiki I'm > talking about. Any suggestions for how to automate this scraping process? > I can scrape the HTML off every page, but what I really want is to get the > wikitext off of every page. > > Bart Humphries > bart.humphr...@gmail.com > (909)529-BART(2278) > _______________________________________________ > Wikitech-l mailing list > Wikitech-l@lists.wikimedia.org > https://lists.wikimedia.org/mailman/listinfo/wikitech-l _______________________________________________ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l