https://bugzilla.wikimedia.org/show_bug.cgi?id=62468
Bug ID: 62468
Summary: Add option to have the Internet Archiver (and/or other
robots) retrieve raw wikitext of all pages
Product: MediaWiki
Version: 1.23-git
Hardware: All
OS: All
Status: NEW
Severity: enhancement
Priority: Unprioritized
Component: General/Unknown
Assignee: [email protected]
Reporter: [email protected]
Web browser: ---
Mobile Platform: ---
I propose to add an option to have the Internet Archiver (and/or other robots)
retrieve raw wikitext. This way, if a wiki goes down, it will be possible to
more easily create a successor wiki by gathering that data from the Internet
Archive. As it is now, all that can be obtained are the parsed pages. That's
okay for a static archive, but one might want to revive the wiki for further
editing. Also, Archive should be opened
I would say that someone should write a script to convert the parsed pages
retrieved from the Internet Archive back into wikitext, but that will run into
problems with templates and such, unless it's designed to identify them and
recreate them. It would be a much easier and cleaner solution to just make the
wikitext available from the get-go.
--
You are receiving this mail because:
You are the assignee for the bug.
You are on the CC list for the bug.
_______________________________________________
Wikibugs-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikibugs-l