Not exactly the same thing, but there is a set of grabber scripts designed to get and import an entire wiki via the api, but these are currently mostly just mw maintenance and a few python scripts that only expect to be used with an actual second mediawiki instance.

That being said, it probably would be a good idea to rearchitect them to be a bit less redundant with each other, and in so doing also make their usage a tad more flexible anyway, such as for creating dumps and backups instead. Whatever the case, at very least they might make a useful reference point.

See: https://www.mediawiki.org/wiki/Manual:Grabbers

-I

On 15/03/2019 18:43, John wrote:
Unfortunately that doesn’t really work unless it’s a fairly small wiki. If
it’s bigger the export request times out.

What I was hoping for was a tool that utilizes either special export or the
API to build a dump file using multiple requests.

I can probably write something to do the same thing, but was hoping for an
existing solution

On Fri, Mar 15, 2019 at 1:12 PM Sterkebak <[email protected]> wrote:

XML export if enabled should do the trick.

On Fri, 15 Mar 2019 at 18:04, John <[email protected]> wrote:

Are there any tools that allow you to create a dump of a mediawiki
install
that doesn’t require direct database access? My primary focus is on
creating a backup of the wiki contents.
Thanks
_______________________________________________
MediaWiki-l mailing list
To unsubscribe, go to:
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l

--
Met vriendelijke groet,

Huib Laurens
_______________________________________________
MediaWiki-l mailing list
To unsubscribe, go to:
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l

_______________________________________________
MediaWiki-l mailing list
To unsubscribe, go to:
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l



_______________________________________________
MediaWiki-l mailing list
To unsubscribe, go to:
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l

Reply via email to