Thanks Arlo for replying.

Can you please give me some referenc elink to the native parser pf
Mediawiki that you have suggested? A native parser will always be the
easiest way to cater our need. We want to convert the pages pf our
Mediawiki (1.17.5) to either pdf or html pages . All the
attachments (images), comments should also come as part of the output file.

Thanks
Ayaskant

On Fri, Feb 7, 2020 at 9:55 PM Arlo Breault <[email protected]> wrote:

>
> > On Feb 6, 2020, at 4:35 AM, Ayaskant Swain <[email protected]>
> wrote:
> >
> > Hi Parsoid Support Team,
> >
> > I am reaching out to you to know about the usage of this tool. We have a
> very old version 1.17.5 of Mediawiki in our organization and want to
> convert the pages of it to html pages and store it on disk for archiving.
> As you know internally Mediawiki stores pages as WikiText.
> >
> > Can parsoid (https://www.mediawiki.org/wiki/Parsoid) help us here?
>
> Maybe?  It's very likely that Parsoid will have some compatibility
> issues that you'll need to hack around.
>
>
> > I also saw the documentation of VisualEditor extension (
> https://www.mediawiki.org/wiki/VisualEditor) which uses parsoid
> internally to convert wikitext pages. Which tool among these 2 should we
> use to do my job? Can you please suggest?
>
> Parsoid is not included in VE, it just offers an API for VE to query.
> Adding VE to the mix is an unnecessary complication.
>
>
> > can parsoid be used as a standalone application or tool instead of VE?
>
> Yes
>
>
> > If we use any of them do we need to just provide the url of our
> Mediawiki page (example - https://<our_dns_host>/wiki/TestPage or do we
> need to extract the content from DB which is in WikiText format and feed it
> to parsoid for converting it to html page?
>
> Parsoid has traditionally interacted with MediaWiki's action API
> (the thing at /api.php).  You would not need to do any manual extraction.
>
>
> There seems to be an active project similar to what you're describing
> at, https://github.com/openzim/mwoffliner
>
> However, it might be less complicated to just use the parser
> that ships with the MediaWiki version you're running.  In other
> words, screen scrape the pages MediaWiki is already serving you.
>
>
> > Thanks
> > _______________________________________________
> > Wikitext-l mailing list
> > [email protected]
> > https://lists.wikimedia.org/mailman/listinfo/wikitext-l
>
>
_______________________________________________
Wikitext-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikitext-l

Reply via email to