On 9/27/2010 2:58 PM, Chad wrote:
> This. Tim sums up the consensus very well with that commit summary.
> He also made some comments on the history of wikitext and alternative
> parsers on foundation-l back in Jan '09[0]. Worth a read (starting mainly
> at ""Parser" is a convenient and short name for it").
>
> While a real parser is a nice pipe dream, in practice not a single project
> to "rewrite the parser" has succeeded in the years of people trying. Like
> Aryeh says, if you can pull it off and make it practical, hats off to you.
>
For my own IX work I've written a wikimedia markup parser in C#
based on the Irony framework. It fails to parse about 0.5% of pages in
wikipedia and is oblivious to a lot of the stranger stuff [like the HTML
intrusions] but it does a good job of eating infoboxes and making sense
of internal and external links. Now, the strange stuff + the parse
fails would probably be impossible to handle in a rational way...
_______________________________________________
Wikitech-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikitech-l