2010-09-27 22:46, Paul Houle skrev:
>    On 9/27/2010 2:58 PM, Chad wrote:
>    
>> This. Tim sums up the consensus very well with that commit summary.
>> He also made some comments on the history of wikitext and alternative
>> parsers on foundation-l back in Jan '09[0]. Worth a read (starting mainly
>> at ""Parser" is a convenient and short name for it").
>>
>> While a real parser is a nice pipe dream, in practice not a single project
>> to "rewrite the parser" has succeeded in the years of people  trying. Like
>> Aryeh says, if you can pull it off and make it practical, hats off to you.
>>
>>      
>       For my own IX work I've written a wikimedia markup parser in C#
> based on the Irony framework.  It fails to parse about 0.5% of pages in
> wikipedia

What do you mean with "fail".  It assigns slightly incorrect semantic to a
construction? It fails to accept the input?  It crashes?

>   and is oblivious to a lot of the stranger stuff [like the HTML
> intrusions] but it does a good job of eating infoboxes and making sense
> of internal and external links.  Now,  the strange stuff + the parse
> fails would probably be impossible to handle in a rational way...
>    

I disagree.  I believe that there is a rational way to handle all kinds 
of input.

/Andreas


_______________________________________________
Wikitech-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Reply via email to