2009/9/25 Dmitriy Sintsov <ques...@rambler.ru>:

> XML is used to store "human-produced" rich formatted text by many
> standalone and web apps. XML parsers are also very strict and spitting
> errors. As it's been mentioned recently, XML is really good for bots,
> too, for that reason (the input is "error-free" tree).
> I wonder, If the browsers can handle tag soup, most probably MediaWiki
> parser can handle wikitext soup, too? Eg, instead of parsing error,
> properly close the nodes. The existing wikitext of millions of articles
> has to be converted by commandline upgrade script in case the wikitext
> will be abandoned. Though I wonder whether it's possible to keep
> wikitext editing mode for backwards compatibility by using the same
> method online.


Wikitext started as a shorthand for HTML. The horrible things that
have happened since then are from ad-hoc additions to a parser and no
formal spec, leaving the parser behaviour as, literally, the only
definition.


- d.

_______________________________________________
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Reply via email to