On Tue, Apr 5, 2011 at 7:45 AM, Ashar Voultoiz <[email protected]> wrote:
> On 05/04/11 04:47, Tim Starling wrote:
>  > Speaking of "fast", I did a quick benchmark of the [[Barack Obama]]
>  > article with templates pre-expanded. It took 22 seconds in HipHop and
>  > 112 seconds in Zend, which is not bad, for a first attempt. I reckon
>  > it would do better if a few of the regular expressions were replaced
>  > with tight loops.
> <snip>
> I have imported in my local wiki the english [[Barack Obama]] article
> with all its dependencies. I can not have it parsed under either 256MB
> max memory or 1 minute max execution time limits.
> Hiphop helps, but there is still a highly broken code somewhere in our
> PHP source code.  No matter how much hacks we throw at bad code, the
> algorithm still need to get fixed.

For comparison: WYSIFTW parses [[Barak Obama]] in 3.5 sec on my iMac,
and in 4.4 sec on my MacBook (both Chrome 12).

Yes, it doesn't do template/variable replacing, and it's probably full
of corner cases that break; OTOH, it's JavaScript running in a
browser, which should make it much slower than a dedicated server
setup running precompiled PHP.

So, maybe another hard look at the MediaWiki parser is in order?

Cheers,
Magnus

_______________________________________________
Wikitech-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Reply via email to