Bartosz Dziewoński wrote:
>It's also why #ifexist is expensive: it needs a separate database query
>for each time it's used, to check for a single page, because it's
>impossible to determine the list of pages to check in advance.

I'm not sure I understand the impossibility here.

When the expensive parser function count feature was added, I remember
this issue being discussed and my memory is that it seemed possible to
batch the ifexist lookups in a similar way to how we batch regular
internal link lookups against the pagelinks table, but nobody was
interested in implementing it at the time.

If the wikitext is parsed/evaluated on page save, I don't see why ifexist
lookups would be impossible to batch. We're already using the pagelinks
table for the ifexist functionality to properly work, as I understand it
(cf. <https://phabricator.wikimedia.org/T14019>).

MZMcBride



_______________________________________________
Wikitech-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Reply via email to