https://bugzilla.wikimedia.org/show_bug.cgi?id=62209

--- Comment #2 from Dimitris Kontokostas <jimk...@gmail.com> ---
Thanks,

I already saw the &exintro option so, one question to understand this. 

when I use this call:
http://en.wikipedia.org/w/api.php?action=query&prop=extracts&exintro=&explaintext=&titles=Athens

does this extension loads the whole page, convert it to html and then return
the first section? 
if not this extension is perfect for our purpose and don't read the rest :)

if yes, we would like to avoid loading the whole page as it would slow down our
extraction.

What we do so far is to take the wiki markup of the page up to the first
section and feed it in the mw "parse" api call [1] which normally returns html.
Then we hack into the mw core to return cleaned text. 

So, the request is to add the "text" and "title" parameters in your api. When
they are given, instead of parsing the page by title you will parse the "text"
parameter ("title" is used for  magic words like {{PAGENAME}}), get the html
and clean it the same way you do now.

Cheers,
Dimitris

[1] https://www.mediawiki.org/wiki/API:Parsing_wikitext#parse

-- 
You are receiving this mail because:
You are the assignee for the bug.
You are on the CC list for the bug.
_______________________________________________
Wikibugs-l mailing list
Wikibugs-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikibugs-l

Reply via email to