Hi Brad,

You mentioned "a while back" for "apcontinue", show recent was it? This
dump generator is attempting to archive all sorts of versions of MediaWiki,
or so unless we write a backward compatibility handler in the script itself.

...and I agree, the code is in a total mess. We need to get someone to
rewrite the whole thing, soon.

On Fri, Nov 9, 2012 at 11:50 PM, Brad Jorsch <[email protected]> wrote:

> You're searching for the continue parameter as "apfrom", but this was
> changed to "apcontinue" a while back. Changing line 162 to something
> like this should probably do it:
>
>     m = re.findall(r'<allpages (?:apfrom|apcontinue)="([^>]+)" />', xml)
>
> Note that for full correctness, you probably should omit both apfrom
> and apcontinue entirely from params the first time around, and send
> back whichever of the two is found by the above line in subsequent
> queries.
>
> Also, why in the world aren't you using an XML parser (or a JSON
> parser with format=json) to process the API response instead of trying
> to parse the XML using regular expressions?!
>
> On Fri, Nov 9, 2012 at 2:27 AM, Federico Leva (Nemo) <[email protected]>
> wrote:
> > It's completely broken:
> > https://code.google.com/p/wikiteam/issues/detail?id=56
> > It will download only a fraction of the wiki, 500 pages at most per
> > namespace.
>
> _______________________________________________
> Mediawiki-api mailing list
> [email protected]
> https://lists.wikimedia.org/mailman/listinfo/mediawiki-api
>



-- 
Regards,
Hydriz

We've created the greatest collection of shared knowledge in history. Help
protect Wikipedia. Donate now: http://donate.wikimedia.org
_______________________________________________
Mediawiki-api mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/mediawiki-api

Reply via email to