Hmm, I had a similar trouble maybe 2 years ago, I will try to remember it. Somehow it was perhaps related to flagged and unflagged versions.

Ah, that's interesting.  Is there a different API call for flagged pages?


Or is there a more explicit way to iterate over the history? Maybe a lower level function that I should be using?


Please make sure you use the newest Pywiki version, send a piece of code and tell us which wiki do you speak about (I think Obama has to have an article approximately in all Wikipedias).


Using the latest nightly and the attached script, you can see that it only iterates up to 2006 for Mr. Obama...



import wikipedia

urlname = 'Barack_Obama'

site = wikipedia.Site('en')
page = wikipedia.Page(site, urlname)
history = page.fullVersionHistory(
    getAll=True, reverseOrder=True, revCount=500)

print "Got history for %s of length = %d", (urlname, len(history))

n = 0
for (revisionId, editTimestamp, username, content) in history:
    print n, len(content), editTimestamp
    n += 1

print 'finished!'


output:

<snip>
1391 39143 2006-09-01T17:26:38Z
1392 39112 2006-09-01T17:28:49Z
finished!




any advice?


Thanks!

jrf

import wikipedia

urlname = 'Barack_Obama'

site = wikipedia.Site('en')
page = wikipedia.Page(site, urlname)
history = page.fullVersionHistory(
    getAll=True, reverseOrder=True, revCount=500)


print "Got history for %s of length = %d", (urlname, len(history))

n = 0
for (revisionId, editTimestamp, username, content) in history:
    print n, len(content), editTimestamp
    n += 1

print 'finished!'
_______________________________________________
Pywikipedia-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/pywikipedia-l

Reply via email to