Took the tip. Though I did the monobook force manually myself. That
patch may have aged and missed new things, also that one for Export was
redundant (Export is a raw page with no skin).
Actually, I was planning on doing something one up on what you thought
of. Though, when I first thought of it, the pywikipediabot guys rejected it.
I was going to tweak the function that grabs family files. In a way so
that when you use -family:wikia_/name/ and it cannot find a Wikia
family, it will fallback to a special class which will use the API to
grab namespace and other information about the wiki. Then it will save
that data to a flat file (What's the point of building a brand new
family file for each wiki, just store it all in a flatfile) and thus the
framework will work all over wikia without ever needing a family file,
unless your wiki is desynced from Wikia (ie: Uncyclopedia), or is one of
the few which use the /w format cause they moved to Wikia from another
place and need link compatibility.
Of course, that would be changed so that now you would simply use
-family:/name/ when I get around to building that for this repo.
As for your patch.
I did some file searching before I made my change.
But I did it in the reverse of you and found some other info:
- export_address is the function which gives out the export page address.
-- _GetAll.getData is the only thing which calls for that address
(besides perhaps one of my scripts or something real odd we don't need
to worry about)
--- getall is the function which scripts actually use to grab multiple
pages of data.
I don't know what tree leads replace.py to getall cause I didn't do
enough hunting. But however, you know the "Getting ## pages from
/somesite/..." that shows up as output in replace.py to indicate data
being grabbed. That is part of getall, and that's the only place which
has that output.
The issue with editing get other than the fact that you are also
applying that replacement to the edit page text you should not be
applying it to, is that there are 7 other scripts which call getall. Not
only that, but that includes the pagegenerators (which are likely what
replace.py is actually using to get the data), and even the
disambiguation scripts. So by editing get instead of _GetAll the scripts
which use getall or a page generator like they are supposed to are all
still broken.
~Daniel Friesen(Dantman) of:
-The Gaiapedia (http://gaia.wikia.com)
-Wikia ACG on Wikia.com (http://wikia.com/wiki/Wikia_ACG)
-and Wiki-Tools.com (http://wiki-tools.com)
C Stafford wrote:
re:
http://svn.nadir-point.com/viewvc/wikia-pywikibot/wikipedia.py?r1=10&r2=9&pathrev=10
see
https://sourceforge.net/tracker/index.php?func=detail&aid=1885569&group_id=93107&atid=603139
also, in general
https://sourceforge.net/tracker/index.php?func=detail&aid=1916496&group_id=93107&atid=603141
since i was doing random bot work all around wikia, i got tired of
manually creating a family file for each wiki, especially when i knew
i was only going to be botting there this one time.
i figured it would be worth the time to write something to generate
the family files dynamically (well, with some cache) from the wiki's
current namespace structure (yay Special:Export), than to have an
archive of only certain ones.
so i wrote one. works quite nicely too, random sample url from the
wiki goes in, family file come out.
sadly, its only here on my local machine right now (i had it online
briefly for some testing, thanks again manticore and jack). i'll see
what i can do it get it cleaned up (and protected against attacks) and
with a better interface and put online for people.
-ubrfzy
_______________________________________________
Wikia-l mailing list
[email protected]
http://lists.wikia.com/mailman/listinfo/wikia-l
_______________________________________________
Wikia-l mailing list
[email protected]
http://lists.wikia.com/mailman/listinfo/wikia-l