Revision: 7114
Author:   alexsh
Date:     2009-08-05 21:40:30 +0000 (Wed, 05 Aug 2009)

Log Message:
-----------
Remove all 'useAPI=True' tag for function set useAPI default True

Modified Paths:
--------------
    trunk/pywikipedia/casechecker.py
    trunk/pywikipedia/catlib.py
    trunk/pywikipedia/imageuncat.py
    trunk/pywikipedia/maintenance/family_check.py
    trunk/pywikipedia/welcome.py
    trunk/pywikipedia/wikipedia.py

Modified: trunk/pywikipedia/casechecker.py
===================================================================
--- trunk/pywikipedia/casechecker.py    2009-08-05 21:39:21 UTC (rev 7113)
+++ trunk/pywikipedia/casechecker.py    2009-08-05 21:40:30 UTC (rev 7114)
@@ -200,14 +200,14 @@
             wlpage = self.whitelists[self.site.lang]
             wikipedia.output(u'Loading whitelist from %s' % wlpage)
             wlparams = {
-                        'action'    : 'query',
-                        'prop'      : 'links',
-                        'titles'    : wlpage,
-                        'redirects' : '',
-                        'indexpageids' : '',
-                        }
+                'action'    : 'query',
+                'prop'      : 'links',
+                'titles'    : wlpage,
+                'redirects' : '',
+                'indexpageids' : '',
+            }
 
-            data = query.GetData(self.site.lang, wlparams, wikipedia.verbose, 
useAPI=True, encodeTitle=False)
+            data = query.GetData(wlparams, wikipedia.getSite(self.site.lang), 
wikipedia.verbose, encodeTitle=False)
             if len(data['query']['pageids']) == 1:
                 pageid = data['query']['pageids'][0]
                 links = data['query']['pages'][pageid]['links']

Modified: trunk/pywikipedia/catlib.py
===================================================================
--- trunk/pywikipedia/catlib.py 2009-08-05 21:39:21 UTC (rev 7113)
+++ trunk/pywikipedia/catlib.py 2009-08-05 21:40:30 UTC (rev 7114)
@@ -533,8 +533,7 @@
         'cmtitle'   :CatName,
         }
 
-    data = query.GetData(params, site = site, 
-                    useAPI = True, encodeTitle = False)
+    data = query.GetData(params, site, encodeTitle = False)
     categories_parsed.append(CatName)
     try:
         members = data['query']['categorymembers']

Modified: trunk/pywikipedia/imageuncat.py
===================================================================
--- trunk/pywikipedia/imageuncat.py     2009-08-05 21:39:21 UTC (rev 7113)
+++ trunk/pywikipedia/imageuncat.py     2009-08-05 21:40:30 UTC (rev 7114)
@@ -1243,7 +1243,7 @@
         'leend'     :today.strftime(dateformat)
         }
 
-    data = query.GetData(params, site, useAPI = True, encodeTitle = False)
+    data = query.GetData(params, site, encodeTitle = False)
     try:
         for item in data['query']['logevents']:
             result.append(item['title'])
@@ -1279,7 +1279,7 @@
         'rctype'    :'edit|log',
         }
 
-    data = query.GetData(params, site, useAPI = True, encodeTitle = False)
+    data = query.GetData(params, site, encodeTitle = False)
     try:
         for item in data['query']['recentchanges']:
             result.append(item['title'])

Modified: trunk/pywikipedia/maintenance/family_check.py
===================================================================
--- trunk/pywikipedia/maintenance/family_check.py       2009-08-05 21:39:21 UTC 
(rev 7113)
+++ trunk/pywikipedia/maintenance/family_check.py       2009-08-05 21:40:30 UTC 
(rev 7114)
@@ -17,7 +17,7 @@
                 'meta': 'siteinfo',
                 'siprop': 'namespaces'}
     try:
-        data = query.GetData(predata, site = site, useAPI = 
True)['query']['namespaces']
+        data = query.GetData(predata, site)['query']['namespaces']
     except wikipedia.ServerError, e:
         wikipedia.output(u'Warning! %s: %s' % (site, e))
         return

Modified: trunk/pywikipedia/welcome.py
===================================================================
--- trunk/pywikipedia/welcome.py        2009-08-05 21:39:21 UTC (rev 7113)
+++ trunk/pywikipedia/welcome.py        2009-08-05 21:40:30 UTC (rev 7114)
@@ -506,8 +506,7 @@
         'usprop'    :'blockinfo',
         }
 
-    data = query.GetData(params,
-                    useAPI = True, encodeTitle = False)
+    data = query.GetData(params, encodeTitle = False)
     # If there's not the blockedby parameter (that means the user isn't 
blocked), it will return False otherwise True.
     try:
         blockedBy = data['query']['users'][0]['blockedby']

Modified: trunk/pywikipedia/wikipedia.py
===================================================================
--- trunk/pywikipedia/wikipedia.py      2009-08-05 21:39:21 UTC (rev 7113)
+++ trunk/pywikipedia/wikipedia.py      2009-08-05 21:40:30 UTC (rev 7114)
@@ -919,8 +919,7 @@
             'prop'      :'info',
             'titles'    :self.title(),
             }
-        data = query.GetData(params,
-                        useAPI = True, encodeTitle = False)        
+        data = query.GetData(params, encodeTitle = False)        
         pageid = data['query']['pages'].keys()[0]
         if data['query']['pages'][pageid].keys()[0] == 'lastrevid':
             return data['query']['pages'][pageid]['lastrevid'] # if ok,
@@ -952,8 +951,7 @@
             'tllimit'   :tllimit,
             }
 
-        data = query.GetData(params,
-                        useAPI = True, encodeTitle = False)
+        data = query.GetData(params, encodeTitle = False)
         try:
             pageid = data['query']['pages'].keys()[0]
         except KeyError:
@@ -1327,7 +1325,7 @@
         #if titles:
         #    predata['titles'] = query.ListToParam(titles)
         
-        text = query.GetData(predata, useAPI = True)['query']['pages']
+        text = query.GetData(predata)['query']['pages']
         
         for pageid in text:
             if text[pageid].has_key('missing'):
@@ -2765,7 +2763,7 @@
             'titles'    :self.title(),
             }
         try:
-            data = query.GetData(params, useAPI = True, encodeTitle = 
False)['query']['pages']
+            data = query.GetData(params, encodeTitle = False)['query']['pages']
         except KeyError:
             raise NoPage(u'API Error, nothing found in the APIs')
 
@@ -2829,7 +2827,7 @@
             'titles'    :self.title(),
             'iiprop'    :'url',
         }
-        imagedata = query.GetData(params, useAPI = True, encodeTitle = False)
+        imagedata = query.GetData(params, encodeTitle = False)
         try:
             url=imagedata['query']['pages'].values()[0]['imageinfo'][0]['url']
 #        urlR = re.compile(r'<div class="fullImageLink" id="file">.*?<a 
href="(?P<url>[^ ]+?)"(?! class="image")|<span class="dangerousLink"><a 
href="(?P<url2>.+?)"', re.DOTALL)
@@ -2899,7 +2897,7 @@
             'prop'      :'imageinfo',
             'titles'    :self.title(),
             }
-        data = query.GetData(params, useAPI = True, encodeTitle = False)
+        data = query.GetData(params, encodeTitle = False)
         try:
             # We don't know the page's id, if any other better idea please 
change it
             pageid = data['query']['pages'].keys()[0]
@@ -2921,7 +2919,7 @@
                 'iiprop'    :'sha1',
                 }
             # First of all we need the Hash that identify an image
-            data = query.GetData(params, useAPI = True, encodeTitle = False)
+            data = query.GetData(params, encodeTitle = False)
             pageid = data['query']['pages'].keys()[0]
             try:
                 hash_found = 
data['query']['pages'][pageid][u'imageinfo'][0][u'sha1']
@@ -4505,7 +4503,7 @@
                 'meta': 'userinfo',
                 'uiprop': 'blockinfo',
             }
-            data = query.GetData(params, self, useAPI = 
True)['query']['userinfo']
+            data = query.GetData(params, self)['query']['userinfo']
             return data.has_key('blockby')
         except NotImplementedError:
             return False
@@ -5121,7 +5119,7 @@
                         'meta':'allmessages',
                     }
                     try:
-                        datas = query.GetData(params, useAPI = 
True)['query']['allmessages']
+                        datas = query.GetData(params)['query']['allmessages']
                     except KeyError:
                         raise ServerError("The APIs don't return data, the 
site may be down")
                     except NotImplementedError:
@@ -5222,7 +5220,7 @@
                 'meta': 'userinfo',
                 'uiprop': 
'blockinfo|groups|rights|hasmsg|ratelimits|preferencestoken',
             }
-            text = query.GetData(params, site = self, useAPI = True, 
sysop=sysop)['query']['userinfo']
+            text = query.GetData(params, self, 
sysop=sysop)['query']['userinfo']
             ##output('%s' % text) # for debug use only
         else:
             url = self.edit_address('Non-existing_page')
@@ -5503,8 +5501,7 @@
         if leuser is not None: params['leuser'] = leuser
         if letitle is not None: params['letitle'] = letitle
         while True:
-            data = query.GetData(params,
-                            useAPI = True, encodeTitle = False)
+            data = query.GetData(params, encodeTitle = False)
             try:
                 imagesData = data['query']['logevents']
             except KeyError:
@@ -5572,8 +5569,7 @@
         if rcshow is not None: params['rcshow'] = rcshow
         if rctype is not None: params['rctype'] = rctype
         while True:
-            data = query.GetData(params,
-                            useAPI = True, encodeTitle = False)
+            data = query.GetData(params, encodeTitle = False)
             try:
                 rcData = data['query']['recentchanges']
             except KeyError:
@@ -5709,7 +5705,7 @@
                 'rnlimit': '1',
                 #'': '',
             }
-            data = query.GetData(params, useAPI = True)
+            data = query.GetData(params)
             return Page(self, data['query']['random'][0]['title'])
         else:
             """Yield random page via Special:Random"""
@@ -5727,7 +5723,7 @@
                 'rnlimit': '1',
                 'rnredirect': '1',
             }
-            data = query.GetData(params, useAPI = True)
+            data = query.GetData(params)
             return Page(self, data['query']['random'][0]['title'])
         else:
             """Yield random redirect page via Special:RandomRedirect."""
@@ -5781,7 +5777,7 @@
             params['apfrom'] = start
             if throttle:
                 get_throttle()
-            data = query.GetData(params, useAPI = True)
+            data = query.GetData(params)
             
             for p in data['query']['allpages']:
                 yield Page(self, p['title'])
@@ -5973,7 +5969,7 @@
                     'euquery': url,
                 }
                 while True:
-                    data = query.GetData(params, useAPI = True)
+                    data = query.GetData(params)
                     if data['query']['exturlusage'] == []:
                         break
                     for pages in data['query']['exturlusage']:
@@ -6586,7 +6582,7 @@
             'list'      :'allimages',
             'aisha1'    :hash_found,
         }
-        allimages = query.GetData(params, site = getSite(self.lang, 
self.family), useAPI = True, encodeTitle = False)['query']['allimages']
+        allimages = query.GetData(params, getSite(self.lang, self.family), 
encodeTitle = False)['query']['allimages']
         files = list()
         for imagedata in allimages:
             image = imagedata[u'name']



_______________________________________________
Pywikipedia-svn mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/pywikipedia-svn

Reply via email to