This information is available mostly pre-calculated in the CirrusSearch
dumps at http://dumps.wikimedia.your.org/other/cirrussearch/current/
Each article is represented by a line of json in those dumps. There is a
field called 'incoming_links' which is the number of unique articles with
links
I would second the recommendation of using the dumps for such a large
graphing project. If it's more than a couple hundred pages the API/database
queries can get bulky
On Sun, Mar 18, 2018 at 5:07 PM Brian Wolff wrote:
> Hi,
>
> You can run longer queries by getting access to
Hi,
You can run longer queries by getting access to toolforge (
https://wikitech.wikimedia.org/wiki/Portal:Toolforge) and running from the
command line.
However the query in question might still take an excessively long time
(if you are doing all of wikipedia). I would expect that query to
Hi there,
I'm a final year Mathematics student at the University of Bristol, and I'm
studying Wikipedia as a graph for my project.
I'd like to get data regarding the number of outgoing links on each page,
and the number of pages with links to each page. I have already
inquired about this with