The list of articles was decided here phabricator.wikimedia.org/T120504

There's a page for a few categories, a very short article (Campus
honeymoon), a long one (Barack Obama).

As I mentioned in the task it is trivial to run the report with a different
sample set and I'm happy to if somebody has the interest to visualize a
different set of articles.

Adam for example in the task posted different sets of articles based on
other criteria like page views (
https://phabricator.wikimedia.org/T120504#1900287), it'd be interesting to
run those and see if the trends on navbox and reference sizes hold up.

We're also thinking about running a similar report across a bigger dataset
in a more aggregated way, maybe the top 100.000 articles on pageviews with
the sizes weighted with the pageviews number to get a more global
understanding, but we haven't gotten around to it yet (it would be a new
more global, less per-page one).
On Jan 20, 2016 6:53 PM, "Federico Leva (Nemo)" <[email protected]> wrote:

> Joaquin Oltra Hernandez, 20/01/2016 16:55:
>
>> We've added Mediawiki parser content analysis to the content analysis
>> report that the Reading web team performed last quarter.
>>
>
> Thanks. It would be useful to understand what your dataset is: I see 9
> page titles, presumably fetched from the English Wikipedia. Is this your
> dataset? How did you ensure it's representative of what users see?
>
> Nemo
>
_______________________________________________
Mobile-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/mobile-l

Reply via email to