https://bugzilla.wikimedia.org/show_bug.cgi?id=60826
Toby Negrin tneg...@wikimedia.org changed:
What|Removed |Added
CC|
https://bugzilla.wikimedia.org/show_bug.cgi?id=60826
--- Comment #5 from Erik Zachte ezac...@wikimedia.org ---
First step is done:
==adapting wikistats scripts==
*new argument -F to force processing full archive dumps (regardless of dump
size)
*Wikistats now can handle segmented dumps
(which
https://bugzilla.wikimedia.org/show_bug.cgi?id=60826
Erik Zachte ezac...@wikimedia.org changed:
What|Removed |Added
CC|dvanli...@gmail.com
https://bugzilla.wikimedia.org/show_bug.cgi?id=60826
Erik Zachte ezac...@wikimedia.org changed:
What|Removed |Added
Priority|Unprioritized |High
--- Comment
https://bugzilla.wikimedia.org/show_bug.cgi?id=60826
--- Comment #1 from Bingle bingle-ad...@wikimedia.org ---
Prioritization and scheduling of this bug is tracked on Mingle card
https://wikimedia.mingle.thoughtworks.com/projects/analytics/cards/cards/1429
--
You are receiving this mail
https://bugzilla.wikimedia.org/show_bug.cgi?id=60826
--- Comment #2 from Toby Negrin tneg...@wikimedia.org ---
I'd really like to see if we can use hadoop for further processing of the
dumps.
We can easily set up a hadoop instance in labs -- anybody interested in taking
a crack at this?
-Toby