https://bugzilla.wikimedia.org/show_bug.cgi?id=60826

--- Comment #5 from Erik Zachte <ezac...@wikimedia.org> ---
First step is done: 

==adapting wikistats scripts==

*new argument -F to force processing full archive dumps (regardless of dump
size)
*Wikistats now can handle segmented dumps 
(which BTW differ in file name for wp:de and wp:en) 
e.g. see first 100 lines or so in http://dumps.wikimedia.org/enwiki/20140304/ 
*Wikistats can now also see for segmented dumps if there was a error during
dump generation 
(by parsing dump job status report 'index.html') and looking for 'failed' in
appropriate sections
if found switch to other dump format, 
if none of dumps format is valid switch to older dump

Second step has started

==collect counts from full archive dumps for Wikipedias only on stat1==

*this will run for several weeks probably
* see for progress
http://stats.wikimedia.org/WikiCountsJobProgressCurrentStat1.html

Third step needs to be done

==merge results from stat1 into stat1002==

*make small script that merges values (missing values only) from 
stat1:/[..]/StatisticsMonthly.csv into 
stat1002:/[..]//StatisticsMonthly.csv 
as part of monthly Wikistats cycle stat1002:/[..]count_report_publish_wp.sh

-- 
You are receiving this mail because:
You are the assignee for the bug.
You are on the CC list for the bug.
_______________________________________________
Wikibugs-l mailing list
Wikibugs-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikibugs-l

Reply via email to