https://bugzilla.wikimedia.org/show_bug.cgi?id=60826
--- Comment #5 from Erik Zachte <[email protected]> --- First step is done: ==adapting wikistats scripts== *new argument -F to force processing full archive dumps (regardless of dump size) *Wikistats now can handle segmented dumps (which BTW differ in file name for wp:de and wp:en) e.g. see first 100 lines or so in http://dumps.wikimedia.org/enwiki/20140304/ *Wikistats can now also see for segmented dumps if there was a error during dump generation (by parsing dump job status report 'index.html') and looking for 'failed' in appropriate sections if found switch to other dump format, if none of dumps format is valid switch to older dump Second step has started ==collect counts from full archive dumps for Wikipedias only on stat1== *this will run for several weeks probably * see for progress http://stats.wikimedia.org/WikiCountsJobProgressCurrentStat1.html Third step needs to be done ==merge results from stat1 into stat1002== *make small script that merges values (missing values only) from stat1:/[..]/StatisticsMonthly.csv into stat1002:/[..]//StatisticsMonthly.csv as part of monthly Wikistats cycle stat1002:/[..]count_report_publish_wp.sh -- You are receiving this mail because: You are the assignee for the bug. You are on the CC list for the bug. _______________________________________________ Wikibugs-l mailing list [email protected] https://lists.wikimedia.org/mailman/listinfo/wikibugs-l
