Stefan.petrea has uploaded a new change for review. https://gerrit.wikimedia.org/r/61005
Change subject: Squashed commit of the following: ...................................................................... Squashed commit of the following: commit a2ad65cbef62cfe68bbb484a0662465bd19cc501 Author: Petrea Corneliu Stefan <ste...@garage-coding.com> Date: Fri Apr 26 17:28:27 2013 +0300 Fault tolerant tests * In case some required modules don't exist skip some tests so test plan is still executed Change-Id: I96352bb0109f6eeeeff962d8a1d13ea20fc75f7a --- M pageviews_reports/lib/PageViews/Model/Sequential.pm M pageviews_reports/t/01-bot-detection.t 2 files changed, 34 insertions(+), 22 deletions(-) git pull ssh://gerrit.wikimedia.org:29418/analytics/wikistats refs/changes/05/61005/1 diff --git a/pageviews_reports/lib/PageViews/Model/Sequential.pm b/pageviews_reports/lib/PageViews/Model/Sequential.pm index 5f344d3..15669bf 100644 --- a/pageviews_reports/lib/PageViews/Model/Sequential.pm +++ b/pageviews_reports/lib/PageViews/Model/Sequential.pm @@ -3,7 +3,7 @@ use warnings; use File::Basename qw/basename/; use Data::Dumper; -use PageViews::BotDetector; +#use PageViews::BotDetector; use Time::Piece; use Carp; @@ -99,22 +99,22 @@ # The values will indicate what kind of url it is. # -sub build_accepted_url_regex2 { - my $ra = Regexp::Assemble->new(); - for my $p1 ("http","https") { - for my $p2 ("","www.") { - for my $l(@accept_lang_array) { - for my $d (@accept_domain_suffixes) { - $ra->add("$p1://$p2$l.m.$d.org/wiki/"); - $ra->add("$p1://$p2$l.m.$d.org/w/index.php"); - $ra->add("$p1://$p2$l.m.$d.org/w/api.php\?action=mobile"); - $ra->add("$p1://$p2$l.m.$d.org/w/api.php\?action=mobileview"); - }; - }; - }; - }; - return $ra->re(); -} +#sub build_accepted_url_regex2 { + #my $ra = Regexp::Assemble->new(); + #for my $p1 ("http","https") { + #for my $p2 ("","www.") { + #for my $l(@accept_lang_array) { + #for my $d (@accept_domain_suffixes) { + #$ra->add("$p1://$p2$l.m.$d.org/wiki/"); + #$ra->add("$p1://$p2$l.m.$d.org/w/index.php"); + #$ra->add("$p1://$p2$l.m.$d.org/w/api.php\?action=mobile"); + #$ra->add("$p1://$p2$l.m.$d.org/w/api.php\?action=mobileview"); + #}; + #}; + #}; + #}; + #return $ra->re(); +#} sub new { @@ -147,11 +147,11 @@ counts_discarded_mimetype => {}, counts_mimetype => {}, - bdetector => PageViews::BotDetector->new(), + #bdetector => PageViews::BotDetector->new(), }; $raw_obj->{accept_re} = build_accepted_url_regex1(), - $raw_obj->{bdetector}->load_ip_ranges(); - $raw_obj->{bdetector}->load_useragent_regex(); + #$raw_obj->{bdetector}->load_ip_ranges(); + #$raw_obj->{bdetector}->load_useragent_regex(); my $obj = bless $raw_obj,$class; diff --git a/pageviews_reports/t/01-bot-detection.t b/pageviews_reports/t/01-bot-detection.t index aef5bf7..d2604c1 100644 --- a/pageviews_reports/t/01-bot-detection.t +++ b/pageviews_reports/t/01-bot-detection.t @@ -1,9 +1,21 @@ #!/usr/bin/env perl use strict; use warnings; -use Test::More qw/no_plan/; +use Test::More; + +eval { + require Regexp::Assemble; + require Net::Patricia; +}; + +if($@) { + plan skip_all => "Regexp::Assemble and Net::Patricia not found"; +} else { + plan no_plan => 0; +}; + use lib "./lib"; -use PageViews::BotDetector; +require PageViews::BotDetector; my $b = PageViews::BotDetector->new; $b->load_ip_ranges(); -- To view, visit https://gerrit.wikimedia.org/r/61005 To unsubscribe, visit https://gerrit.wikimedia.org/r/settings Gerrit-MessageType: newchange Gerrit-Change-Id: I96352bb0109f6eeeeff962d8a1d13ea20fc75f7a Gerrit-PatchSet: 1 Gerrit-Project: analytics/wikistats Gerrit-Branch: master Gerrit-Owner: Stefan.petrea <ste...@garage-coding.com> _______________________________________________ MediaWiki-commits mailing list MediaWiki-commits@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/mediawiki-commits