Hello community,
here is the log from the commit of package perl-WWW-RobotRules for
openSUSE:Factory checked in at 2012-03-01 17:25:32
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Comparing /work/SRC/openSUSE:Factory/perl-WWW-RobotRules (Old)
and /work/SRC/openSUSE:Factory/.perl-WWW-RobotRules.new (New)
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Package is "perl-WWW-RobotRules", Maintainer is ""
Changes:
--------
--- /work/SRC/openSUSE:Factory/perl-WWW-RobotRules/perl-WWW-RobotRules.changes
2011-09-23 12:39:23.000000000 +0200
+++
/work/SRC/openSUSE:Factory/.perl-WWW-RobotRules.new/perl-WWW-RobotRules.changes
2012-03-01 17:25:34.000000000 +0100
@@ -1,0 +2,6 @@
+Mon Feb 20 10:48:55 UTC 2012 - [email protected]
+
+- updated to 6.02
+ * Restore perl-5.8.1 compatiblity.
+
+-------------------------------------------------------------------
Old:
----
WWW-RobotRules-6.01.tar.gz
New:
----
WWW-RobotRules-6.02.tar.gz
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Other differences:
------------------
++++++ perl-WWW-RobotRules.spec ++++++
--- /var/tmp/diff_new_pack.JbuBZ9/_old 2012-03-01 17:25:38.000000000 +0100
+++ /var/tmp/diff_new_pack.JbuBZ9/_new 2012-03-01 17:25:38.000000000 +0100
@@ -1,7 +1,7 @@
#
# spec file for package perl-WWW-RobotRules
#
-# Copyright (c) 2011 SUSE LINUX Products GmbH, Nuernberg, Germany.
+# Copyright (c) 2012 SUSE LINUX Products GmbH, Nuernberg, Germany.
#
# All modifications and additions to the file contributed by third parties
# remain the property of their copyright owners, unless otherwise agreed
@@ -16,25 +16,23 @@
#
-
Name: perl-WWW-RobotRules
-Version: 6.01
-Release: 1
-License: GPL-1.0+ or Artistic-1.0
+Version: 6.02
+Release: 0
%define cpan_name WWW-RobotRules
Summary: database of robots.txt-derived permissions
-Url: http://search.cpan.org/dist/WWW-RobotRules/
+License: Artistic-1.0 or GPL-1.0+
Group: Development/Libraries/Perl
-#Source:
http://www.cpan.org/authors/id/G/GA/GAAS/WWW-RobotRules-%{version}.tar.gz
-Source: %{cpan_name}-%{version}.tar.gz
-BuildRequires: perl(Fcntl)
-BuildRequires: perl(URI) >= 1.10
+Url: http://search.cpan.org/dist/WWW-RobotRules/
+Source:
http://www.cpan.org/authors/id/G/GA/GAAS/%{cpan_name}-%{version}.tar.gz
+BuildArch: noarch
+BuildRoot: %{_tmppath}/%{name}-%{version}-build
BuildRequires: perl
BuildRequires: perl-macros
-Requires: perl(Fcntl)
+BuildRequires: perl(URI) >= 1.10
+#BuildRequires: perl(WWW::RobotRules)
+#BuildRequires: perl(WWW::RobotRules::AnyDBM_File)
Requires: perl(URI) >= 1.10
-BuildRoot: %{_tmppath}/%{name}-%{version}-build
-BuildArch: noarch
%{perl_requires}
%description
@@ -43,9 +41,35 @@
can use the _/robots.txt_ file to forbid conforming robots from accessing
parts of their web site.
+The parsed files are kept in a WWW::RobotRules object, and this object
+provides methods to check if access to a given URL is prohibited. The same
+WWW::RobotRules object can be used for one or more parsed _/robots.txt_
+files on any number of hosts.
+
+The following methods are provided:
+
+* $rules = WWW::RobotRules->new($robot_name)
+
+ This is the constructor for WWW::RobotRules objects. The first argument
+ given to new() is the name of the robot.
+
+* $rules->parse($robot_txt_url, $content, $fresh_until)
+
+ The parse() method takes as arguments the URL that was used to retrieve
+ the _/robots.txt_ file, and the contents of the file.
+
+* $rules->allowed($uri)
+
+ Returns TRUE if this robot is allowed to retrieve this URL.
+
+* $rules->agent([$name])
+
+ Get/set the agent name. NOTE: Changing the agent name will clear the
+ robots.txt rules and expire times out of the cache.
%prep
%setup -q -n %{cpan_name}-%{version}
+find . -type f -print0 | xargs -0 chmod 644
%build
%{__perl} Makefile.PL INSTALLDIRS=vendor
@@ -59,11 +83,8 @@
%perl_process_packlist
%perl_gen_filelist
-%clean
-%{__rm} -rf %{buildroot}
-
%files -f %{name}.files
-%defattr(644,root,root,755)
+%defattr(-,root,root,755)
%doc Changes README
%changelog
++++++ WWW-RobotRules-6.01.tar.gz -> WWW-RobotRules-6.02.tar.gz ++++++
diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn'
'--exclude=.svnignore' old/WWW-RobotRules-6.01/Changes
new/WWW-RobotRules-6.02/Changes
--- old/WWW-RobotRules-6.01/Changes 2011-03-13 16:08:40.000000000 +0100
+++ new/WWW-RobotRules-6.02/Changes 2012-02-18 14:07:15.000000000 +0100
@@ -1,4 +1,11 @@
_______________________________________________________________________________
+2012-02-18 WWW-RobotRules 6.02
+
+Restore perl-5.8.1 compatiblity.
+
+
+
+_______________________________________________________________________________
2011-03-13 WWW-RobotRules 6.01
Added legal notice and updated the meta repository link
diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn'
'--exclude=.svnignore' old/WWW-RobotRules-6.01/META.yml
new/WWW-RobotRules-6.02/META.yml
--- old/WWW-RobotRules-6.01/META.yml 2011-03-13 16:08:54.000000000 +0100
+++ new/WWW-RobotRules-6.02/META.yml 2012-02-18 14:07:29.000000000 +0100
@@ -1,6 +1,6 @@
--- #YAML:1.0
name: WWW-RobotRules
-version: 6.01
+version: 6.02
abstract: database of robots.txt-derived permissions
author:
- Gisle Aas <[email protected]>
@@ -13,16 +13,16 @@
requires:
AnyDBM_File: 0
Fcntl: 0
- perl: 5.008008
+ perl: 5.008001
URI: 1.10
resources:
MailingList: mailto:[email protected]
- repository:
http://github.com/gisle/libwww-perl/tree/WWW-RobotRules/master
+ repository: http://github.com/gisle/www-robotrules
no_index:
directory:
- t
- inc
-generated_by: ExtUtils::MakeMaker version 6.56
+generated_by: ExtUtils::MakeMaker version 6.57_05
meta-spec:
url: http://module-build.sourceforge.net/META-spec-v1.4.html
version: 1.4
diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn'
'--exclude=.svnignore' old/WWW-RobotRules-6.01/Makefile.PL
new/WWW-RobotRules-6.02/Makefile.PL
--- old/WWW-RobotRules-6.01/Makefile.PL 2011-03-13 16:08:36.000000000 +0100
+++ new/WWW-RobotRules-6.02/Makefile.PL 2012-02-18 13:57:41.000000000 +0100
@@ -1,6 +1,6 @@
#!perl -w
-require 5.008008;
+require 5.008001;
use strict;
use ExtUtils::MakeMaker;
@@ -10,7 +10,7 @@
ABSTRACT_FROM => 'lib/WWW/RobotRules.pm',
AUTHOR => 'Gisle Aas <[email protected]>',
LICENSE => "perl",
- MIN_PERL_VERSION => 5.008008,
+ MIN_PERL_VERSION => 5.008001,
PREREQ_PM => {
'AnyDBM_File' => 0,
'Fcntl' => 0,
@@ -18,7 +18,7 @@
},
META_MERGE => {
resources => {
- repository =>
'http://github.com/gisle/libwww-perl/tree/WWW-RobotRules/master',
+ repository => 'http://github.com/gisle/www-robotrules',
MailingList => 'mailto:[email protected]',
}
},
diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn'
'--exclude=.svnignore' old/WWW-RobotRules-6.01/lib/WWW/RobotRules.pm
new/WWW-RobotRules-6.02/lib/WWW/RobotRules.pm
--- old/WWW-RobotRules-6.01/lib/WWW/RobotRules.pm 2011-03-13
16:08:40.000000000 +0100
+++ new/WWW-RobotRules-6.02/lib/WWW/RobotRules.pm 2012-02-18
14:05:13.000000000 +0100
@@ -1,6 +1,6 @@
package WWW::RobotRules;
-$VERSION = "6.01";
+$VERSION = "6.02";
sub Version { $VERSION; }
use strict;
--
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]