Author: qboosh Date: Sun Mar 13 07:11:53 2011 GMT Module: packages Tag: HEAD ---- Log message: - new
---- Files affected: packages/perl-WWW-RobotRules: perl-WWW-RobotRules.spec (NONE -> 1.1) (NEW) ---- Diffs: ================================================================ Index: packages/perl-WWW-RobotRules/perl-WWW-RobotRules.spec diff -u /dev/null packages/perl-WWW-RobotRules/perl-WWW-RobotRules.spec:1.1 --- /dev/null Sun Mar 13 08:11:53 2011 +++ packages/perl-WWW-RobotRules/perl-WWW-RobotRules.spec Sun Mar 13 08:11:48 2011 @@ -0,0 +1,85 @@ +# $Revision$, $Date$ +# +# Conditional build: +%bcond_without tests # do not perform "make test" +# +%define pdir WWW +%define pnam RobotRules +%include /usr/lib/rpm/macros.perl +Summary: WWW::RobotRules - database of robots.txt-derived permissions +Summary(pl.UTF-8): WWW::RobotRules - baza danych uprawnień z robots.txt +Name: perl-WWW-RobotRules +Version: 6.00 +Release: 1 +# same as perl +License: GPL v1+ or Artistic +Group: Development/Languages/Perl +Source0: http://www.cpan.org/modules/by-module/WWW/%{pdir}-%{pnam}-%{version}.tar.gz +# Source0-md5: 4b5f4838bf1de44057b0bd6240b05dd7 +URL: http://search.cpan.org/dist/WWW-RobotRules/ +BuildRequires: perl-devel >= 1:5.8.8 +BuildRequires: rpm-perlprov >= 4.1-13 +%if %{with tests} +BuildRequires: perl-URI >= 1.10 +%endif +Conflicts: perl-libwww < 6 +BuildArch: noarch +BuildRoot: %{tmpdir}/%{name}-%{version}-root-%(id -u -n) + +%description +This module parses /robots.txt files as specified in "A Standard for +Robot Exclusion", at <http://www.robotstxt.org/wc/norobots.html>. +Webmasters can use the /robots.txt file to forbid conforming robots +from accessing parts of their web site. + +The parsed files are kept in a WWW::RobotRules object, and this object +provides methods to check if access to a given URL is prohibited. The +same WWW::RobotRules object can be used for one or more parsed +/robots.txt files on any number of hosts. + +%description -l pl.UTF-8 +Ten moduł analizuje pliki /robots.txt opisane w dokumencie "A Standard +for Robot Exclusion" (<http://www.robotstxt.org/wc/norobots.html>). +Webmasterzy mogą używać pliku /robots.txt, aby zabronić automatom +zgodnym z tym standardem dostępu do pewnych plików na stronie WWW. + +Przeanalizowane pliki są trzymane w obiekcie WWW::RobotRules, a obiekt +ten udostępnia metody do sprawdzania, czy dostęp do danego URL-a jest +zabroniony. Ten sam obiekt WWW::RobotRules może być używany dla +jednego lub większej liczby przeanalizowanych plików /robots.txt z +różnych hostów. + +%prep +%setup -q -n %{pdir}-%{pnam}-%{version} + +%build +%{__perl} Makefile.PL \ + INSTALLDIRS=vendor +%{__make} + +%{?with_tests:%{__make} test} + +%install +rm -rf $RPM_BUILD_ROOT + +%{__make} pure_install \ + DESTDIR=$RPM_BUILD_ROOT + +%clean +rm -rf $RPM_BUILD_ROOT + +%files +%defattr(644,root,root,755) +%doc Changes README +%{perl_vendorlib}/WWW/RobotRules.pm +%{perl_vendorlib}/WWW/RobotRules +%{_mandir}/man3/WWW::RobotRules*.3pm* + +%define date %(echo `LC_ALL="C" date +"%a %b %d %Y"`) +%changelog +* %{date} PLD Team <[email protected]> +All persons listed below can be reached at <cvs_login>@pld-linux.org + +$Log$ +Revision 1.1 2011/03/13 07:11:48 qboosh +- new ================================================================ _______________________________________________ pld-cvs-commit mailing list [email protected] http://lists.pld-linux.org/mailman/listinfo/pld-cvs-commit
