Hello community,
here is the log from the commit of package perl-WWW-RobotRules for openSUSE:Factory checked in at 2012-03-01 17:25:32
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Comparing /work/SRC/openSUSE:Factory/perl-WWW-RobotRules (Old)
and /work/SRC/openSUSE:Factory/.perl-WWW-RobotRules.new (New)
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Package is "perl-WWW-RobotRules", Maintainer is ""
Changes:
--------
--- /work/SRC/openSUSE:Factory/perl-WWW-RobotRules/perl-WWW-RobotRules.changes 2011-09-23 12:39:23.000000000 +0200
+++ /work/SRC/openSUSE:Factory/.perl-WWW-RobotRules.new/perl-WWW-RobotRules.changes 2012-03-01 17:25:34.000000000 +0100
@@ -1,0 +2,6 @@
+Mon Feb 20 10:48:55 UTC 2012 - coolo@suse.com
+
+- updated to 6.02
+ * Restore perl-5.8.1 compatiblity.
+
+-------------------------------------------------------------------
Old:
----
WWW-RobotRules-6.01.tar.gz
New:
----
WWW-RobotRules-6.02.tar.gz
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Other differences:
------------------
++++++ perl-WWW-RobotRules.spec ++++++
--- /var/tmp/diff_new_pack.JbuBZ9/_old 2012-03-01 17:25:38.000000000 +0100
+++ /var/tmp/diff_new_pack.JbuBZ9/_new 2012-03-01 17:25:38.000000000 +0100
@@ -1,7 +1,7 @@
#
# spec file for package perl-WWW-RobotRules
#
-# Copyright (c) 2011 SUSE LINUX Products GmbH, Nuernberg, Germany.
+# Copyright (c) 2012 SUSE LINUX Products GmbH, Nuernberg, Germany.
#
# All modifications and additions to the file contributed by third parties
# remain the property of their copyright owners, unless otherwise agreed
@@ -16,25 +16,23 @@
#
-
Name: perl-WWW-RobotRules
-Version: 6.01
-Release: 1
-License: GPL-1.0+ or Artistic-1.0
+Version: 6.02
+Release: 0
%define cpan_name WWW-RobotRules
Summary: database of robots.txt-derived permissions
-Url: http://search.cpan.org/dist/WWW-RobotRules/
+License: Artistic-1.0 or GPL-1.0+
Group: Development/Libraries/Perl
-#Source: http://www.cpan.org/authors/id/G/GA/GAAS/WWW-RobotRules-%{version}.tar.gz
-Source: %{cpan_name}-%{version}.tar.gz
-BuildRequires: perl(Fcntl)
-BuildRequires: perl(URI) >= 1.10
+Url: http://search.cpan.org/dist/WWW-RobotRules/
+Source: http://www.cpan.org/authors/id/G/GA/GAAS/%{cpan_name}-%{version}.tar.gz
+BuildArch: noarch
+BuildRoot: %{_tmppath}/%{name}-%{version}-build
BuildRequires: perl
BuildRequires: perl-macros
-Requires: perl(Fcntl)
+BuildRequires: perl(URI) >= 1.10
+#BuildRequires: perl(WWW::RobotRules)
+#BuildRequires: perl(WWW::RobotRules::AnyDBM_File)
Requires: perl(URI) >= 1.10
-BuildRoot: %{_tmppath}/%{name}-%{version}-build
-BuildArch: noarch
%{perl_requires}
%description
@@ -43,9 +41,35 @@
can use the _/robots.txt_ file to forbid conforming robots from accessing
parts of their web site.
+The parsed files are kept in a WWW::RobotRules object, and this object
+provides methods to check if access to a given URL is prohibited. The same
+WWW::RobotRules object can be used for one or more parsed _/robots.txt_
+files on any number of hosts.
+
+The following methods are provided:
+
+* $rules = WWW::RobotRules->new($robot_name)
+
+ This is the constructor for WWW::RobotRules objects. The first argument
+ given to new() is the name of the robot.
+
+* $rules->parse($robot_txt_url, $content, $fresh_until)
+
+ The parse() method takes as arguments the URL that was used to retrieve
+ the _/robots.txt_ file, and the contents of the file.
+
+* $rules->allowed($uri)
+
+ Returns TRUE if this robot is allowed to retrieve this URL.
+
+* $rules->agent([$name])
+
+ Get/set the agent name. NOTE: Changing the agent name will clear the
+ robots.txt rules and expire times out of the cache.
%prep
%setup -q -n %{cpan_name}-%{version}
+find . -type f -print0 | xargs -0 chmod 644
%build
%{__perl} Makefile.PL INSTALLDIRS=vendor
@@ -59,11 +83,8 @@
%perl_process_packlist
%perl_gen_filelist
-%clean
-%{__rm} -rf %{buildroot}
-
%files -f %{name}.files
-%defattr(644,root,root,755)
+%defattr(-,root,root,755)
%doc Changes README
%changelog
++++++ WWW-RobotRules-6.01.tar.gz -> WWW-RobotRules-6.02.tar.gz ++++++
diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/WWW-RobotRules-6.01/Changes new/WWW-RobotRules-6.02/Changes
--- old/WWW-RobotRules-6.01/Changes 2011-03-13 16:08:40.000000000 +0100
+++ new/WWW-RobotRules-6.02/Changes 2012-02-18 14:07:15.000000000 +0100
@@ -1,4 +1,11 @@
_______________________________________________________________________________
+2012-02-18 WWW-RobotRules 6.02
+
+Restore perl-5.8.1 compatiblity.
+
+
+
+_______________________________________________________________________________
2011-03-13 WWW-RobotRules 6.01
Added legal notice and updated the meta repository link
diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/WWW-RobotRules-6.01/META.yml new/WWW-RobotRules-6.02/META.yml
--- old/WWW-RobotRules-6.01/META.yml 2011-03-13 16:08:54.000000000 +0100
+++ new/WWW-RobotRules-6.02/META.yml 2012-02-18 14:07:29.000000000 +0100
@@ -1,6 +1,6 @@
--- #YAML:1.0
name: WWW-RobotRules
-version: 6.01
+version: 6.02
abstract: database of robots.txt-derived permissions
author:
- Gisle Aas