There is a url normalizing feature in

        conf/regex-normalize.xml

For example I used the following pattern to say that pages with &limit and &limitstart are the same on a JasperSoft forum

<!-- Jasper normalize jaspersoft URLs:
     (1) &limit=6&limitstart=0 means the same as the page w/o any limit
     (2) catid=10&id=NNN means the same as id=NNN&catid=10  -->
<regex-normalize>
<regex>
<pattern>(\?|\&amp;|\&amp;amp;)limit=6(\&amp;|\&amp;amp;) limitstart=0$</pattern>
  <substitution></substitution>
</regex>
<regex>
<pattern>(\?|\&amp;|\&amp;amp;)(id=[0-9]+)(\&amp;|\&amp;amp;) (catid=10)(.*)</pattern>
  <substitution>$1$4$3$2$5</substitution>
</regex>
</regex-normalize>


On Oct 8, 2008, at 4:23 AM, Detlef Müller-Solger wrote:

Hi,

in Germany it is reported, that one big show stopper for Nutch is the fact, that there for example are often identical webpage’s which can be addressed by different URLs. For example by requesting

www.xyz.de/information
or by
www.xyz.de/information/
or by
www.xyz.de/information/index

From my point of view due to the different URLs Nutch is indexing those webpages unfortuneately three times. Is there a method to avoid the indexing of these doublets? For example by comparing all information of the webpage excluding the URL.

Note: A Fliter like "reduce URL generally of "/index"" is no solution because in other cases of the same run "/index" may be needed or the same Webpage can be adressed also by other URL Syntax.

Thanx

Detlef Müller-Solger



Reply via email to