I seem to have been able to fix this by using the server_aliases directive,
which is much better than blocking the URL, since the document(s) returned is
valid, and should be indexed, not ignored.

-----Original Message-----
From: Jim [mailto:[EMAIL PROTECTED] 
Sent: 28 September 2004 04:21
To: BROCKINGTON, Michael -Syntegra UK
Cc: [EMAIL PROTECTED]
Subject: Re: [htdig] Mangled Domain Info


On Thu, 23 Sep 2004 [EMAIL PROTECTED] wrote:

> I have just noticed that my new 3.1.6 setup is giving me duplicate 
> results for many files. The reason for this is that in several places 
> in my site, users have managed to put a final dot at the end of the 
> domain portion of the URL, ie: my.valid.domain./some/file.name

You might try adding something like .org./ .com./ .net./ etc. to the
exclude_urls attribute.

Jim


********************************************************************

This email may contain information which is privileged or confidential. If you are not 
the intended recipient of this email, please notify the sender immediately and delete 
it without reading, copying, storing, forwarding or disclosing its contents to any 
other person
Thank you

Check us out at http://www.btsyntegra.com

********************************************************************



-------------------------------------------------------
This SF.Net email is sponsored by: YOU BE THE JUDGE. Be one of 170
Project Admins to receive an Apple iPod Mini FREE for your judgement on
who ports your project to Linux PPC the best. Sponsored by IBM.
Deadline: Sept. 24. Go here: http://sf.net/ppc_contest.php
_______________________________________________
ht://Dig general mailing list: <[EMAIL PROTECTED]>
ht://Dig FAQ: http://htdig.sourceforge.net/FAQ.html
List information (subscribe/unsubscribe, etc.)
https://lists.sourceforge.net/lists/listinfo/htdig-general

Reply via email to