I seem to have been able to fix this by using the server_aliases directive, which is much better than blocking the URL, since the document(s) returned is valid, and should be indexed, not ignored.
-----Original Message----- From: Jim [mailto:[EMAIL PROTECTED] Sent: 28 September 2004 04:21 To: BROCKINGTON, Michael -Syntegra UK Cc: [EMAIL PROTECTED] Subject: Re: [htdig] Mangled Domain Info On Thu, 23 Sep 2004 [EMAIL PROTECTED] wrote: > I have just noticed that my new 3.1.6 setup is giving me duplicate > results for many files. The reason for this is that in several places > in my site, users have managed to put a final dot at the end of the > domain portion of the URL, ie: my.valid.domain./some/file.name You might try adding something like .org./ .com./ .net./ etc. to the exclude_urls attribute. Jim ******************************************************************** This email may contain information which is privileged or confidential. If you are not the intended recipient of this email, please notify the sender immediately and delete it without reading, copying, storing, forwarding or disclosing its contents to any other person Thank you Check us out at http://www.btsyntegra.com ******************************************************************** ------------------------------------------------------- This SF.Net email is sponsored by: YOU BE THE JUDGE. Be one of 170 Project Admins to receive an Apple iPod Mini FREE for your judgement on who ports your project to Linux PPC the best. Sponsored by IBM. Deadline: Sept. 24. Go here: http://sf.net/ppc_contest.php _______________________________________________ ht://Dig general mailing list: <[EMAIL PROTECTED]> ht://Dig FAQ: http://htdig.sourceforge.net/FAQ.html List information (subscribe/unsubscribe, etc.) https://lists.sourceforge.net/lists/listinfo/htdig-general