Hi,
I'm trying to implement a duplicate detection method that doesn't delete
duplicate pages from the index. Essentially, I want to be able to
display all the duplicate URLs for a page in the search results instead
of just the one that was kept in the index.
There are two (potentially more) ways that I can think of to implement this.
1. Offline duplicate detection which deletes the pages from the index
but stores references to the deleted pages with the copy that is kept.
The search results can then display all the URLs that have the same content.
2. Duplicate detection at search time that groups identical/similar
pages together. This method has the advantage that one could implement
duplicate detection that is sensitive to the query terms. However, it
would add a performance penalty to the search.
I not very familiar with the Nutch API though I know there's a MD5
signature based deduping method in place and a Signature class to extend
for offline duplicate detection. I was wondering if anyone had tried
search time deduping and what would be good places to try and implement it.
Any other suggestions/advice would be great.
Thanks,
- Shailesh
-------------------------------------------------------
This SF.Net email is sponsored by xPML, a groundbreaking scripting language
that extends applications into web and mobile media. Attend the live webcast
and join the prime developer group breaking into this new coding territory!
http://sel.as-us.falkag.net/sel?cmd=lnk&kid=110944&bid=241720&dat=121642
_______________________________________________
Nutch-developers mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/nutch-developers