For "fuzzy" you're going to pay one way or another.
You can use ngram analyzers on indexed content and queries which will
add IO costs ("files" becomes "fi","fil", "file","il","ile","iles" in
both your query and index) or you can use some form of query-time edit
distance comparison on "files" and pay the CPU costs. You can use
WordNet and "files" becomes "registers". You can examine large volumes
of user queries and look at what is the most likely interpretation. You
can use Soundex and then if you're lucky files==philes but there's no
room for error and they either match or they dont - there is no measure
of similarity.
There's no free lunch here.
Timo Nentwig wrote:
On Saturday 24 November 2007 18:28:48 markharw00d wrote:
term. You can limit the number of edit distance comparisons conducted by
setting the minimum prefix length. This is a property of the QueryParser
Well, javadoc: "prefixLength - length of common (non-fuzzy) prefix". So, this
is some kind of "wildcard fuzzy" but not real fuzzy anymore.
I understand the optimitation but right now I hardly can image a reasonable
use-case. Who care whether the levenstein distance is a the beginnen, middle
or end of word, .e.g when searching fuzzy for "philes" I want to
find "files"...
---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]