lance bowler wrote on 12/10/11 10:01 AM: > Hi, > > Trying things out and a few quick questions: > > - is it possible to collapse duplicate results based on a field? eg, you > may have many results from a particular website, so I may want to collapse > these and display a More link. Can do?
No existing feature for that afaik. If I were attempting it, I would probably define some kind of unique site id in my docs (like a base URI) as a field and then include that in my sort_spec. Then could prune results based on that field. > > - I want to do something similar to Google's Cached page: ie, display a > website page from my store with the search terms highlighted (ie, not using > Lucy's normal excerpt/highlighter since the page is not coming from the > index itself, but from cached pages). Lucy's highlighter does a great job > of highlighting words based on stemming, can this somehow be hooked into to > display a page from disk? Check out HTML::HiLiter and/or Search::Tools::HiLiter (which HTML::HiLiter uses underneath). I wrote that explicitly for the purpose you're describing. > > - I read somewhere on the list about concurrency maybe being a problem (ie, > remote cluster searchers can only perform 1 search at a time). surely if > you have 10 or more visitors performing parallel searches they'll block > until each one completes...? If I'm understanding your concern correctly, I don't think concurrency as you're describing it is a problem. Multiple Searchers can be open against the same index simultaneously, each handling parallel searches. The issue you might be referring to with the now-deprecated LucyX::Remote::Search(Server|Client) was that requests to the server were being executed serially rather than in parallel. That issue should be addressed now in trunk with the new ClusterSearcher. -- Peter Karman . http://peknet.com/ . [email protected]
