lance bowler wrote on 12/10/11 10:01 AM:
> Hi,
> 
> Trying things out and a few quick questions:
> 
> - is it possible to collapse duplicate results based on a field?  eg, you
> may have many results from a particular website, so I may want to collapse
> these and display a More link.  Can do?

No existing feature for that afaik. If I were attempting it, I would probably
define some kind of unique site id in my docs (like a base URI) as a field and
then include that in my sort_spec. Then could prune results based on that field.

> 
> - I want to do something similar to Google's Cached page:  ie, display a
> website page from my store with the search terms highlighted (ie, not using
> Lucy's normal excerpt/highlighter since the page is not coming from the
> index itself, but from cached pages).  Lucy's highlighter does a great job
> of highlighting words based on stemming, can this somehow be hooked into to
> display a page from disk?

Check out HTML::HiLiter and/or Search::Tools::HiLiter (which HTML::HiLiter uses
underneath). I wrote that explicitly for the purpose you're describing.


> 
> - I read somewhere on the list about concurrency maybe being a problem (ie,
> remote cluster searchers can only perform 1 search at a time).  surely if
> you have 10 or more visitors performing parallel searches they'll block
> until each one completes...?

If I'm understanding your concern correctly, I don't think concurrency as you're
describing it is a problem. Multiple Searchers can be open against the same
index simultaneously, each handling parallel searches. The issue you might be
referring to with the now-deprecated LucyX::Remote::Search(Server|Client) was
that requests to the server were being executed serially rather than in
parallel. That issue should be addressed now in trunk with the new 
ClusterSearcher.

-- 
Peter Karman  .  http://peknet.com/  .  [email protected]

Reply via email to