I used the XML, I think using HTTP is important.  No caching on the client yet, 
that is a good idea, however my personal goal is to have an index that is 
updated every 30 seconds or less and so am not sure about caching on the 
client.  The caching can be handled by the Solr servers, that should be fine.  
If it works correctly then the architecture is very simple requiring 2 layers.  
The first is a Solr layer, the second is the client layer essentially running 
many threads in parallel per request.  Seems like this would scale cheaply by 
adding more hardware on both layers.  

>  If you are using RMI you could
either borrow from or subclass Lucene's MultiSearcher that implements
this stuff.

Yeah this is the real issue, if there are any general outlines of the best way 
to do this with Solr.  Perhaps a separate Solr call for the docFreqs?  Or could 
this be returned in the current /select call?  I'm still trying to figure this 
part out.  

----- Original Message ----
From: Yonik Seeley <[EMAIL PROTECTED]>
To: solr-dev@lucene.apache.org
Sent: Thursday, May 18, 2006 5:21:58 PM
Subject: Re: Making RemoteSearchable like client for Solr

On 5/18/06, jason rutherglen <[EMAIL PROTECTED]> wrote:
> A solution to an index that requires 100s of millions of documents is to 
> distributed the documents over multiple servers.  I thought I had the 
> RemoteSearchable like client for Solr pretty well done

Great!  Can you share what approach you followed?
Is caching done on the subsearchers, and not the supersearcher?
Are you using RMI, or XML/HTTP?

> , however this normalized scoring with weights throws a bit of a kink.

It certainly does... not easy stuff.  If you are using RMI you could
either borrow from or subclass Lucene's MultiSearcher that implements
this stuff.


-Yonik



Reply via email to