Just wondering, is there an easy way to load the whole index into ram? On Wed, Sep 9, 2009 at 4:22 PM, Alex Baranov <alex.barano...@gmail.com>wrote:
> There is a good article on how to scale the Lucene/Solr solution: > > > http://www.lucidimagination.com/Community/Hear-from-the-Experts/Articles/Scaling-Lucene-and-Solr > > Also, if you have heavy load on the server (large amount of concurrent > requests) then I'd suggest to consider loading the index into RAM. It > worked > well for me on the project with 140+ million documents and 30 concurrent > user requests per second. If your index can be placed in RAM you can reduce > the architecture complexity. > > Alex Baranov > > On Wed, Sep 9, 2009 at 5:10 PM, Elaine Li <elaine.bing...@gmail.com> > wrote: > > > Hi, > > > > I have 20 million docs on solr. If my query would return more than > > 10,000 results, the response time will be very very long. How to > > resolve such problem? Can I slice my docs into pieces and let the > > query operate within one piece at a time so the response time and > > response data will be more managable? Thanks. > > > > Elaine > > >