Hi,

I'm curious about what Tomislav is asking about, too -- how do searchers know 
when to reopen the index?  That is, say you have a cluster of fetchers and 
every once in a while you end up with a newer version of an index (or indices), 
and say that you simply scp those indices to searchers, how do you signal the 
searcher webapps to go and reopen the new index?

In Solr-land, for example, this is done by issuing a "commit" command, which 
tells the Solr IndexSearcher to, among other things, reopen the index.  In pure 
Lucene-land, you check the index version via IndexReader.  How about in 
Nutchlandia?

Also, is scp-ing/rsyncing the index over to searcher boxes the way to go?  I 
didn't see this covered on the Wiki, but maybe I didn't search well enough? ;)

Thanks,
Otis

----- Original Message ----
From: Tomislav Poljak <[EMAIL PROTECTED]>
To: [email protected]
Sent: Friday, September 7, 2007 6:52:58 PM
Subject: Re: help with hardware requirements

Hi,
what would be a recommended hardware specification for a machine running
searcher web application with 15K users per day witch uses index of 100K
urls (crawling is done by other machine)? What is a good practice for
getting index from crawl machine to search machine (if using separate
machines for crawl and search)?

Thanks,
      Tomislav




Reply via email to