I think nutch has a distributed lucene implementation. I could have used nutch straightaway, but I have a different crawler, and also dont want to use NDFS(which is used by nutch) . What I have proposed earlier is basically based on mapReduce paradigm, which is used by nutch as well.

It would be nice to get some articles specifically detailing out the distributed architecture used in nutch.

prasen

Samuru Jackson wrote:

Does it make any sense ? Also would like to know if there are other ways
to distribute lucene's indexing/searching  ?

I'm interested in such a distributed architecture too.

What I have got in mind is some kind of lucene index cluster where you
have got several machines having subindexes in memory. So if you have
got a a searchquery the machines should perfom fast because the index
is in memory and no hard disk access is performed.

Is there anything like this available?



---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to