I think nutch has a distributed lucene implementation. I could have used
nutch straightaway, but I have a different crawler, and also dont want
to use NDFS(which is used by nutch) . What I have proposed earlier is
basically based on mapReduce paradigm, which is used by nutch as well.
It would be nice to get some articles specifically detailing out the
distributed architecture used in nutch.
prasen
Samuru Jackson wrote:
Does it make any sense ? Also would like to know if there are other ways
to distribute lucene's indexing/searching ?
I'm interested in such a distributed architecture too.
What I have got in mind is some kind of lucene index cluster where you
have got several machines having subindexes in memory. So if you have
got a a searchquery the machines should perfom fast because the index
is in memory and no hard disk access is performed.
Is there anything like this available?
---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]