On Fri, 2009-03-13 at 19:42 -0700, buddha1021 wrote:
hi dennis:
...
> I am confident that hadoop can process the large datas of the www search
engine! But lucene? I am afraid of the limited size of lucene's index per
server is very little ,10G? or 30G? this is not enough for the www search
> engine! IMO, this is a bottleneck!
I agree that the actual problem/solution of accessing lucene indexes is
to keep them small. What does the possibility of having a clouded index
serve if accessing it takes hours?
For me here should lie one of nutch core competences: making search in
BIG indexes fast (as fast as in SMALL indexes).
I would suggest looking at Katta (http://katta.sourceforge.net/).
It's one of several projects where the goal is to support very large
Lucene indexes via distributed shards. Solr has also added federated
search support.
-- Ken
--
Ken Krugler
+1 530-210-6378