Laurent Laborde wrote:
> 
> On Fri, Dec 26, 2008 at 2:35 AM, buddha1021 <[email protected]> wrote:
>>
>> hi all:
>>  I am very interested in the nutch! I want to ask some questions about
>> nutch:
>>  (1)Can nutch search 1 billion(=1000 millions) pages that the size of the
>> page's data will achive 10T(=10000G) bytes? one page's size ==10k .
> 
> Nutch use hadoop, which rely on HDFS to store it's data.
> and HDFS can certainly handle 10TB.
> ..........................................................................
> I know hadoop can handle 10TB,but this only means that nutch can store
> 10TB datas.when people use nutch to search,the speed of the search is also
> affected by lucene,can lucene search 10TB datas based on the hadoop?
> especially the speed mustn't be slow!
> ..............................................................................................................
>>  (2)If nutch can do this ,what about the speed of the search ,compared
>> with
>> google ? Can the speed of the search meet the people's requirement ?
> 
> I don't know. you should take a look at lucence performance.
> http://www.google.fr/search?q=lucene+performance
> ..................................................................................
> For nutch ,lucene use hadoop's RPC.call to search,so,lucene itself's
> performance is not enough.
> ...................................................................................
>>  (3)If nutch can do this ,how many nodes would be required?
> 
> i'd like to know too :)
> .........................................................................
> If one node has 4G ram and 260G disk,how many nodes woud be required?
> thank you!:)
> -- 
> F4FQM
> Kerunix Flan
> Laurent Laborde
> 
> 
:-):-)
-- 
View this message in context: 
http://www.nabble.com/the-question-of-the-nutch%27s-ability%21-tp21171116p21171888.html
Sent from the Nutch - User mailing list archive at Nabble.com.

Reply via email to