That's the paradigm of Hadoop's Map-Reduce. 

> -----Original Message-----
> From: Samuel LEMOINE [mailto:[EMAIL PROTECTED] 
> Sent: Thursday, August 23, 2007 2:48 PM
> To: [email protected]
> Subject: "Moving Computation is Cheaper than Moving Data"
> 
> When I read the Hadoop documentation:
> The Hadoop Distributed File System: Architecture and Design
> (http://lucene.apache.org/hadoop/hdfs_design.html)
> 
> a paragraph hold my attention:
> 
> 
>       "Moving Computation is Cheaper than Moving Data"
> 
> A computation requested by an application is much more 
> efficient if it is executed near the data it operates on. 
> This is especially true when the size of the data set is 
> huge. This minimizes network congestion and increases the 
> overall throughput of the system. The assumption is that it 
> is often better to migrate the computation closer to where 
> the data is located rather than moving the data to where the 
> application is running. HDFS provides interfaces for 
> applications to move themselves closer to where the data is located.
> 
> 
> 
> 
> I'd like to know how to perform that, espacially with the aim 
> of distributed Lucene search ? Which Hadoop classes should I 
> use to do that ?
> 
> Thanks in advance,
> 
> Samuel
> 

Reply via email to