Dear Wiki user, You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change notification.
The following page has been changed by JamesOmdahl: http://wiki.apache.org/hadoop/PoweredBy ------------------------------------------------------------------------------ * [http://www.weblab.infosci.cornell.edu/ Cornell University Web Lab] * Generating web graphs on 100 nodes (dual 2.4GHz Xeon Processor, 2 GB RAM, 72GB Hard Drive) + + * [http://www.enormo.com/ Enormo] + * 4 nodes cluster (32 cores, 1TB). + * We use Hadoop to filter and index our listings, removing exact duplicates and grouping similar ones. + * We plan to use Pig very shortly to produce statistics. * [http://www.systems.ethz.ch/education/courses/hs08/map-reduce/ ETH Zurich Systems Group] * We are using Hadoop in a course that we are currently teaching: "Massively Parallel Data Analysis with MapReduce". The course projects are based on real use-cases from biological data analysis. @@ -156, +161 @@ * Using Hadoop on EC2 to process documents from a continuous web crawl and distributed training of support vector machines * Using HDFS for large archival data storage - * [http://www.properazzi.com/ Properazzi] - * 4 nodes cluster (32 cores, 1TB). - * We use Hadoop to filter and index our listings, removing exact duplicates and grouping similar ones. - * We plan to use Pig very shortly to produce statistics. - * [http://www.quantcast.com/ Quantcast] * 3000 cores, 3500TB. 1PB+ processing each day. * Hadoop scheduler with fully custom data path / sorter
