Dear Wiki user, You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change notification.
The following page has been changed by xavierstevens: http://wiki.apache.org/hadoop/PoweredBy ------------------------------------------------------------------------------ * We use Hadoop to store copies of internal log and dimension data sources and use it as a source for reporting/analytics and machine learning. * Currently have a 320 machine cluster with 2,560 cores and about 1.3 PB raw storage. Each (commodity) node has 8 cores and 4 TB of storage. * We are heavy users of both streaming as well as the Java apis. We have built a higher level data warehousing framework using these features called Hive (see the [https://issues.apache.org/jira/browse/HADOOP-3601 JIRA ticket]). We have also written a read-only FUSE implementation over hdfs. + + * [http://www.newscorp.com/management/fim.html FOX Interactive Media] + * 20 machine cluster (8 cores/machine, 1TB/machine storage) + * 10 machine cluster (8 cores/machine, 1TB/machine storage) + * In process of creating a third 50-node cluster + * Use for log analysis, data mining and machine learning + * [http://www.hadoop.co.kr/ Hadoop Korean User Group], a Korean Local Community Team Page. * 50 node cluster In the Korea university network environment. * Pentium 4 PC, HDFS 4TB Storage * Used for development projects * Retrieving and Analyzing Biomedical Knowledge * Latent Semantic Analysis, Collaborative Filtering - - * [http://www.newscorp.com/management/fim.html FOX Interactive Media] - * 12 node cluster (8 cores/node, 1TB/node storage) - * 10 node cluster (8 cores/node, 1TB/node storage) - * In process of creating a third 50-node cluster with more storage/node - * Use for log analysis, data mining and machine learning * [http://www.google.com Google] * [http://www.google.com/intl/en/press/pressrel/20071008_ibm_univ.html University Initiative to Address Internet-Scale Computing Challenges]
