Does commodity hardware come with ECC memory? Since Hadoop apps tend to move large amounts of data around, ECC memory seems pretty important.
With just two machines, you might be limited since you need to run multiple components (Namenode, job tracker, etc) on one machine. Two machines seems quite low... I would go with 10, if cost is same. Also 10 machines might eat up more power.
Raghu. Chris Fellows wrote:
Hello, Much of the hadoop documentation speaks to large clusters of commodity machines. There is a debate on our end about which would be better: a small number of high performance machines (2 boxes with 4 quad core processors) or X number of commodity machines. I feel that disk I/O might be the bottle neck with the 2 high perf machines (though I did just read in the FAQ about being able to split the dfs-data across multiple drives). So this is a "which would rather" question. If you were setting up a cluster of machines to perform data rollups/aggregation (and other mapred tasks) on files in the .25-1TB size, which would rather have: 1. 2 4 quad core machines with your choice on RAM and number of drives 2. 10 (or more) commodity machines (as defined on the hadoop wiki) And of course a "why?" would be very helpful. Thanks!