You will also want to run hardware tests on the machines in your cluster to make sure memory, disk, and network is working properly. We use tools such as memtest86+ and Doug Ledfors Memory Test Script to do burn ins. Here is a link to different test programs for linux.

http://linuxquality.sunsite.dk/articles/testsuites/

Dennis Kubes

Owen O'Malley wrote:

On Apr 23, 2007, at 7:39 AM, Steve Schlosser wrote:

I've got a small hadoop cluster running (5 nodes today, going to 15+
soon), and I'd like to do some benchmarking.  My question to the group
is - what is the first benchmark you run on a new cluster?

I usually use random-writer to generate some random data (it defaults to 10g/node) and then use sort to sort it. Sort provides a pretty decent simple testcase for moving a lot of data through map/reduce.

-- Owen

Reply via email to