M, See http://wiki.apache.org/hadoop/FAQ - "3.6. I want to make a large cluster smaller by taking out a bunch of nodes simultaneously. How can this be done?"
This explains how to decomission nodes by moving the data off of the existing node. It's fairly easy and painless (just add the nodename to the slaves.exclude file and notify dfs) and once the data is off the node you could swap-out the disks and then re-introduce the node back into the cluster with larger drives (removing the nodename from slaves.exclude). Chris On 17 May 2012 02:55, Mayuran Yogarajah <mayuran.yogara...@casalemedia.com>wrote: > Our cluster has several nodes which have smaller disks than other nodes and > as a result fill up quicker. > > I am looking to move data off these nodes and onto the others. > > > > Here is what I am planning to do: > > 1) On the nodes with smaller disks, set dfs.datanode.du.reserved to a > larger value > > 2) Restart data nodes > > 3) Run balancer > > > > Will this have the desired effect? > > If there is a better way to accomplish this please let me know. > > > > Thanks, > > M > >