hi,Joarder Kamal: The following steps may help. 1. Stop your datanode daemon which you want to add a new HDD. 2. modify the data.dfs.dir property in hdfs-site.xml as mentioned by Geelong Yao. 3. start the DataNode daemon again.
Ling Kun On Mon, May 6, 2013 at 12:30 AM, Rahul Bhattacharjee < [email protected]> wrote: > I think the question here is as how to add new HDD volumn into an already > existing formatted HDFS cluster. > Not sure , by just adding the directory in data.dfs.dir would help. > > > On Fri, May 3, 2013 at 3:28 PM, Håvard Wahl Kongsgård < > [email protected]> wrote: > >> go for ext3 or ext4 >> >> >> On Fri, May 3, 2013 at 8:32 AM, Joarder KAMAL <[email protected]> wrote: >> >>> Hi, >>> >>> I have a running HDFS cluster (Hadoop/HBase) consists of 4 nodes and >>> the initial hard disk (/dev/vda1) size is 10G only. Now I have a second >>> hard drive /dev/vdb of 60GB size and want to add it into my existing HDFS >>> cluster. How can I format the new hard disk (and in which format? XFS?) and >>> mount it to work with HDFS >>> >>> Default HDFS directory is situated in >>> /usr/local/hadoop-1.0.4/hadoop-datastore >>> And I followed this link for installation. >>> >>> http://ankitasblogger.blogspot.com.au/2011/01/hadoop-cluster-setup.html >>> >>> Many thanks in advance :) >>> >>> >>> Regards, >>> Joarder Kamal >>> >> >> >> >> -- >> Håvard Wahl Kongsgård >> Data Scientist >> Faculty of Medicine & >> Department of Mathematical Sciences >> NTNU >> >> http://havard.dbkeeping.com/ >> > > -- http://www.lingcc.com
