Hi!,all the steps should be done in the namenode. you can execute the "-report" twice one before the "-refreshNodes" and one later then compare the result!
jonhson.ian wrote: > > On Mon, Jun 1, 2009 at 12:35 AM, HRoger <hanxianyongro...@163.com> wrote: >> >> You should do thart in the right way as the follow steps: >> 1.create a new file named as excludes under $HADOOP_HOME with the >> datanode >> hostname(IP) in it by one name every line. >> 2.edit the hadoop-site.xml by adding >> <property> >> <name>dfs.hosts.exclude</name> >> <value>excludes</ value> >> </property> >> and save it. >> 3.execute the command "bin/hadoop dfsadmin -refreshNodes" in the namenode >> host. >> 4.when the step 3 finished,you can run "bin/hadoop dfsadmin -report and >> check the result. >> > > > I executed above steps all in Namenode and I got following message > (without restart hadoop): > > ----------------- dump of screeen ----------------------- > > $ bin/hadoop dfsadmin -refreshNodes > [had...@hdt0 hadoop-0.19.1]$ bin/hadoop dfsadmin -report > Safe mode is ON > Configured Capacity: 152863682560 (142.37 GB) > Present Capacity: 84421242880 (78.62 GB) > DFS Remaining: 84370862080 (78.58 GB) > DFS Used: 50380800 (48.05 MB) > DFS Used%: 0.06% > > ------------------------------------------------- > Datanodes available: 1 (3 total, 2 dead) > > Name: 10.61.0.5:50010 > Decommission Status : Decommission in progress > Configured Capacity: 152863682560 (142.37 GB) > DFS Used: 50380800 (48.05 MB) > Non DFS Used: 68442439680 (63.74 GB) > DFS Remaining: 84370862080(78.58 GB) > DFS Used%: 0.03% > DFS Remaining%: 55.19% > Last contact: Mon Jun 01 17:32:59 CST 2009 > > > Name: 10.61.0.7 > Decommission Status : Normal > Configured Capacity: 0 (0 KB) > DFS Used: 0 (0 KB) > Non DFS Used: 0 (0 KB) > DFS Remaining: 0(0 KB) > DFS Used%: 100% > DFS Remaining%: 0% > Last contact: Thu Jan 01 08:00:00 CST 1970 > > > Name: 10.61.0.143 > Decommission Status : Normal > Configured Capacity: 0 (0 KB) > DFS Used: 0 (0 KB) > Non DFS Used: 0 (0 KB) > DFS Remaining: 0(0 KB) > DFS Used%: 100% > DFS Remaining%: 0% > Last contact: Thu Jan 01 08:00:00 CST 1970 > > ----------------------------------------------------------------- > > two nodes has been dead... hmm... what happen? > and any help? > > > Thanks again, > > Ian > > -- View this message in context: http://www.nabble.com/DataNode-not-started-up-and-%22org.apache.hadoop.ipc.RemoteException%22--is-thrown-out-tp23791017p23812616.html Sent from the Hadoop core-user mailing list archive at Nabble.com.