Thanks Lohit. The key point I missed was that dfs.hosts.exclude should exist in before starting the namenode. It worked after restarting hdfs.

-Chris



On Jul 8, 2008, at 3:56 PM, lohit wrote:

there are few things which aren't documented.
- you should have defined full path of file as part of dfs.hosts.exclude before starting the namenode. This file must exists, could be a zero length file. - While the system is running, you add the hostname (fully qualified) to this file and then invoke hadoop dfsadmin -refreshNodes - You should have enough free datanodes in the cluster so that the blocks from this node can be replicated to other nodes. Eg if you replication factor is 3 its good to have atleast 4 datanodes before you decomission any one of them. - dfs.namenode.decommission.interval defines interval at which namenode checks if decomission is complete, after which it removes from its list and out of service.

PS : IP address instead of hostname in excludes file should also work.

Thanks,
Lohit

----- Original Message ----
From: Chris Kline <[EMAIL PROTECTED]>
To: core-user@hadoop.apache.org
Sent: Tuesday, July 8, 2008 9:38:36 AM
Subject: Cannot decommission on 16.4

I followed the instruction on the wiki and searched JIRA tickets for
more info, and still cannot decommission a node on 16.4.  I've tried
different combinations of hostnames and IPs and nothing seems to work.

Has anyone successfully decommissioned a node on 16.4?  If so, was
there some trick?  I'm using the exclude method.

-Chris

Reply via email to