Thanks Lohit. The key point I missed was that dfs.hosts.exclude
should exist in before starting the namenode. It worked after
restarting hdfs.
-Chris
On Jul 8, 2008, at 3:56 PM, lohit wrote:
there are few things which aren't documented.
- you should have defined full path of file as part of
dfs.hosts.exclude before starting the namenode. This file must
exists, could be a zero length file.
- While the system is running, you add the hostname (fully
qualified) to this file and then invoke hadoop dfsadmin -refreshNodes
- You should have enough free datanodes in the cluster so that the
blocks from this node can be replicated to other nodes. Eg if you
replication factor is 3 its good to have atleast 4 datanodes before
you decomission any one of them.
- dfs.namenode.decommission.interval defines interval at which
namenode checks if decomission is complete, after which it removes
from its list and out of service.
PS : IP address instead of hostname in excludes file should also work.
Thanks,
Lohit
----- Original Message ----
From: Chris Kline <[EMAIL PROTECTED]>
To: core-user@hadoop.apache.org
Sent: Tuesday, July 8, 2008 9:38:36 AM
Subject: Cannot decommission on 16.4
I followed the instruction on the wiki and searched JIRA tickets for
more info, and still cannot decommission a node on 16.4. I've tried
different combinations of hostnames and IPs and nothing seems to work.
Has anyone successfully decommissioned a node on 16.4? If so, was
there some trick? I'm using the exclude method.
-Chris