[ 
https://issues.apache.org/jira/browse/HADOOP-3500?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley resolved HADOOP-3500.
-----------------------------------

    Resolution: Duplicate

I suspect the browser hiccuped and you got two instances of the same bug. 
(HADOOP-3499)

> decommission node is both in the "Live Datanodes" with "In Service" status, 
> and in the "Dead Datanodes" of the dfs namenode web ui.
> -----------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: HADOOP-3500
>                 URL: https://issues.apache.org/jira/browse/HADOOP-3500
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.17.0
>         Environment: linux-2.6.9
>            Reporter: lixiangna
>   Original Estimate: 72h
>  Remaining Estimate: 72h
>
> try to decommission a node by the following the steps:
> (1) write the hostname of node which will be decommissioned in a file (the 
> exclude file)
> (2) specified the absolute path of the exclude file as a configuration 
> parameter dfs.hosts.exclude.
> (3) run "bin/hadoop dfsadmin -refreshNodes".
> It is surprising that the node is found both in the "Live Datanodes" with "In 
> Service" status, and in the "Dead Datanodes" of the dfs namenode web ui. When 
> copy new data to the HDFS, its Used size is increasing as other 
> un-decommissioned nodes. Obviously it is in service. Restarting the HDFS or 
> waiting a long time(two day) havn't make the decommission yet.
> the more strange thing, If nodes are configured as the include nodes by 
> similar steps, then these include nodes and
> the exclude node are all only in the "Dead Datanodes" lists. 
> I did many times tests in both 0.17.0 and 0.15.1. The results is same. So i 
> think there maybe bugs.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to