dfs.hosts.exclude not working as expected
-----------------------------------------

                 Key: HADOOP-3433
                 URL: https://issues.apache.org/jira/browse/HADOOP-3433
             Project: Hadoop Core
          Issue Type: Bug
          Components: dfs
    Affects Versions: 0.16.3
            Reporter: Christian Kunz


We had to decommission a lot of hosts.

Therefore, we added them to dfs.hosts.exclude and called 'dfsadmin 
-refreshNodes'.
The list of excluded nodes appeared in the list of dead nodes (and still 
remained the list of live nodes), but there was no replication going on for 
more than 20 hours (no NameSystem.addStoredBlock messages in the namenode log).

A few hours ago we stopped a datanode from that list. After it moved from the 
live node list to the dead node list (double entry), replication started 
immediately and completed after about 1 hour (replicated ~ 10,000 blocks).

Somehow, mere exclusion does not trigger replication as it should.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to