dfs.hosts.exclude not working as expected
-----------------------------------------
Key: HADOOP-3433
URL: https://issues.apache.org/jira/browse/HADOOP-3433
Project: Hadoop Core
Issue Type: Bug
Components: dfs
Affects Versions: 0.16.3
Reporter: Christian Kunz
We had to decommission a lot of hosts.
Therefore, we added them to dfs.hosts.exclude and called 'dfsadmin
-refreshNodes'.
The list of excluded nodes appeared in the list of dead nodes (and still
remained the list of live nodes), but there was no replication going on for
more than 20 hours (no NameSystem.addStoredBlock messages in the namenode log).
A few hours ago we stopped a datanode from that list. After it moved from the
live node list to the dead node list (double entry), replication started
immediately and completed after about 1 hour (replicated ~ 10,000 blocks).
Somehow, mere exclusion does not trigger replication as it should.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.