Add a test to make sure that node decomission does not get blocked by 
underreplicated blocks in an unclosed file
----------------------------------------------------------------------------------------------------------------

                 Key: HDFS-694
                 URL: https://issues.apache.org/jira/browse/HDFS-694
             Project: Hadoop HDFS
          Issue Type: Test
    Affects Versions: 0.21.0
            Reporter: Hairong Kuang
             Fix For: 0.21.0


We have a cluster that took  much longer time to decommission datanodes than 
usual. It turned out to be caused by an open file. In HDFS 0.20 when a file is 
open, NN does not schedule any block belonged to this file to replicate. So if 
an open file has an under-replicated block with a replica on a decommission 
datanode, the decommission won't complete until the file is closed.

The new append implementation changed the replication strategy that any 
complete block in an unclosed file will be scheduled to be replicated if it 
becomes under-replicated. This jira aims to add a test to make sure that node 
decommission does not get blocked by under-replicated blocks in an open file.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to