namenode should reject datanodes which send impossible block reports
--------------------------------------------------------------------

                 Key: HDFS-1282
                 URL: https://issues.apache.org/jira/browse/HDFS-1282
             Project: Hadoop HDFS
          Issue Type: Bug
          Components: data-node, name-node
    Affects Versions: 0.20.1
            Reporter: Andrew Ryan


Over the past few weeks we've had several datanodes with bad disks that suffer 
ext3 corruption, and consequently start reporting impossible values for how 
full they are. This particular node, for example, has a configured capacity of 
10.86TB but reports 1733.95TB used, for a total of 15973.57% utilization.

Node     Last Contact   Admin State     Configured Capacity (TB)        Used 
(TB)       Non DFS  Used (TB)      Remaining  (TB)         Used  (%)       Used 
 (%)       Remaining (%)   Blocks 
hadoop2254       44     In Service      10.86   1733.95 0       5.24    
15973.57   48.25        65602 

If we can avoid generating such bogus data on the datanode that would be great. 
 But if the namenode receives such an impossible block report, it should 
definitely consider that datanode to be not trustworthy, and in my opinion, 
make it dead.

The "fix" in our case was either to fsck or replace the bad disk.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to