A block report processing may incorrect cause the namenode to delete blocks
----------------------------------------------------------------------------
Key: HADOOP-1135
URL: https://issues.apache.org/jira/browse/HADOOP-1135
Project: Hadoop
Issue Type: Bug
Components: dfs
Reporter: dhruba borthakur
Assigned To: dhruba borthakur
When a block report arrives at the namenode, the namenode goes through all the
blocks on that datanode. If a block is not valid it is marked for deletion. The
blocks-to-be-deleted are sent to the datanode as a response to the next
heartbeat RPC. The namenode sends only 100 blocks-to-be-deleted at a time. This
was introduced as part of hadoop-994. The bug is that if the number of
blocks-to-be-deleted exceeds 100, then that namenode marks all the remaining
blocks in the block report for deletion.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.