HDFS should repair corrupted files
----------------------------------
Key: HADOOP-855
URL: https://issues.apache.org/jira/browse/HADOOP-855
Project: Hadoop
Issue Type: Bug
Components: dfs
Reporter: Wendy Chien
Assigned To: Wendy Chien
While reading if we discover a mismatch between a block and checksum, we want
to report this back to the namenode to delete the corrupted block or crc.
To implement this, we need to do the following:
DFSInputStream
1. move DFSInputStream out of DFSClient
2. add member variable to keep track of current datanode (the chosen node)
DistributedFileSystem
1. change reportChecksumFailure parameter crc from int to FSInputStream (needed
to be able to delete it).
2. determine specific block and datanode from DFSInputStream passed to
reportChecksumFailure
3. call namenode to delete block/crc vis DFSClient
ClientProtocol
1. add method to ask namenode to delete certain blocks on specifc datanode.
Namenode
1. add ability to delete certain blocks on specific datanode
--
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators:
https://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see: http://www.atlassian.com/software/jira