Shouldn't this be targeted for 0.14?

Doug

Hairong Kuang (JIRA) wrote:
     [ 
https://issues.apache.org/jira/browse/HADOOP-1124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HADOOP-1124:
----------------------------------

    Priority: Major  (was: Blocker)

ChecksumFileSystem does not handle ChecksumError correctly
----------------------------------------------------------

                Key: HADOOP-1124
                URL: https://issues.apache.org/jira/browse/HADOOP-1124
            Project: Hadoop
         Issue Type: Bug
         Components: fs
   Affects Versions: 0.12.0
           Reporter: Hairong Kuang
        Assigned To: Hairong Kuang
            Fix For: 0.13.0


When handle ChecksumError, the checksumed file system tries to recover by 
rereading from a different replica.
I have three comments:
1. One bug in the code is that when retrying, the object that computes checksum 
does not get restored to the old state.
2. The code also assumes that the first byte read and the byte being read when ChecksumError occurs are in the same block. 3. It would be more efficient if we roll back to the first byte in the chunk that's being checksumed instead of rolling back to the first byte that was read.

Reply via email to