[ 
https://issues.apache.org/jira/browse/HDFS-9719?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15188354#comment-15188354
 ] 

Kai Zheng commented on HDFS-9719:
---------------------------------

Thanks [~umamaheswararao] for the review and great comments!
bq. I am seeing some variable names still like “toRecoverLen”. Can we take 
chance in this patch to change them like toReconstructLen ?
Yes. I will fix the variable and check one more time for such instances.
bq. doReadMinimum : this method name looks to be wrong. Its actually reading 
from minimum required sourced data nodes. But this name looks like it is 
reading minimum data length/what?
The concern seems reasonable. I will rename: doReadMinimum => 
doReadMinimumSources, and the same, readMinimum => readMinimumSources.
bq. Also I assume StripedReader itself should handle multiple chunk/cell 
readers.
Yeah, *Striped* should itself already be of the meaning of handling multiple 
units, source datanodes or target datanodes.
bq. how about renaming class name like StripedReaders -> StripedReader and 
StripedReader -> StripedChunkReader and the same comment applies for 
StripedWriter*
Pretty good suggestions for more readable component names! Thanks a lot!

Will update the patch accordingly soon. ...

> Refactoring ErasureCodingWorker into smaller reusable constructs
> ----------------------------------------------------------------
>
>                 Key: HDFS-9719
>                 URL: https://issues.apache.org/jira/browse/HDFS-9719
>             Project: Hadoop HDFS
>          Issue Type: Sub-task
>            Reporter: Kai Zheng
>            Assignee: Kai Zheng
>         Attachments: HDFS-9719-v1.patch, HDFS-9719-v2.patch, 
> HDFS-9719-v3.patch, HDFS-9719-v4.patch, HDFS-9719-v5.patch, HDFS-9719-v6.patch
>
>
> This would suggest and refactor {{ErasureCodingWorker}} into smaller 
> constructs to be reused in other places like block group checksum computing 
> in datanode side. As discussed in HDFS-8430 and implemented in HDFS-9694 
> patch, checksum computing for striped block groups would be distributed to 
> datanode in the group, where data block data should be able to be 
> reconstructed when missed/corrupted to recompute the block checksum. The most 
> needed codes are in the current ErasureCodingWorker and could be reused in 
> order to avoid duplication. Fortunately, we have very good and complete 
> tests, which would make the refactoring much easier. The refactoring will 
> also help a lot for subsequent tasks in phase II for non-striping erasure 
> coded files and blocks. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to