[ https://issues.apache.org/jira/browse/HDFS-16776?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17871534#comment-17871534 ]
ASF GitHub Bot commented on HDFS-16776: --------------------------------------- zhengchenyu commented on PR #6964: URL: https://github.com/apache/hadoop/pull/6964#issuecomment-2272574201 @tomscut @zhangshuyan0 @haiyang1987 I think the problem is that ec's code is too dependent on the original process, the original process is based on continuous block copy. Many bugs come from over-reliance on this process and unnecessary parameter passing. I submit [HDFS-17542](https://github.com/apache/hadoop/pull/6915), try to rearrange these code. After this pr, I think there are no need to check the length of target. Can you please review HDFS-17542? As for the maintenance, maybe I think it is not just maintainance. There are some imperfections in calculating the ec replica state, mainly because the uniqueness of the internal blocks is not taken into account when calculating NumberReplicas. Although some issues attempt to solve these problems (such as the inaccurate calculation of DECOMMISSIONING solved by HDFS-14920), it is not thorough. [HDFS-17542](https://github.com/apache/hadoop/pull/6915) introduce NumberReplicasStriped, I think it is a better way. > Erasure Coding: The length of targets should be checked when DN gets a > reconstruction task > ------------------------------------------------------------------------------------------ > > Key: HDFS-16776 > URL: https://issues.apache.org/jira/browse/HDFS-16776 > Project: Hadoop HDFS > Issue Type: Improvement > Components: erasure-coding > Affects Versions: 3.4.0, 3.3.5 > Reporter: Ruinan Gu > Assignee: Ruinan Gu > Priority: Major > Labels: pull-request-available > Fix For: 3.4.0, 3.3.5, 3.2.5 > > > The length of targets should be checked when DN gets a EC reconstruction > task.For some reason (HDFS-14768, HDFS-16739) , the length of targets will be > larger than additionalReplRequired which causes some elements in targets get > the default value 0. It may trigger the bug which leads to the data > corrupttion just like HDFS-14768. -- This message was sent by Atlassian Jira (v8.20.10#820010) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org