[ 
https://issues.apache.org/jira/browse/HDFS-12923?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16290087#comment-16290087
 ] 

Lei (Eddy) Xu commented on HDFS-12923:
--------------------------------------

[~elgoiri] yes, sure.  In {{FSDirConcatOp#verifySrcFiles()}}, it has already 
checked the EC policy ID between files, so a {{HadoopIllegalArgumentException}} 
will be thrown.

{code:title=FSDirConcatOp.java}
private static INodeFile[] verifySrcFiles(...) {
....
 if(srcINodeFile.getErasureCodingPolicyID() !=
          targetINode.getErasureCodingPolicyID()) {
        throw new HadoopIllegalArgumentException("Source file " + src
            + " and target file " + targetIIP.getPath()
            + " have different erasure coding policy");
      }
}
{code}

> DFS.concat should throw exception if files have different EC policies. 
> -----------------------------------------------------------------------
>
>                 Key: HDFS-12923
>                 URL: https://issues.apache.org/jira/browse/HDFS-12923
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: erasure-coding
>    Affects Versions: 3.0.0
>            Reporter: Lei (Eddy) Xu
>            Priority: Critical
>             Fix For: 3.0.0
>
>
> {{DFS#concat}} appends blocks from different files to a single file. However, 
> if these files have different EC policies, or mixed with replicated and EC 
> files, the resulted file would be problematic to read, because the EC codec 
> is defined in INode instead of in a block. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org

Reply via email to