[ 
https://issues.apache.org/jira/browse/HDFS-795?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-795:
-----------------------------

          Component/s: hdfs client
             Priority: Critical  (was: Major)
    Affects Version/s: 0.20.1

Upgrading to critical since this is reproducible and causes complete pipeline 
failure for writers.

> DFS Write pipeline does not detect defective datanode correctly in some cases 
> (HADOOP-3339)
> -------------------------------------------------------------------------------------------
>
>                 Key: HDFS-795
>                 URL: https://issues.apache.org/jira/browse/HDFS-795
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: hdfs client
>    Affects Versions: 0.20.1
>            Reporter: Raghu Angadi
>            Priority: Critical
>             Fix For: 0.20.2
>
>         Attachments: toreproduce-5796.patch
>
>
> HDFS write pipeline does not select the correct datanode in some error cases. 
> One example : say DN2 is the second datanode and write to it times out since 
> it is in a bad state.. pipeline actually removes the first datanode. If such 
> a datanode happens to be the last one in the pipeline, write is aborted 
> completely with a hard error.
> Essentially the error occurs when writing to a downstream datanode fails 
> rather than reading. This bug was actually fixed in 0.18 (HADOOP-3339). But 
> HADOOP-1700 essentially reverted it. I am not sure why.
> It is absolutely essential for HDFS to handle failures on subset of datanodes 
> in a pipeline. We should not have at least known bugs that lead to hard 
> failures.
> I will attach patch for a hack that illustrates this problem. Still thinking 
> of how an automated test would look like for this one. 
> My preferred target for  this fix is 0.20.1.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to