Kihwal Lee created HDFS-10178:
---------------------------------
Summary: Permanent write failures can happen if pipeline
recoveries occur for the first packet
Key: HDFS-10178
URL: https://issues.apache.org/jira/browse/HDFS-10178
Project: Hadoop HDFS
Issue Type: Bug
Reporter: Kihwal Lee
Priority: Critical
We have observed that write fails permanently if the first packet doesn't go
through properly and pipeline recovery happens. If the packet header is sent
out, but the data portion of the packet does not reach one or more datanodes in
time, the pipeline recovery will be done against the 0-byte partial block.
If additional datanodes are added, the block is transferred to the new nodes.
After the transfer, each node will have a meta file containing the header and
0-length data block file. The pipeline recovery seems to work correctly up to
this point, but write fails when actual data packet is resent.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)