[
https://issues.apache.org/jira/browse/HDFS-6867?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14103161#comment-14103161
]
Zhe Zhang commented on HDFS-6867:
---------------------------------
How about starting a new block while recovering the partially-failed block in
background? This will cause a small block in the middle of the file, which is
suboptimal. To mitigate we can create a new block with a given size -- (default
block size) - (size of the partial block being recovered). Then we can merge
the recovered and the new blocks.
> For DFSOutputStream, do pipeline recovery for a single block in the background
> ------------------------------------------------------------------------------
>
> Key: HDFS-6867
> URL: https://issues.apache.org/jira/browse/HDFS-6867
> Project: Hadoop HDFS
> Issue Type: Bug
> Components: hdfs-client
> Affects Versions: 2.0.0-alpha
> Reporter: Colin Patrick McCabe
>
> For DFSOutputStream, we should be able to do pipeline recovery in the
> background, while the user is continuing to write to the file. This is
> especially useful for long-lived clients that write to an HDFS file slowly.
--
This message was sent by Atlassian JIRA
(v6.2#6252)