[ 
https://issues.apache.org/jira/browse/HDFS-989?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13145563#comment-13145563
 ] 

Todd Lipcon commented on HDFS-989:
----------------------------------

With hsync/hflush at a checksum boundary, we re-write the last checksum into 
the checksum file at the same time as we append the new data. There is a race 
during which the new data is there but not the checksum info -- but during 
recovery operations I believe we deal with this situation by ignoring checksum 
errors on the last checksum-chunk if there are no replicas with a valid 
last-chunk.
                
> Flush/sync do not work on Hadoop LocalFileSystem
> ------------------------------------------------
>
>                 Key: HDFS-989
>                 URL: https://issues.apache.org/jira/browse/HDFS-989
>             Project: Hadoop HDFS
>          Issue Type: Bug
>    Affects Versions: 0.20.1
>            Reporter: Nathan Marz
>
> They seem to be no-ops. This is really easy to reproduce, just open a file 
> using FileSystem/getLocal(new Configuration()), write data to the output 
> stream, and then try to flush/sync. I also tried creating the output stream 
> with a buffer size of 1, but that had no effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to