[ 
https://issues.apache.org/jira/browse/HADOOP-15074?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16518779#comment-16518779
 ] 

Harish Jaiprakash commented on HADOOP-15074:
--------------------------------------------

[~arpitagarwal], SequenceFileWriter depends upon FSDataOutputStream object 
returned by FileSystem.create/append, which does not expose the 
hsync(EnumSet<SyncFlag> syncFlags). The DFSOutputStream gets wrapped into a 
FSDataOutputStream so it's not possible to fix in SequenceFile.

This bug makes it hard to implement producer / consumer using sequence files. 
We are a bit stuck on this. When would length get persisted, if hsync is never 
called with UPDATE_LENGTH? Is there like a periodic update of length or udpate 
when a block is full and written or only when close is called?

> SequenceFile#Writer flush does not update the length of the written file.
> -------------------------------------------------------------------------
>
>                 Key: HADOOP-15074
>                 URL: https://issues.apache.org/jira/browse/HADOOP-15074
>             Project: Hadoop Common
>          Issue Type: Bug
>            Reporter: Mukul Kumar Singh
>            Assignee: Shashikant Banerjee
>            Priority: Major
>
> SequenceFile#Writer flush does not update the length of the file. This 
> happens because as part of the flush, {{UPDATE_LENGTH}} flag is not passed to 
> the DFSOutputStream#hsync.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to