[
https://issues.apache.org/jira/browse/HADOOP-5553?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12688896#action_12688896
]
Chris Douglas commented on HADOOP-5553:
---------------------------------------
The binary reader for a file format doesn't seem like the right place to inject
record projection or record skipping, particularly through user code. if the
reader were modified to read lazily, then whatever is consuming the value bytes
can discard whatever it pleases from the stream.
Are you sure that SequenceFile fulfills your requirements? If you need such
specific control, perhaps a less general format would be more fitting.
> Change modifier of SequenceFile.CompressedBytes and
> SequenceFile.UncompressedBytes from private to public
> ---------------------------------------------------------------------------------------------------------
>
> Key: HADOOP-5553
> URL: https://issues.apache.org/jira/browse/HADOOP-5553
> Project: Hadoop Core
> Issue Type: Improvement
> Reporter: He Yongqiang
> Attachments: Hadoop-5553-2.patch, Hadoop-5553-3.patch,
> Hadoop-5553.patch
>
>
> SequenceFile.rawValue() provides the only interface to navigate the
> underlying bytes. And with some little work on implementing a customized
> ValueBytes can avoid reading all bytes into memory. Unfortunately, the
> current nextRawValue will cast the passing ValueBytes to either private class
> CompressedBytes or private class UnCompressedBytes, this will disallow user
> further extension.
> I can not see any reason that CompressedBytes and UnCompressedBytes should be
> set to private. And since the ValueBytes is public and nextValue() casts it
> to either CompressedBytes or UnCompressedBytes, i think it would be better if
> they are public.
> I am stuck now by this issue, really appracited if this got resolved as soon
> as possible.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.