[
https://issues.apache.org/jira/browse/ARROW-8127?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Wes McKinney resolved ARROW-8127.
---------------------------------
Fix Version/s: 0.17.0
Resolution: Fixed
Issue resolved by pull request 6637
[https://github.com/apache/arrow/pull/6637]
> [C++] [Parquet] Incorrect column chunk metadata for multipage batch writes
> --------------------------------------------------------------------------
>
> Key: ARROW-8127
> URL: https://issues.apache.org/jira/browse/ARROW-8127
> Project: Apache Arrow
> Issue Type: Bug
> Components: C++
> Reporter: TP Boudreau
> Assignee: TP Boudreau
> Priority: Minor
> Labels: pull-request-available
> Fix For: 0.17.0
>
> Attachments: multipage-batch-write.cc
>
> Time Spent: 1.5h
> Remaining Estimate: 0h
>
> When writing to a buffered column writer using PLAIN encoding, if the size of
> the batch supplied for writing exceeds the page size for the writer, the
> resulting file has an incorrect data_page_offset set in its column chunk
> metadata. This causes an exception to be thrown when reading the file (file
> appears to be too short to the reader).
> For example, the attached code, which attempts to write a batch of 262145
> Int32's (= 1048576 + 4 bytes) using the default page size of 1048576 bytes
> (with buffered writer, PLAIN encoding), fails on reading, throwing the error:
> "Tried reading 1048678 bytes starting at position 1048633 from file but only
> got 333".
> The error is caused by the second page write tripping the conditional here
> https://github.com/apache/arrow/blob/master/cpp/src/parquet/column_writer.cc#L302,
> in the serialized in-memory writer wrapped by the buffered writer.
> The fix builds the metadata with offsets from the terminal sink rather than
> the in memory buffered sink. A PR is coming.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)