[
https://issues.apache.org/jira/browse/FLINK-7243?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16647573#comment-16647573
]
ASF GitHub Bot commented on FLINK-7243:
---------------------------------------
HuangZhenQiu commented on issue #6483: [FLINK-7243][flink-formats] Add parquet
input format
URL: https://github.com/apache/flink/pull/6483#issuecomment-429234463
@fhueske
Thanks for the review. Resolved all of the comments except unit tests for
the checkpointing logic
1) For the question of "instead of always reading as Row and from there
converting to the other types?"
In Parquet's interface, a converter is needed for each type of result.
Record can be convert to row by recursively put children in particular index,
but Map has to do it with Key. To reduce code duplication, I use the row as
intermediate representation. So type conversion can be put in sub class of
ParquetInputFormat.
2) I will add unit test for checkpoint logic tomorrow night.
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]
> Add ParquetInputFormat
> ----------------------
>
> Key: FLINK-7243
> URL: https://issues.apache.org/jira/browse/FLINK-7243
> Project: Flink
> Issue Type: Sub-task
> Components: Table API & SQL
> Reporter: godfrey he
> Assignee: Zhenqiu Huang
> Priority: Major
> Labels: pull-request-available
>
> Add a {{ParquetInputFormat}} to read data from a Apache Parquet file.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)