[ 
https://issues.apache.org/jira/browse/FLINK-7243?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16647573#comment-16647573
 ] 

ASF GitHub Bot commented on FLINK-7243:
---------------------------------------

HuangZhenQiu commented on issue #6483: [FLINK-7243][flink-formats] Add parquet 
input format
URL: https://github.com/apache/flink/pull/6483#issuecomment-429234463
 
 
   @fhueske 
   Thanks for the review. Resolved all of the comments except unit tests for 
the checkpointing logic 
   1) For the question of "instead of always reading as Row and from there 
converting to the other types?"
   
   In Parquet's interface, a converter is needed for each type of result. 
Record can be convert to row by recursively put children in particular index, 
but Map has to do it with Key. To reduce code duplication, I use the row as 
intermediate representation. So type conversion can be put in sub class of 
ParquetInputFormat.
   
   2) I will add unit test for checkpoint logic tomorrow night. 
   
   
   

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


> Add ParquetInputFormat
> ----------------------
>
>                 Key: FLINK-7243
>                 URL: https://issues.apache.org/jira/browse/FLINK-7243
>             Project: Flink
>          Issue Type: Sub-task
>          Components: Table API & SQL
>            Reporter: godfrey he
>            Assignee: Zhenqiu Huang
>            Priority: Major
>              Labels: pull-request-available
>
> Add a {{ParquetInputFormat}} to read data from a Apache Parquet file. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to