[
https://issues.apache.org/jira/browse/HUDI-69?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17084560#comment-17084560
]
Vinoth Chandar commented on HUDI-69:
------------------------------------
. I was wondering if we can just wrap the FileFormat (Parquet/ORC both have
formats inside Spark) , reuse its record reader for reading parquet/orc -> Row
and also use our existing LogReader classes to read the log blocks are Row
(instead of GenericRecord.. or we can for now do GenericRecord -> Row ).. This
means, we need to redesign our CompactedRecordScanner etc classes to be generic
and not implicitly assume it merging Avro/ArrayWritable per se. Must be doable.
> Support realtime view in Spark datasource #136
> ----------------------------------------------
>
> Key: HUDI-69
> URL: https://issues.apache.org/jira/browse/HUDI-69
> Project: Apache Hudi (incubating)
> Issue Type: New Feature
> Components: Spark Integration
> Reporter: Vinoth Chandar
> Assignee: Yanjia Gary Li
> Priority: Major
> Fix For: 0.6.0
>
>
> https://github.com/uber/hudi/issues/136
--
This message was sent by Atlassian Jira
(v8.3.4#803005)