[ 
https://issues.apache.org/jira/browse/NIFI-3724?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16236590#comment-16236590
 ] 

ASF GitHub Bot commented on NIFI-3724:
--------------------------------------

Github user bbende commented on the issue:

    https://github.com/apache/nifi/pull/1712
  
    @nellashapiro123 it would probably be best to ask this on the mailing lists:
    https://nifi.apache.org/mailing_lists.html
    
    If you send an email, please provide more info about your flow like which 
reader and writer is FetchParquet using? what schema access strategy is each 
reader and writer using? and if using schema access by name, what is the value 
of the schema.name attribute coming into FetchParquet?


> Add Put/Fetch Parquet Processors
> --------------------------------
>
>                 Key: NIFI-3724
>                 URL: https://issues.apache.org/jira/browse/NIFI-3724
>             Project: Apache NiFi
>          Issue Type: Improvement
>            Reporter: Bryan Bende
>            Assignee: Bryan Bende
>            Priority: Minor
>             Fix For: 1.2.0
>
>
> Now that we have the record reader/writer services currently in master, it 
> would be nice to have reader and writers for Parquet. Since Parquet's API is 
> based on the Hadoop Path object, and not InputStreams/OutputStreams, we can't 
> really implement direct conversions to and from Parquet in the middle of a 
> flow, but we can we can perform the conversion by taking any record format 
> and writing to a Path as Parquet, or reading Parquet from a Path and writing 
> it out as another record format.
> We should add a PutParquet that uses a record reader and writes records to a 
> Path as Parquet, and a FetchParquet that reads Parquet from a path and writes 
> out records to a flow file using a record writer.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to