[ 
https://issues.apache.org/jira/browse/CAMEL-13573?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17326566#comment-17326566
 ] 

Miguel Costa commented on CAMEL-13573:
--------------------------------------

Maybe this is something basic but can you point me to the project where we 
could create a PR? I would like to use this in a Kafka Camel Sink Connector but 
I understood that this would need to be done at the Camel level and not in the 
Camel Connectors so I would not be sure on how this could be extended.

Would it be something similar to this setup in Avro?
https://github.com/apache/camel/blob/main/components/camel-avro/src/main/java/org/apache/camel/dataformat/avro/AvroDataFormat.java

Because we did this conversion to Parquet in other places (like in Apache 
Flink) using org.apache.parquet.hadoop.ParquetWriter but I've also seen it in 
other projects where actually first people transform the data into Avro and 
then pass it to Parquet using the and I've seen it done in a way that it just 
first passes to Avro and then to Parquet using 
org.apache.parquet.avro.AvroParquetWriter


> Parquet Dataformat: supporting parquet files in marshal / unmarshal
> -------------------------------------------------------------------
>
>                 Key: CAMEL-13573
>                 URL: https://issues.apache.org/jira/browse/CAMEL-13573
>             Project: Camel
>          Issue Type: New Feature
>            Reporter: Matteo Redaelli
>            Priority: Major
>             Fix For: Future
>
>
> It would be useweful to read / store data (marshal / unmarshal) in parquet 
> format ([https://parquet.apache.org/]). 
> It is widely used in bigdata projects / datalakes 
>  
> I did not found any info about it in the documentation...
>  
> Thanks
> Matteo



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to