Not too familiar wth Nifi myself. Would this still target an use-case like what 
pratyaksh mentioned? 
For delta streamer specifically, we are moving more and more towards continuous 
mode, where 
Hudi writing and compaction are amanged by a single long running spark 
application. 

Would Nifi also help us manage compactions when working with Hudi datasource or 
just writing plain spark Hudi pipelines? 

On 2019/09/18 08:18:44, Taher Koitawala <taher...@gmail.com> wrote: 
> That's another way of doing things. I want to know if someone wrote
> something like PutParquet. Which directly can write data to Hudi. AFAIK I
> don't think anyone has.
> 
> That will really be powerful.
> 
> On Wed, Sep 18, 2019, 1:37 PM Pratyaksh Sharma <pratyaks...@gmail.com>
> wrote:
> 
> > Hi Taher,
> >
> > In the initial phase of our CDC pipeline, we were using Hudi with Nifi.
> > Nifi was being used to read Binlog file of mysql and to push that data to
> > some Kafka topic. This topic was then getting consumed by DeltaStreamer. So
> > Nifi was indirectly involved in that flow.
> >
> > On Wed, Sep 18, 2019 at 10:29 AM Taher Koitawala <taher...@gmail.com>
> > wrote:
> >
> > > Hi All,
> > >           Just wanted to know has anyone tried to write data to Hudi
> > with a
> > > Nifi flow?
> > >
> > > Perhaps may be just a csv file on local to Hudi dataset? If not then lets
> > > try that!
> > >
> > > Regards,
> > > Taher Koitawala
> > >
> >
> 

Reply via email to