[ 
https://issues.apache.org/jira/browse/HUDI-184?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16977090#comment-16977090
 ] 

vinoyang commented on HUDI-184:
-------------------------------

bq. Also there are probably some bigger questions to answer first? e.g if we 
are targetting the streaming APIs, then whats the execution model? In Spark 
Streaming, we commit after each micro batch.  When do we commit for Flink 
writing? 

>From a high-level abstraction, the execution model of Flink is DataFlow. From 
>a low-level abstraction, in unbounded streaming, window is a mechanism that 
>split the unbounded stream into bounded stream. Maybe we can use the window in 
>Flink to mapping the micro-batch(RDD) in Spark.

> Integrate Hudi with Apache Flink
> --------------------------------
>
>                 Key: HUDI-184
>                 URL: https://issues.apache.org/jira/browse/HUDI-184
>             Project: Apache Hudi (incubating)
>          Issue Type: New Feature
>          Components: Write Client
>            Reporter: vinoyang
>            Assignee: vinoyang
>            Priority: Major
>
> Apache Flink is a popular streaming processing engine.
> Integrating Hudi with Flink is a valuable work.
> The discussion mailing thread is here: 
> [https://lists.apache.org/api/source.lua/1533de2d4cd4243fa9e8f8bf057ffd02f2ac0bec7c7539d8f72166ea@%3Cdev.hudi.apache.org%3E]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to