Thanks Zhuoyu for starting this discussion.

Looks good to me overall.

> Scenario 1: When you explicitly know which fields to use. In this case, the 
> system supports the retrieval of data from a single table and nested datasets.

This scenario just like using Flink SQL to read data from Mongo CDC
and project columns (including nested columns), and write data to
Paimon?

Best,
Jingsong

On Wed, Aug 9, 2023 at 9:23 AM 陈卓宇 <[email protected]> wrote:
>
> Hi devs,
>
> We would like to start a discussion about PIP-7: SyncAction based on
> MongoDB[1].
>
> As a streaming lake, users can utilize Paimon integrated with Flink CDC to
> accomplish the entire process of ingesting data into the lake. During this
> process, the absence of a fixed data schema in MongoDB often leads to an
> increase in the complexity of data synchronization development. I propose
> the development of a SyncAction based on MongoDB, which can assist users in
> efficiently ingesting MongoDB data into a data lake or data warehouse.
> Looking forward to hearing from you, thanks.
>
>
> [1]
> https://cwiki.apache.org/confluence/display/PAIMON/PIP-7%3A+SyncAction+based+on+MongoDB
>
> Best,
> Zhuoyu Chen

Reply via email to