[
https://issues.apache.org/jira/browse/FLINK-6573?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17443565#comment-17443565
]
ZhuoYu Chen commented on FLINK-6573:
------------------------------------
[~arvid] {color:#333333}Thank you for reminding me{color}.{color:#333333}I have
completed most of the work, but there are some shortcomings :1.
{color}{color:#333333}When the data sink to mongodb is an array,mongodb will
return an exception, and you need to transfer the array to List 2 before sink
data.{color}{color:#333333}Specify the primary key 3. {color}{color:#333333}The
requirement is to organize the code to meet FLink PR specification{color}
!image-2021-11-15-14-41-07-514.png!
> Flink MongoDB Connector
> -----------------------
>
> Key: FLINK-6573
> URL: https://issues.apache.org/jira/browse/FLINK-6573
> Project: Flink
> Issue Type: New Feature
> Components: Connectors / Common
> Affects Versions: 1.2.0
> Environment: Linux Operating System, Mongo DB
> Reporter: Nagamallikarjuna
> Assignee: ZhuoYu Chen
> Priority: Not a Priority
> Labels: stale-assigned
> Attachments: image-2021-11-15-14-35-04-364.png,
> image-2021-11-15-14-35-40-394.png
>
> Original Estimate: 672h
> Remaining Estimate: 672h
>
> Hi Community,
> Currently we are using Flink in the current Project. We have huge amount of
> data to process using Flink which resides in Mongo DB. We have a requirement
> of parallel data connectivity in between Flink and Mongo DB for both
> reads/writes. Currently we are planning to create this connector and
> contribute to the Community.
> I will update the further details once I receive your feedback
> Please let us know if you have any concerns.
--
This message was sent by Atlassian Jira
(v8.20.1#820001)