[
https://issues.apache.org/jira/browse/FLINK-21643?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17296685#comment-17296685
]
Maciej Obuchowski edited comment on FLINK-21643 at 3/7/21, 1:45 AM:
--------------------------------------------------------------------
As I already have solution (working on production with Oracle) on my Flink
fork, I'll provide it as draft solution.
EDIT: here: https://github.com/apache/flink/pull/15102
was (Author: mobuchowski):
As I already have solution (working on production with Oracle) on my Flink
fork, I'll provide it as draft solution.
> JDBC sink should be able to execute statements on multiple tables
> -----------------------------------------------------------------
>
> Key: FLINK-21643
> URL: https://issues.apache.org/jira/browse/FLINK-21643
> Project: Flink
> Issue Type: New Feature
> Components: Connectors / JDBC
> Affects Versions: 1.12.2
> Reporter: Maciej Obuchowski
> Priority: Major
> Labels: pull-request-available
>
> Currently datastream JDBC sink supports outputting data only to one table -
> by having to provide SQL template, from which SimpleBatchStatementExecutor
> creates PreparedStatement. Creating multiple sinks, each of which writes data
> to one table is impractical for moderate to large number of tables -
> relational databases don't usually tolerate large number of connections.
> I propose adding DynamicBatchStatementExecutor, which will additionally
> require
> 1) provided mechanism to create SQL statements based on given object
> 2) cache for prepared statements
> 3) mechanism for determining which statement should be used for given object
--
This message was sent by Atlassian Jira
(v8.3.4#803005)