[
https://issues.apache.org/jira/browse/FLINK-15206?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Bowen Li updated FLINK-15206:
-----------------------------
Summary: support dynamic catalog table for truly unified SQL job (was:
support dynamic catalog table for unified SQL job)
> support dynamic catalog table for truly unified SQL job
> -------------------------------------------------------
>
> Key: FLINK-15206
> URL: https://issues.apache.org/jira/browse/FLINK-15206
> Project: Flink
> Issue Type: New Feature
> Components: Table SQL / API
> Reporter: Bowen Li
> Assignee: Bowen Li
> Priority: Major
> Fix For: 1.11.0
>
>
> currently if users have both an online and an offline job with same business
> logic in Flink SQL, their codebase is still not unified. They would keep two
> SQL statements whose only difference is the source (or/and sink) table. E.g.
> {code:java}
> // online job
> insert into x select * from kafka_table;
> // offline backfill job
> insert into x select * from hive_table;
> {code}
> We would like to introduce a "dynamic catalog table". The dynamic catalog
> table acts as a view, and is just an abstract source from actual sources
> behind it under with configurations. When execute a job, depending on the
> configuration, the dynamic catalog table can point to an actual source table.
> A use case for this is the example given above - users want to just keep one
> sql statement as {{insert into x select * from my_source_dynamic_table);}};
> when executed in streaming mode, the {{my_source_dynamic_table}} should point
> to a kafka catalog table, and in batch mode, the {{my_source_dynamic_table}}
> should point to a hive catalog table.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)