Hi Vino & Hequn,
I am now using the table/sql API, if I import the mysql table as a
stream then convert it into a table, it seems that it can also be a workaround
for batch/streaming joining. May I ask what is the difference between the UDTF
method? Does this implementation has some defects?
Best
Henry
> 在 2018年9月22日,上午10:28,Hequn Cheng <[email protected]> 写道:
>
> Hi
>
> +1 for vino's answer.
> Also, this kind of join will be supported in FLINK-9712
> <https://issues.apache.org/jira/browse/FLINK-9712>. You can check more
> details in the jira.
>
> Best, Hequn
>
> On Fri, Sep 21, 2018 at 4:51 PM vino yang <[email protected]
> <mailto:[email protected]>> wrote:
> Hi Henry,
>
> There are three ways I can think of:
>
> 1) use DataStream API, implement a flatmap UDF to access dimension table;
> 2) use table/sql API, implement a UDTF to access dimension table;
> 3) customize the table/sql join API/statement's implementation (and change
> the physical plan)
>
> Thanks, vino.
>
> 徐涛 <[email protected] <mailto:[email protected]>> 于2018年9月21日周五
> 下午4:43写道:
> Hi All,
> Sometimes some “dimension table” need to be joined from the "fact
> table", if data are not joined before sent to Kafka.
> So if the data are joined in Flink, does the “dimension table” have
> to be import as a stream, or there are some other ways can achieve it?
> Thanks a lot!
>
> Best
> Henry