Hi Jark, Sorry for my late reply.
Yes, I’m using the old planner. I’ve tried the blink planner, and it works well. We would like to switch to the blink planner, but we’ve developed some custom features on the old planner, so it would take some time to port the codes. So I might give a try to fix the old planner if it’s not too involved. Best, Paul Lam > 在 2020年3月19日,17:13,Jark Wu <imj...@gmail.com> 写道: > > Hi Paul, > > Are you using old planner? Did you try blink planner? I guess it maybe a bug > in old planner which doesn't work well on new types. > > Best, > Jark > > On Thu, 19 Mar 2020 at 16:27, Paul Lam <paullin3...@gmail.com > <mailto:paullin3...@gmail.com>> wrote: > Hi, > > Recently I upgraded a simple application that inserts static data into a > table from 1.9.0 to 1.10.0, and > encountered a timestamp type incompatibility problem during the table sink > validation. > > The SQL is like: > ``` > insert into kafka.test.tbl_a # schema: (user_name STRING, user_id INT, > login_time TIMESTAMP) > select ("ann", 1000, TIMESTAMP "2019-12-30 00:00:00") > ``` > > And the error thrown: > ``` > Field types of query result and registered TableSink `kafka`.`test`.`tbl_a` > do not match. > Query result schema: [EXPR$0: String, EXPR$1: Integer, EXPR$2: > Timestamp] > TableSink schema: [user_name: String, user_id: Integer, login_time: > LocalDateTime] > ``` > > After some digging, I found the root cause might be that since FLINK-14645 > timestamp fields > defined via TableFactory had been bridged to LocalDateTime, but timestamp > literals are > still backed by java.sql.Timestamp. > > Is my reasoning correct? And is there any workaround? Thanks a lot! > > Best, > Paul Lam >