Hi Akshay, thanks for reaching out to the community. There was a similar question on the mailing list earlier this month [1]. Unfortunately, it just doesn't seem to be supported, yet. The feature request was already created with FLINK-23589 [2].
Best, Matthias [1] https://lists.apache.org/thread.html/r463f748358202d207e4bf9c7fdcb77e609f35bbd670dbc5278dd7615%40%3Cuser.flink.apache.org%3E [2] https://issues.apache.org/jira/browse/FLINK-23589 On Thu, Aug 26, 2021 at 11:07 AM Akshay Agarwal <akshay.agar...@grofers.com> wrote: > Hi everyone, > > We are trying out flink 1.13.1 with kafka topics as avro backend but we > are facing an issue while creating Table SQL that avro doesn't support > precision greater than 3. I am not getting the reason why flink isn't > supporting timestamps greater than 3 (exception > <https://github.com/apache/flink/blob/3555741a12ba9fb65e8db9f731a131ab39d1cfe8/flink-formats/flink-avro/src/main/java/org/apache/flink/formats/avro/typeutils/AvroSchemaConverter.java#L359>) > as avro(1.10.0) does support microseconds precision. > Our kafka records contain a timestamp as YYYY-MM-DD'T'HH-mm-ss.SSSSSS'Z'. > Eventually I have read the records as a custom UDF which drops seconds > offset but wanted to know if there is a better way to handle this also the > reason why isn't it's supported. It would be a great help for us to know > about it and we can build on that. > > *Regards* > *Akshay Agarwal* > > [image: https://grofers.com] <https://grofers.com>