大家好,请教个问题:
我在pyflink中使用SQL DDL创建kafka源,如下:
kafka_source_ddl = """
CREATE TABLE kafka_source_tb (
name VARCHAR,
number INT,
msgtime TIMESTAMP,
WATERMARK FOR msgtime AS msgtime
) WITH (
'connector.type' = 'kafka',
'connector.version' = 'universal',
'connector.topic' = 'mytopic',
'connector.properties.zookeeper.connect' = 'localhost:2181',
'connector.properties.bootstrap.servers' = 'localhost:9092',
'format.type' = 'json',
'format.derive-schema' = 'true'
)
"""
st_env.sql_update(kafka_source_ddl)
在使用窗口时报错,代码如下:
st_env.from_path("kafka_source_tb") \
.window(Slide.over("10.secends").every("1.secends").on("msgtime").alias("msgtime"))
\
.group_by("msgtime") \
.select("msgtime.start as b, msgtime.end as c, msgtime.rowtime as d")
\
报错如下
: org.apache.flink.table.api.ValidationException: A group window expects a time
attribute for grouping in a stream environment.
at
org.apache.flink.table.operations.utils.factories.AggregateOperationFactory.validateStreamTimeAttribute(AggregateOperationFactory.java:293)
at
org.apache.flink.table.operations.utils.factories.AggregateOperationFactory.validateTimeAttributeType(AggregateOperationFactory.java:278)
at
org.apache.flink.table.operations.utils.factories.AggregateOperationFactory.getValidatedTimeAttribute(AggregateOperationFactory.java:271)
at
org.apache.flink.table.operations.utils.factories.AggregateOperationFactory.createResolvedWindow(AggregateOperationFactory.java:233)
at
org.apache.flink.table.operations.utils.OperationTreeBuilder.windowAggregate(OperationTreeBuilder.java:243)
at
org.apache.flink.table.api.internal.TableImpl$WindowGroupedTableImpl.select(TableImpl.java:762)
at
org.apache.flink.table.api.internal.TableImpl$WindowGroupedTableImpl.select(TableImpl.java:747)
我看到有人说直接使用api的rowtime函数好像有bug,就选了用DDL的,这个error是否相关?或者是我哪里写错了?
请大家帮忙看一下
谢谢!