Wei Zhong created FLINK-28764:
---------------------------------
Summary: Support more than 64 distinct aggregate function calls in
one aggregate SQL query
Key: FLINK-28764
URL: https://issues.apache.org/jira/browse/FLINK-28764
Project: Flink
Issue Type: Improvement
Components: Table SQL / Planner
Affects Versions: 1.15.1, 1.14.5, 1.13.6
Reporter: Wei Zhong
Currently Flink SQL does not support more than 64 distinct aggregate function
calls in one aggregate SQL query. We encountered this problem while migrating
batch jobs from spark to flink. The spark job has 79 distinct aggregate
function calls in one aggregate SQL query.
Reproduce code:
{code:java}
public class Test64Distinct {
public static void main(String[] args) {
TableEnvironment tableEnv =
TableEnvironment.create(EnvironmentSettings.inBatchMode());
tableEnv.executeSql("create table datagen_source(id BIGINT, val BIGINT)
with " +
"('connector'='datagen', 'number-of-rows'='1000')");
tableEnv.executeSql("select " +
"count(distinct val * 1), " +
"count(distinct val * 2), " +
"count(distinct val * 3), " +
"count(distinct val * 4), " +
"count(distinct val * 5), " +
"count(distinct val * 6), " +
"count(distinct val * 7), " +
"count(distinct val * 8), " +
"count(distinct val * 9), " +
"count(distinct val * 10), " +
"count(distinct val * 11), " +
"count(distinct val * 12), " +
"count(distinct val * 13), " +
"count(distinct val * 14), " +
"count(distinct val * 15), " +
"count(distinct val * 16), " +
"count(distinct val * 17), " +
"count(distinct val * 18), " +
"count(distinct val * 19), " +
"count(distinct val * 20), " +
"count(distinct val * 21), " +
"count(distinct val * 22), " +
"count(distinct val * 23), " +
"count(distinct val * 24), " +
"count(distinct val * 25), " +
"count(distinct val * 26), " +
"count(distinct val * 27), " +
"count(distinct val * 28), " +
"count(distinct val * 29), " +
"count(distinct val * 30), " +
"count(distinct val * 31), " +
"count(distinct val * 32), " +
"count(distinct val * 33), " +
"count(distinct val * 34), " +
"count(distinct val * 35), " +
"count(distinct val * 36), " +
"count(distinct val * 37), " +
"count(distinct val * 38), " +
"count(distinct val * 39), " +
"count(distinct val * 40), " +
"count(distinct val * 41), " +
"count(distinct val * 42), " +
"count(distinct val * 43), " +
"count(distinct val * 44), " +
"count(distinct val * 45), " +
"count(distinct val * 46), " +
"count(distinct val * 47), " +
"count(distinct val * 48), " +
"count(distinct val * 49), " +
"count(distinct val * 50), " +
"count(distinct val * 51), " +
"count(distinct val * 52), " +
"count(distinct val * 53), " +
"count(distinct val * 54), " +
"count(distinct val * 55), " +
"count(distinct val * 56), " +
"count(distinct val * 57), " +
"count(distinct val * 58), " +
"count(distinct val * 59), " +
"count(distinct val * 60), " +
"count(distinct val * 61), " +
"count(distinct val * 62), " +
"count(distinct val * 63), " +
"count(distinct val * 64), " +
"count(distinct val * 65) from datagen_source").print();
}
} {code}
Exception:
{code:java}
Exception in thread "main" org.apache.flink.table.api.TableException: Sql
optimization: Cannot generate a valid execution plan for the given query:
LogicalSink(table=[*anonymous_collect$1*], fields=[EXPR$0, EXPR$1, EXPR$2,
EXPR$3, EXPR$4, EXPR$5, EXPR$6, EXPR$7, EXPR$8, EXPR$9, EXPR$10, EXPR$11,
EXPR$12, EXPR$13, EXPR$14, EXPR$15, EXPR$16, EXPR$17, EXPR$18, EXPR$19,
EXPR$20, EXPR$21, EXPR$22, EXPR$23, EXPR$24, EXPR$25, EXPR$26, EXPR$27,
EXPR$28, EXPR$29, EXPR$30, EXPR$31, EXPR$32, EXPR$33, EXPR$34, EXPR$35,
EXPR$36, EXPR$37, EXPR$38, EXPR$39, EXPR$40, EXPR$41, EXPR$42, EXPR$43,
EXPR$44, EXPR$45, EXPR$46, EXPR$47, EXPR$48, EXPR$49, EXPR$50, EXPR$51,
EXPR$52, EXPR$53, EXPR$54, EXPR$55, EXPR$56, EXPR$57, EXPR$58, EXPR$59,
EXPR$60, EXPR$61, EXPR$62, EXPR$63, EXPR$64])
+- LogicalAggregate(group=[{}], EXPR$0=[COUNT(DISTINCT $0)],
EXPR$1=[COUNT(DISTINCT $1)], EXPR$2=[COUNT(DISTINCT $2)],
EXPR$3=[COUNT(DISTINCT $3)], EXPR$4=[COUNT(DISTINCT $4)],
EXPR$5=[COUNT(DISTINCT $5)], EXPR$6=[COUNT(DISTINCT $6)],
EXPR$7=[COUNT(DISTINCT $7)], EXPR$8=[COUNT(DISTINCT $8)],
EXPR$9=[COUNT(DISTINCT $9)], EXPR$10=[COUNT(DISTINCT $10)],
EXPR$11=[COUNT(DISTINCT $11)], EXPR$12=[COUNT(DISTINCT $12)],
EXPR$13=[COUNT(DISTINCT $13)], EXPR$14=[COUNT(DISTINCT $14)],
EXPR$15=[COUNT(DISTINCT $15)], EXPR$16=[COUNT(DISTINCT $16)],
EXPR$17=[COUNT(DISTINCT $17)], EXPR$18=[COUNT(DISTINCT $18)],
EXPR$19=[COUNT(DISTINCT $19)], EXPR$20=[COUNT(DISTINCT $20)],
EXPR$21=[COUNT(DISTINCT $21)], EXPR$22=[COUNT(DISTINCT $22)],
EXPR$23=[COUNT(DISTINCT $23)], EXPR$24=[COUNT(DISTINCT $24)],
EXPR$25=[COUNT(DISTINCT $25)], EXPR$26=[COUNT(DISTINCT $26)],
EXPR$27=[COUNT(DISTINCT $27)], EXPR$28=[COUNT(DISTINCT $28)],
EXPR$29=[COUNT(DISTINCT $29)], EXPR$30=[COUNT(DISTINCT $30)],
EXPR$31=[COUNT(DISTINCT $31)], EXPR$32=[COUNT(DISTINCT $32)],
EXPR$33=[COUNT(DISTINCT $33)], EXPR$34=[COUNT(DISTINCT $34)],
EXPR$35=[COUNT(DISTINCT $35)], EXPR$36=[COUNT(DISTINCT $36)],
EXPR$37=[COUNT(DISTINCT $37)], EXPR$38=[COUNT(DISTINCT $38)],
EXPR$39=[COUNT(DISTINCT $39)], EXPR$40=[COUNT(DISTINCT $40)],
EXPR$41=[COUNT(DISTINCT $41)], EXPR$42=[COUNT(DISTINCT $42)],
EXPR$43=[COUNT(DISTINCT $43)], EXPR$44=[COUNT(DISTINCT $44)],
EXPR$45=[COUNT(DISTINCT $45)], EXPR$46=[COUNT(DISTINCT $46)],
EXPR$47=[COUNT(DISTINCT $47)], EXPR$48=[COUNT(DISTINCT $48)],
EXPR$49=[COUNT(DISTINCT $49)], EXPR$50=[COUNT(DISTINCT $50)],
EXPR$51=[COUNT(DISTINCT $51)], EXPR$52=[COUNT(DISTINCT $52)],
EXPR$53=[COUNT(DISTINCT $53)], EXPR$54=[COUNT(DISTINCT $54)],
EXPR$55=[COUNT(DISTINCT $55)], EXPR$56=[COUNT(DISTINCT $56)],
EXPR$57=[COUNT(DISTINCT $57)], EXPR$58=[COUNT(DISTINCT $58)],
EXPR$59=[COUNT(DISTINCT $59)], EXPR$60=[COUNT(DISTINCT $60)],
EXPR$61=[COUNT(DISTINCT $61)], EXPR$62=[COUNT(DISTINCT $62)],
EXPR$63=[COUNT(DISTINCT $63)], EXPR$64=[COUNT(DISTINCT $64)])
+- LogicalProject(exprs=[[*($1, 1), *($1, 2), *($1, 3), *($1, 4), *($1, 5),
*($1, 6), *($1, 7), *($1, 8), *($1, 9), *($1, 10), *($1, 11), *($1, 12), *($1,
13), *($1, 14), *($1, 15), *($1, 16), *($1, 17), *($1, 18), *($1, 19), *($1,
20), *($1, 21), *($1, 22), *($1, 23), *($1, 24), *($1, 25), *($1, 26), *($1,
27), *($1, 28), *($1, 29), *($1, 30), *($1, 31), *($1, 32), *($1, 33), *($1,
34), *($1, 35), *($1, 36), *($1, 37), *($1, 38), *($1, 39), *($1, 40), *($1,
41), *($1, 42), *($1, 43), *($1, 44), *($1, 45), *($1, 46), *($1, 47), *($1,
48), *($1, 49), *($1, 50), *($1, 51), *($1, 52), *($1, 53), *($1, 54), *($1,
55), *($1, 56), *($1, 57), *($1, 58), *($1, 59), *($1, 60), *($1, 61), *($1,
62), *($1, 63), *($1, 64), *($1, 65)]])
+- LogicalTableScan(table=[[default_catalog, default_database,
datagen_source]])group count must be less than 64.
Please check the documentation for the set of currently supported SQL features.
at
org.apache.flink.table.planner.plan.optimize.program.FlinkVolcanoProgram.optimize(FlinkVolcanoProgram.scala:86)
at
org.apache.flink.table.planner.plan.optimize.program.FlinkChainedProgram.$anonfun$optimize$1(FlinkChainedProgram.scala:59)
at
scala.collection.TraversableOnce.$anonfun$foldLeft$1(TraversableOnce.scala:156)
at
scala.collection.TraversableOnce.$anonfun$foldLeft$1$adapted(TraversableOnce.scala:156)
at scala.collection.Iterator.foreach(Iterator.scala:937)
at scala.collection.Iterator.foreach$(Iterator.scala:937)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1425)
at scala.collection.IterableLike.foreach(IterableLike.scala:70)
at scala.collection.IterableLike.foreach$(IterableLike.scala:69)
at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:156)
at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:154)
at scala.collection.AbstractTraversable.foldLeft(Traversable.scala:104)
at
org.apache.flink.table.planner.plan.optimize.program.FlinkChainedProgram.optimize(FlinkChainedProgram.scala:55)
at
org.apache.flink.table.planner.plan.optimize.BatchCommonSubGraphBasedOptimizer.optimizeTree(BatchCommonSubGraphBasedOptimizer.scala:92)
at
org.apache.flink.table.planner.plan.optimize.BatchCommonSubGraphBasedOptimizer.optimizeBlock(BatchCommonSubGraphBasedOptimizer.scala:57)
at
org.apache.flink.table.planner.plan.optimize.BatchCommonSubGraphBasedOptimizer.$anonfun$doOptimize$1(BatchCommonSubGraphBasedOptimizer.scala:44)
at
org.apache.flink.table.planner.plan.optimize.BatchCommonSubGraphBasedOptimizer.$anonfun$doOptimize$1$adapted(BatchCommonSubGraphBasedOptimizer.scala:44)
at scala.collection.immutable.List.foreach(List.scala:388)
at
org.apache.flink.table.planner.plan.optimize.BatchCommonSubGraphBasedOptimizer.doOptimize(BatchCommonSubGraphBasedOptimizer.scala:44)
at
org.apache.flink.table.planner.plan.optimize.CommonSubGraphBasedOptimizer.optimize(CommonSubGraphBasedOptimizer.scala:78)
at
org.apache.flink.table.planner.delegation.PlannerBase.optimize(PlannerBase.scala:312)
at
org.apache.flink.table.planner.delegation.PlannerBase.translate(PlannerBase.scala:192)
at
org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:1688)
at
org.apache.flink.table.api.internal.TableEnvironmentImpl.executeQueryOperation(TableEnvironmentImpl.java:840)
at
org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:1342)
at
org.apache.flink.table.api.internal.TableEnvironmentImpl.executeSql(TableEnvironmentImpl.java:708)
at com.shopee.di.Test64Distinct.main(Test64Distinct.java:11)
Caused by: org.apache.flink.table.api.TableException: group count must be less
than 64.
at
org.apache.flink.table.planner.plan.rules.logical.DecomposeGroupingSetsRule.onMatch(DecomposeGroupingSetsRule.scala:177)
at
org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:229)
at
org.apache.calcite.plan.volcano.IterativeRuleDriver.drive(IterativeRuleDriver.java:58)
at
org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:510)
at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:312)
at
org.apache.flink.table.planner.plan.optimize.program.FlinkVolcanoProgram.optimize(FlinkVolcanoProgram.scala:62)
... 27 more {code}
--
This message was sent by Atlassian Jira
(v8.20.10#820010)