[
https://issues.apache.org/jira/browse/FLINK-36443?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Xuyang Zhong updated FLINK-36443:
---------------------------------
Fix Version/s: 2.1.0
> Flaky Test: assertEquals in
> PushPartitionIntoTableSourceScanRuleTest.testWithUdf
> --------------------------------------------------------------------------------
>
> Key: FLINK-36443
> URL: https://issues.apache.org/jira/browse/FLINK-36443
> Project: Flink
> Issue Type: Bug
> Components: Table SQL / Planner
> Reporter: William Lin
> Assignee: Muhammet Orazov
> Priority: Minor
> Labels: pull-request-available
> Fix For: 2.1.0
>
>
> The assertEquals function in testWithUdf in
> PushPartitionIntoTableSourceScanRuleTest assumes the partitions in the
> generated plan have a specific order. However, the order is not specific in
> some cases, causing this a flaky test.
> {code:java}
> // code placeholder
> [ERROR]
> org.apache.flink.table.planner.plan.rules.logical.PushPartitionIntoTableSourceScanRuleTest.testWithUdf
> -- Time elapsed: 0.159 s <<< FAILURE! org.opentest4j.AssertionFailedError:
> optimized rel plan ==> expected: < LogicalProject(id=[$0], name=[$1],
> part1=[$2], part2=[$3]) +- LogicalFilter(condition=[>($0, 2)]) +-
> LogicalTableScan(table=[[test_catalog, test_database, MyTable,
> partitions=[{part1=A, part2=1}, {part1=C, part2=1}]]]) > but was: <
> LogicalProject(id=[$0], name=[$1], part1=[$2], part2=[$3]) +-
> LogicalFilter(condition=[>($0, 2)]) +- LogicalTableScan(table=[[test_catalog,
> test_database, MyTable, partitions=[{part1=A, part2=1}, {part2=1,
> part1=C}]]]) > at
> org.apache.flink.table.planner.utils.DiffRepository.assertEquals(DiffRepository.java:438)
> at
> org.apache.flink.table.planner.utils.TableTestUtilBase.assertEqualsOrExpand(TableTestBase.scala:1176)
> at
> org.apache.flink.table.planner.utils.TableTestUtilBase.assertPlanEquals(TableTestBase.scala:1091)
> at
> org.apache.flink.table.planner.utils.TableTestUtilBase.doVerifyPlan(TableTestBase.scala:921)
> at
> org.apache.flink.table.planner.utils.TableTestUtilBase.verifyRelPlan(TableTestBase.scala:467)
> at
> org.apache.flink.table.planner.plan.rules.logical.PushPartitionIntoLegacyTableSourceScanRuleTest.testWithUdf(PushPartitionIntoLegacyTableSourceScanRuleTest.scala:179)
> at java.base/java.lang.reflect.Method.invoke(Method.java:569) at
> java.base/java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:183)
> at
> java.base/java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:197)
> at
> java.base/java.util.stream.ReferencePipeline$2$1.accept(ReferencePipeline.java:179)
> at
> java.base/java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:197)
> at
> java.base/java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:183)
> at
> java.base/java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:197)
> at
> java.base/java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:197)
> at java.base/java.util.Iterator.forEachRemaining(Iterator.java:133) at
> scala.collection.convert.Wrappers$IteratorWrapper.forEachRemaining(Wrappers.scala:31)
> at
> java.base/java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1845)
> at
> java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:509)
> at
> java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:499)
> at
> java.base/java.util.stream.ForEachOps$ForEachOp.evaluateSequential(ForEachOps.java:150)
> at
> java.base/java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(ForEachOps.java:173)
> at
> java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
> at
> java.base/java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:596)
> at
> java.base/java.util.stream.ReferencePipeline$7$1.accept(ReferencePipeline.java:276)
> at
> java.base/java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1625)
> at
> java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:509)
> at
> java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:499)
> at
> java.base/java.util.stream.ForEachOps$ForEachOp.evaluateSequential(ForEachOps.java:150)
> at
> java.base/java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(ForEachOps.java:173)
> at
> java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
> at
> java.base/java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:596)
> at
> java.base/java.util.concurrent.RecursiveAction.exec(RecursiveAction.java:194)
> at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:373)
> at
> java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1182)
> at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1655)
> at
> java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1622)
> at
> java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165)
> Here is the line related to this: String expected2Canonical =
> expected2.replace(Util.LINE_SEPARATOR, "\n"); String actualCanonical =
> actual.replace(Util.LINE_SEPARATOR, "\n");
> Assertions.assertEquals(expected2Canonical, actualCanonical, tag); {code}
--
This message was sent by Atlassian Jira
(v8.20.10#820010)