andygrove commented on code in PR #2566:
URL: https://github.com/apache/arrow-datafusion/pull/2566#discussion_r876075776
##########
datafusion/core/src/optimizer/limit_push_down.rs:
##########
@@ -278,4 +304,116 @@ mod test {
Ok(())
}
+
+ #[test]
+ fn limit_pushdown_with_offset_projection_table_provider() -> Result<()> {
+ let table_scan = test_table_scan()?;
+
+ let plan = LogicalPlanBuilder::from(table_scan)
+ .project(vec![col("a")])?
+ .offset(10)?
+ .limit(1000)?
+ .build()?;
+
+ // Should push the limit down to table provider
+ // When it has a select
+ let expected = "Limit: 1000\
+ \n Offset: 10\
+ \n Projection: #test.a\
+ \n TableScan: test projection=None, limit=1010";
+
+ assert_optimized_plan_eq(&plan, expected);
+
+ Ok(())
+ }
+
+ #[test]
+ fn limit_pushdown_with_offset_after_limit() -> Result<()> {
+ let table_scan = test_table_scan()?;
+
+ let plan = LogicalPlanBuilder::from(table_scan)
+ .project(vec![col("a")])?
+ .limit(1000)?
+ .offset(10)?
+ .build()?;
+
+ // Not push the limit down to table provider
+ // When offset after limit
+ let expected = "Offset: 10\
+ \n Limit: 1000\
+ \n Projection: #test.a\
+ \n TableScan: test projection=None, limit=1000";
+
+ assert_optimized_plan_eq(&plan, expected);
+
+ Ok(())
+ }
+
+ #[test]
+ fn limit_push_down_with_offset_take_smaller_limit() -> Result<()> {
+ let table_scan = test_table_scan()?;
+
+ let plan = LogicalPlanBuilder::from(table_scan)
+ .offset(10)?
+ .limit(1000)?
+ .limit(10)?
+ .build()?;
+
+ // Should push down the smallest limit
+ // Towards table scan
+ // This rule doesn't replace multiple limits
+ let expected = "Limit: 10\
+ \n Limit: 10\
+ \n Offset: 10\
+ \n TableScan: test projection=None, limit=20";
+
+ assert_optimized_plan_eq(&plan, expected);
+
+ Ok(())
+ }
+
+ #[test]
+ fn limit_doesnt_push_down_with_offset_aggregation() -> Result<()> {
+ let table_scan = test_table_scan()?;
+
+ let plan = LogicalPlanBuilder::from(table_scan)
+ .aggregate(vec![col("a")], vec![max(col("b"))])?
+ .offset(10)?
+ .limit(1000)?
+ .build()?;
+
+ // Limit should *not* push down aggregate node
+ let expected = "Limit: 1000\
+ \n Offset: 10\
+ \n Aggregate: groupBy=[[#test.a]], aggr=[[MAX(#test.b)]]\
+ \n TableScan: test projection=None";
+
+ assert_optimized_plan_eq(&plan, expected);
+
+ Ok(())
+ }
+
+ #[test]
+ fn limit_should_push_down_with_offset_union() -> Result<()> {
Review Comment:
I just wanted to be sure that we are not pushing down into these operators
as part of this PR and the tests would help confirm that. Follow-up tickets to
support this would be great.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]