This is an automated email from the ASF dual-hosted git repository.

maxyang pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/cloudberry.git

commit 41b93eebbc22e3efd45f5801cd4287dd930ffecf
Author: zhoujiaqi <[email protected]>
AuthorDate: Thu Jan 2 11:18:51 2025 +0800

    Fix icw test from "Cherry-Pick CUBE grouping set, predicates from subquery 
..."
---
 src/test/regress/expected/dpe_optimizer.out        |  3 +-
 .../regress/expected/groupingsets_optimizer.out    | 49 ++++------------------
 src/test/regress/expected/join_optimizer.out       |  3 +-
 src/test/regress/expected/window_optimizer.out     |  6 ++-
 4 files changed, 15 insertions(+), 46 deletions(-)

diff --git a/src/test/regress/expected/dpe_optimizer.out 
b/src/test/regress/expected/dpe_optimizer.out
index 08978fd662..46ee8c65e5 100644
--- a/src/test/regress/expected/dpe_optimizer.out
+++ b/src/test/regress/expected/dpe_optimizer.out
@@ -278,6 +278,7 @@ explain (costs off, timing off, summary off, analyze) 
select * from pt where exi
          Extra Text: (seg1)   Hash chain length 1.0 avg, 1 max, using 2 of 
262144 buckets.
          ->  Dynamic Seq Scan on pt (actual rows=8 loops=1)
                Number of partitions to scan: 6 
+               Filter: (NOT (ptid IS NULL))
                Partitions scanned:  Avg 2.0 x 3 workers.  Max 2 parts (seg0).
          ->  Hash (actual rows=2 loops=1)
                Buckets: 262144  Batches: 1  Memory Usage: 2049kB
@@ -293,7 +294,7 @@ explain (costs off, timing off, summary off, analyze) 
select * from pt where exi
                                              ->  Seq Scan on t (actual rows=2 
loops=1)
                                                    Filter: (t1 = 
('hello'::text || (tid)::text))
  Optimizer: Pivotal Optimizer (GPORCA)
-(21 rows)
+(22 rows)
 
 select * from pt where exists (select 1 from t where tid = ptid and t1 = 
'hello' || tid);
  dist |   pt1   |  pt2  |    pt3    | ptid 
diff --git a/src/test/regress/expected/groupingsets_optimizer.out 
b/src/test/regress/expected/groupingsets_optimizer.out
index 6a1f249c21..f4990cd25b 100644
--- a/src/test/regress/expected/groupingsets_optimizer.out
+++ b/src/test/regress/expected/groupingsets_optimizer.out
@@ -1670,16 +1670,12 @@ explain (costs off)
                                              Sort Key: share0_ref3.b
                                              ->  Redistribute Motion 3:3  
(slice2; segments: 3)
                                                    Hash Key: share0_ref3.b
-                                                   ->  Partial GroupAggregate
+                                                   ->  Streaming Partial 
HashAggregate
                                                          Group Key: 
share0_ref3.b
-                                                         ->  Sort
-                                                               Sort Key: 
share0_ref3.b
-                                                               ->  Shared Scan 
(share slice:id 2:0)
-                                 ->  GroupAggregate
+                                                         ->  Shared Scan 
(share slice:id 2:0)
+                                 ->  HashAggregate
                                        Group Key: share0_ref4.a
-                                       ->  Sort
-                                             Sort Key: share0_ref4.a
-                                             ->  Shared Scan (share slice:id 
1:0)
+                                       ->  Shared Scan (share slice:id 1:0)
                                  ->  Result
                                        ->  Redistribute Motion 1:3  (slice3)
                                              ->  Finalize Aggregate
@@ -1687,7 +1683,7 @@ explain (costs off)
                                                          ->  Partial Aggregate
                                                                ->  Shared Scan 
(share slice:id 4:0)
  Optimizer: Pivotal Optimizer (GPORCA)
-(40 rows)
+(36 rows)
 
 select a, b, sum(v.x)
   from (values (1),(2)) v(x), gstest_data(v.x)
@@ -2388,23 +2384,7 @@ create table gs_group_1 as
 select g1000, g100, g10, sum(g::numeric), count(*), max(g::text) from
   (select g%1000 as g1000, g%100 as g100, g%10 as g10, g
    from generate_series(0,199999) g) s
-group by cube (g1000,g100,g10);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column(s) named 
'g1000' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make 
sure column(s) chosen are the optimal data distribution key to minimize skew.
-set jit_above_cost to default;
-create table gs_group_2 as
-select g1000, g100, g10, sum(g::numeric), count(*), max(g::text) from
-  (select g/20 as g1000, g/200 as g100, g/2000 as g10, g
-   from generate_series(0,19999) g) s
-group by cube (g1000,g100,g10);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column(s) named 
'g1000' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make 
sure column(s) chosen are the optimal data distribution key to minimize skew.
-create table gs_group_3 as
-select g100, g10, array_agg(g) as a, count(*) as c, max(g::text) as m from
-  (select g/200 as g100, g/2000 as g10, g
-   from generate_series(0,19999) g) s
-group by grouping sets (g100,g10);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause. Creating a NULL policy 
entry.
+group by cube (g1000,g100,g10) distributed by (g1000);
 -- Produce results with hash aggregation.
 set enable_hashagg = true;
 set enable_sort = false;
@@ -2449,23 +2429,8 @@ create table gs_hash_1 as
 select g1000, g100, g10, sum(g::numeric), count(*), max(g::text) from
   (select g%1000 as g1000, g%100 as g100, g%10 as g10, g
    from generate_series(0,199999) g) s
-group by cube (g1000,g100,g10);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column(s) named 
'g1000' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make 
sure column(s) chosen are the optimal data distribution key to minimize skew.
+group by cube (g1000,g100,g10) distributed by (g1000);
 set jit_above_cost to default;
-create table gs_hash_2 as
-select g1000, g100, g10, sum(g::numeric), count(*), max(g::text) from
-  (select g/20 as g1000, g/200 as g100, g/2000 as g10, g
-   from generate_series(0,19999) g) s
-group by cube (g1000,g100,g10);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause -- Using column(s) named 
'g1000' as the Greenplum Database data distribution key for this table.
-HINT:  The 'DISTRIBUTED BY' clause determines the distribution of data. Make 
sure column(s) chosen are the optimal data distribution key to minimize skew.
-create table gs_hash_3 as
-select g100, g10, array_agg(g) as a, count(*) as c, max(g::text) as m from
-  (select g/200 as g100, g/2000 as g10, g
-   from generate_series(0,19999) g) s
-group by grouping sets (g100,g10);
-NOTICE:  Table doesn't have 'DISTRIBUTED BY' clause. Creating a NULL policy 
entry.
 set enable_sort = true;
 set work_mem to default;
 WARNING:  "work_mem": setting is deprecated, and may be removed in a future 
release.
diff --git a/src/test/regress/expected/join_optimizer.out 
b/src/test/regress/expected/join_optimizer.out
index e7db125d49..503c0af151 100644
--- a/src/test/regress/expected/join_optimizer.out
+++ b/src/test/regress/expected/join_optimizer.out
@@ -1933,8 +1933,9 @@ where exists(select * from tenk1 b
                                              ->  Seq Scan on int4_tbl
                                        ->  Index Scan using 
tenk1_thous_tenthous on tenk1 tenk1_1
                                              Index Cond: (tenthous = 
int4_tbl.f1)
+                                             Filter: (NOT (twothousand IS 
NULL))
  Optimizer: Pivotal Optimizer (GPORCA)
-(18 rows)
+(19 rows)
 
 --
 -- More complicated constructs
diff --git a/src/test/regress/expected/window_optimizer.out 
b/src/test/regress/expected/window_optimizer.out
index 78dd907220..59c2635901 100644
--- a/src/test/regress/expected/window_optimizer.out
+++ b/src/test/regress/expected/window_optimizer.out
@@ -3794,11 +3794,11 @@ WINDOW fwd AS (
 INFO:  GPORCA failed to produce a plan, falling back to planner
 DETAIL:  GPDB Expression type: Query Parameter not supported in DXL
 INFO:  GPORCA failed to produce a plan, falling back to planner
-DETAIL:  GPDB Expression type: Query Parameter not supported in DXL
-INFO:  GPORCA failed to produce a plan, falling back to planner
 DETAIL:  Feature not supported: SIRV functions
 INFO:  GPORCA failed to produce a plan, falling back to planner
 DETAIL:  Feature not supported: SIRV functions
+INFO:  GPORCA failed to produce a plan, falling back to planner
+DETAIL:  GPDB Expression type: Query Parameter not supported in DXL
  eq1 | eq2 | eq3 
 -----+-----+-----
  t   | t   | t
@@ -4321,6 +4321,8 @@ EXPLAIN (costs off) SELECT * FROM pg_temp.f(2);
 (2 rows)
 
 SELECT * FROM pg_temp.f(2);
+INFO:  GPORCA failed to produce a plan, falling back to planner
+DETAIL:  GPDB Expression type: Query Parameter not supported in DXL
     f    
 ---------
  {1,2,3}


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to