[ 
https://issues.apache.org/jira/browse/CALCITE-6203?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Caican Cai updated CALCITE-6203:
--------------------------------
    Description: 
I test the avg operator in SparkAdapterTest
{quote}@Test void testGroupBy()

{     final String sql = "select sum(x) as SUM_X, min(y) as MIN_Y, max(y) as 
MAX_Y, avg(x) as AVG_X,"         + "count(*) as CNT_Y, count(distinct y) as 
CNT_DIST_Y\n"         + "from " + VALUES2 + "\n"         + "group by x";     
final String plan = "PLAN="         + "EnumerableCalc(expr#0..5=[\\{inputs}

], expr#6=[CAST($t1):INTEGER NOT NULL], expr#7=[CAST($t2):CHAR(1) NOT NULL], 
expr#8=[CAST($t3):CHAR(1) NOT NULL], expr#9=[CAST($t4):BIGINT NOT NULL], 
SUM_X=[$t6], MIN_Y=[$t7], MAX_Y=[$t8], AVG_X=[$t0], CNT_Y=[$t9], 
CNT_DIST_Y=[$t5])\n"
        + "  EnumerableAggregate(group=[\\{0}], SUM_X=[MIN($2) FILTER $7], 
MIN_Y=[MIN($3) FILTER $7], MAX_Y=[MIN($4) FILTER $7], CNT_Y=[MIN($5) FILTER 
$7], CNT_DIST_Y=[COUNT($1) FILTER $6])\n"
        + "    EnumerableCalc(expr#0..6=[\\{inputs}], expr#7=[0], 
expr#8=[=($t6, $t7)], expr#9=[1], expr#10=[=($t6, $t9)], proj#0..5=[\\{exprs}], 
$g_0=[$t8], $g_1=[$t10])\n"
        + "      EnumerableAggregate(group=[\\{0, 1}], groups=[[\\{0, 1}, 
\\{0}]], SUM_X=[$SUM0($0)], MIN_Y=[MIN($1)], MAX_Y=[MAX($1)], CNT_Y=[COUNT()], 
$g=[GROUPING($0, $1)])\n"
        + "        EnumerableValues(tuples=[[\\{ 1, 'a' }, \\{ 2, 'b' }, \\{ 1, 
'b' }, \\{ 2, 'c' }, \\{ 2, 'c' }]])\n";

    final String[] expectedResult =

{         "SUM_X=2; MIN_Y=a; MAX_Y=b; AVG_X=1; CNT_Y=2; CNT_DIST_Y=2",         
"SUM_X=6; MIN_Y=b; MAX_Y=c; AVG_X=2; CNT_Y=3; CNT_DIST_Y=2"     }

;

    sql(sql).returnsUnordered(expectedResult)
        .explainContains(plan);
  }
{quote}
 

It can be found that there is no corresponding avg operator in 
EnumerableAggregate and EnumerableCalc. Why is this? Spark sql will split the 
avg operator into sum/count form. I don’t know if this is a bug.

 

 

  was:
I test the avg operator in SparkAdapterTest

@Test void testGroupBy() {
    final String sql = "select sum(x) as SUM_X, min(y) as MIN_Y, max(y) as 
MAX_Y, avg(x) as AVG_X,"
        + "count(*) as CNT_Y, count(distinct y) as CNT_DIST_Y\n"
        + "from " + VALUES2 + "\n"
        + "group by x";

    final String plan = "PLAN="
        + "EnumerableCalc(expr#0..5=[\{inputs}], expr#6=[CAST($t1):INTEGER NOT 
NULL], expr#7=[CAST($t2):CHAR(1) NOT NULL], expr#8=[CAST($t3):CHAR(1) NOT 
NULL], expr#9=[CAST($t4):BIGINT NOT NULL], SUM_X=[$t6], MIN_Y=[$t7], 
MAX_Y=[$t8], AVG_X=[$t0], CNT_Y=[$t9], CNT_DIST_Y=[$t5])\n"
        + "  EnumerableAggregate(group=[\{0}], SUM_X=[MIN($2) FILTER $7], 
MIN_Y=[MIN($3) FILTER $7], MAX_Y=[MIN($4) FILTER $7], CNT_Y=[MIN($5) FILTER 
$7], CNT_DIST_Y=[COUNT($1) FILTER $6])\n"
        + "    EnumerableCalc(expr#0..6=[\{inputs}], expr#7=[0], expr#8=[=($t6, 
$t7)], expr#9=[1], expr#10=[=($t6, $t9)], proj#0..5=[\{exprs}], $g_0=[$t8], 
$g_1=[$t10])\n"
        + "      EnumerableAggregate(group=[\{0, 1}], groups=[[\{0, 1}, \{0}]], 
SUM_X=[$SUM0($0)], MIN_Y=[MIN($1)], MAX_Y=[MAX($1)], CNT_Y=[COUNT()], 
$g=[GROUPING($0, $1)])\n"
        + "        EnumerableValues(tuples=[[\{ 1, 'a' }, \{ 2, 'b' }, \{ 1, 
'b' }, \{ 2, 'c' }, \{ 2, 'c' }]])\n";

    final String[] expectedResult = {
        "SUM_X=2; MIN_Y=a; MAX_Y=b; AVG_X=1; CNT_Y=2; CNT_DIST_Y=2",
        "SUM_X=6; MIN_Y=b; MAX_Y=c; AVG_X=2; CNT_Y=3; CNT_DIST_Y=2"
    };

    sql(sql).returnsUnordered(expectedResult)
        .explainContains(plan);
  }

 

It can be found that there is no corresponding avg operator in 
EnumerableAggregate and EnumerableCalc. Why is this? Spark sql will split the 
avg operator into sum/count form. I don’t know if this is a bug.

 

 


> Avg operator test in SparkAdapterTest
> -------------------------------------
>
>                 Key: CALCITE-6203
>                 URL: https://issues.apache.org/jira/browse/CALCITE-6203
>             Project: Calcite
>          Issue Type: Test
>          Components: spark
>    Affects Versions: 1.36.0
>            Reporter: Caican Cai
>            Priority: Minor
>             Fix For: 1.37.0
>
>
> I test the avg operator in SparkAdapterTest
> {quote}@Test void testGroupBy()
> {     final String sql = "select sum(x) as SUM_X, min(y) as MIN_Y, max(y) as 
> MAX_Y, avg(x) as AVG_X,"         + "count(*) as CNT_Y, count(distinct y) as 
> CNT_DIST_Y\n"         + "from " + VALUES2 + "\n"         + "group by x";     
> final String plan = "PLAN="         + "EnumerableCalc(expr#0..5=[\\{inputs}
> ], expr#6=[CAST($t1):INTEGER NOT NULL], expr#7=[CAST($t2):CHAR(1) NOT NULL], 
> expr#8=[CAST($t3):CHAR(1) NOT NULL], expr#9=[CAST($t4):BIGINT NOT NULL], 
> SUM_X=[$t6], MIN_Y=[$t7], MAX_Y=[$t8], AVG_X=[$t0], CNT_Y=[$t9], 
> CNT_DIST_Y=[$t5])\n"
>         + "  EnumerableAggregate(group=[\\{0}], SUM_X=[MIN($2) FILTER $7], 
> MIN_Y=[MIN($3) FILTER $7], MAX_Y=[MIN($4) FILTER $7], CNT_Y=[MIN($5) FILTER 
> $7], CNT_DIST_Y=[COUNT($1) FILTER $6])\n"
>         + "    EnumerableCalc(expr#0..6=[\\{inputs}], expr#7=[0], 
> expr#8=[=($t6, $t7)], expr#9=[1], expr#10=[=($t6, $t9)], 
> proj#0..5=[\\{exprs}], $g_0=[$t8], $g_1=[$t10])\n"
>         + "      EnumerableAggregate(group=[\\{0, 1}], groups=[[\\{0, 1}, 
> \\{0}]], SUM_X=[$SUM0($0)], MIN_Y=[MIN($1)], MAX_Y=[MAX($1)], 
> CNT_Y=[COUNT()], $g=[GROUPING($0, $1)])\n"
>         + "        EnumerableValues(tuples=[[\\{ 1, 'a' }, \\{ 2, 'b' }, \\{ 
> 1, 'b' }, \\{ 2, 'c' }, \\{ 2, 'c' }]])\n";
>     final String[] expectedResult =
> {         "SUM_X=2; MIN_Y=a; MAX_Y=b; AVG_X=1; CNT_Y=2; CNT_DIST_Y=2",        
>  "SUM_X=6; MIN_Y=b; MAX_Y=c; AVG_X=2; CNT_Y=3; CNT_DIST_Y=2"     }
> ;
>     sql(sql).returnsUnordered(expectedResult)
>         .explainContains(plan);
>   }
> {quote}
>  
> It can be found that there is no corresponding avg operator in 
> EnumerableAggregate and EnumerableCalc. Why is this? Spark sql will split the 
> avg operator into sum/count form. I don’t know if this is a bug.
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to