Repository: hive Updated Branches: refs/heads/master 89e75c785 -> ddd4c9aea
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out b/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out index fa1480e..71ccea5 100644 --- a/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out +++ b/ql/src/test/results/clientpositive/spark/groupby_cube1.q.out @@ -42,21 +42,21 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() - keys: key (type: string), val (type: string), 0 (type: int) + keys: key (type: string), val (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE value expressions: _col3 (type: bigint) Reducer 2 Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE @@ -107,21 +107,21 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() - keys: key (type: string), val (type: string), 0 (type: int) + keys: key (type: string), val (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE value expressions: _col3 (type: bigint) Reducer 2 Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE @@ -198,26 +198,26 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() - keys: _col0 (type: string), _col1 (type: string), 0 (type: int) + keys: _col0 (type: string), _col1 (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE value expressions: _col3 (type: bigint) Reducer 2 Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: bigint) + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE File Output Operator @@ -288,20 +288,20 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT val) - keys: key (type: string), 0 (type: int), val (type: string) + keys: key (type: string), 0 (type: bigint), val (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string) + key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint) Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col2 Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE @@ -367,12 +367,12 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() - keys: key (type: string), val (type: string), 0 (type: int) + keys: key (type: string), val (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: rand() (type: double) Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE @@ -381,12 +381,12 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE @@ -395,7 +395,7 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE @@ -473,12 +473,12 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT val) - keys: key (type: string), 0 (type: int), val (type: string) + keys: key (type: string), 0 (type: bigint), val (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string) + key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE @@ -486,12 +486,12 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: int) + key expressions: _col0 (type: string), _col1 (type: bigint) sort order: ++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE @@ -500,7 +500,7 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) mode: final outputColumnNames: _col0, _col2 Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE @@ -591,12 +591,12 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(1) - keys: key (type: string), val (type: string), 0 (type: int) + keys: key (type: string), val (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: rand() (type: double) Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE @@ -612,12 +612,12 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(1) - keys: key (type: string), val (type: string), 0 (type: int) + keys: key (type: string), val (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: rand() (type: double) Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE @@ -626,12 +626,12 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE @@ -640,7 +640,7 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE @@ -661,12 +661,12 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) Statistics: Num rows: 4 Data size: 1200 Basic stats: COMPLETE Column stats: NONE @@ -675,7 +675,7 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/spark/groupby_rollup1.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/groupby_rollup1.q.out b/ql/src/test/results/clientpositive/spark/groupby_rollup1.q.out index 460e7db..cf3721d 100644 --- a/ql/src/test/results/clientpositive/spark/groupby_rollup1.q.out +++ b/ql/src/test/results/clientpositive/spark/groupby_rollup1.q.out @@ -42,21 +42,21 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() - keys: key (type: string), val (type: string), 0 (type: int) + keys: key (type: string), val (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE value expressions: _col3 (type: bigint) Reducer 2 Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE @@ -127,20 +127,20 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT val) - keys: key (type: string), 0 (type: int), val (type: string) + keys: key (type: string), 0 (type: bigint), val (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string) + key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint) Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col2 Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE @@ -206,12 +206,12 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() - keys: key (type: string), val (type: string), 0 (type: int) + keys: key (type: string), val (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: rand() (type: double) Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE @@ -220,12 +220,12 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE @@ -234,7 +234,7 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE @@ -306,12 +306,12 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT val) - keys: key (type: string), 0 (type: int), val (type: string) + keys: key (type: string), 0 (type: bigint), val (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string) + key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE @@ -319,12 +319,12 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: int) + key expressions: _col0 (type: string), _col1 (type: bigint) sort order: ++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 2 Data size: 600 Basic stats: COMPLETE Column stats: NONE @@ -333,7 +333,7 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) mode: final outputColumnNames: _col0, _col2 Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE @@ -424,12 +424,12 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(1) - keys: key (type: string), val (type: string), 0 (type: int) + keys: key (type: string), val (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: rand() (type: double) Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE @@ -445,12 +445,12 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(1) - keys: key (type: string), val (type: string), 0 (type: int) + keys: key (type: string), val (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: rand() (type: double) Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE @@ -459,12 +459,12 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE @@ -473,7 +473,7 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE @@ -494,12 +494,12 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) Statistics: Num rows: 3 Data size: 900 Basic stats: COMPLETE Column stats: NONE @@ -508,7 +508,7 @@ STAGE PLANS: Reduce Operator Tree: Group By Operator aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 1 Data size: 300 Basic stats: COMPLETE Column stats: NONE http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/spark/limit_pushdown2.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/limit_pushdown2.q.out b/ql/src/test/results/clientpositive/spark/limit_pushdown2.q.out index f6f0043..3bd49b4 100644 --- a/ql/src/test/results/clientpositive/spark/limit_pushdown2.q.out +++ b/ql/src/test/results/clientpositive/spark/limit_pushdown2.q.out @@ -966,21 +966,21 @@ STAGE PLANS: Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: avg(_col2) - keys: _col0 (type: string), _col1 (type: string), 0 (type: int) + keys: _col0 (type: string), _col1 (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE value expressions: _col3 (type: struct<count:bigint,sum:double,input:double>) Reducer 2 Reduce Operator Tree: Group By Operator aggregations: avg(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 750 Data size: 7968 Basic stats: COMPLETE Column stats: NONE @@ -1051,21 +1051,21 @@ STAGE PLANS: Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: avg(_col2) - keys: _col0 (type: string), _col1 (type: string), 0 (type: int) + keys: _col0 (type: string), _col1 (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE value expressions: _col3 (type: struct<count:bigint,sum:double,input:double>) Reducer 2 Reduce Operator Tree: Group By Operator aggregations: avg(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 750 Data size: 7968 Basic stats: COMPLETE Column stats: NONE http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/tez/multi_count_distinct.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/multi_count_distinct.q.out b/ql/src/test/results/clientpositive/tez/multi_count_distinct.q.out index bb08f16..5f5a852 100644 --- a/ql/src/test/results/clientpositive/tez/multi_count_distinct.q.out +++ b/ql/src/test/results/clientpositive/tez/multi_count_distinct.q.out @@ -43,14 +43,14 @@ Stage-0 PARTITION_ONLY_SHUFFLE [RS_9] Group By Operator [GBY_8] (rows=1 width=24) Output:["_col0","_col1","_col2"],aggregations:["count(_col0)","count(_col1)","count(_col2)"] - Select Operator [SEL_6] (rows=13 width=97) + Select Operator [SEL_6] (rows=13 width=101) Output:["_col0","_col1","_col2"] - Group By Operator [GBY_5] (rows=13 width=97) + Group By Operator [GBY_5] (rows=13 width=101) Output:["_col0","_col1","_col2","_col3"],keys:KEY._col0, KEY._col1, KEY._col2, KEY._col3 <-Map 1 [SIMPLE_EDGE] SHUFFLE [RS_4] PartitionCols:_col0, _col1, _col2, _col3 - Group By Operator [GBY_3] (rows=13 width=97) + Group By Operator [GBY_3] (rows=13 width=101) Output:["_col0","_col1","_col2","_col3"],keys:_col0, _col1, _col2, 0 Select Operator [SEL_1] (rows=9 width=93) Output:["_col0","_col1","_col2"] @@ -123,14 +123,14 @@ Stage-0 PARTITION_ONLY_SHUFFLE [RS_9] Group By Operator [GBY_8] (rows=1 width=40) Output:["_col0","_col1","_col2","_col3","_col4"],aggregations:["count(_col0)","count(_col1)","count(_col2)","count(_col3)","count(_col4)"] - Select Operator [SEL_6] (rows=22 width=97) + Select Operator [SEL_6] (rows=22 width=101) Output:["_col0","_col1","_col2","_col3","_col4"] - Group By Operator [GBY_5] (rows=22 width=97) + Group By Operator [GBY_5] (rows=22 width=101) Output:["_col0","_col1","_col2","_col3"],keys:KEY._col0, KEY._col1, KEY._col2, KEY._col3 <-Map 1 [SIMPLE_EDGE] SHUFFLE [RS_4] PartitionCols:_col0, _col1, _col2, _col3 - Group By Operator [GBY_3] (rows=22 width=97) + Group By Operator [GBY_3] (rows=22 width=101) Output:["_col0","_col1","_col2","_col3"],keys:_col0, _col1, _col2, 0 Select Operator [SEL_1] (rows=9 width=93) Output:["_col0","_col1","_col2"] http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/vector_grouping_sets.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/vector_grouping_sets.q.out b/ql/src/test/results/clientpositive/vector_grouping_sets.q.out index f08c270..fc8290e 100644 --- a/ql/src/test/results/clientpositive/vector_grouping_sets.q.out +++ b/ql/src/test/results/clientpositive/vector_grouping_sets.q.out @@ -162,18 +162,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 1:string, ConstantVectorExpression(val 0) -> 30:int + keyExpressions: col 1:string, ConstantVectorExpression(val 0) -> 30:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: s_store_id (type: string), 0 (type: int) + keys: s_store_id (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1 Statistics: Num rows: 24 Data size: 51264 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: int) + key expressions: _col0 (type: string), _col1 (type: bigint) sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkOperator native: false @@ -196,7 +196,7 @@ STAGE PLANS: enableConditionsNotMet: hive.execution.engine mr IN [tez, spark] IS false Reduce Operator Tree: Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) mode: mergepartial outputColumnNames: _col0 Statistics: Num rows: 12 Data size: 25632 Basic stats: COMPLETE Column stats: NONE @@ -273,18 +273,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 1:string, ConstantVectorExpression(val 0) -> 30:int + keyExpressions: col 1:string, ConstantVectorExpression(val 0) -> 30:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: _col0 (type: string), 0 (type: int) + keys: _col0 (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1 Statistics: Num rows: 24 Data size: 51264 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: int) + key expressions: _col0 (type: string), _col1 (type: bigint) sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkOperator native: false @@ -307,12 +307,12 @@ STAGE PLANS: enableConditionsNotMet: hive.execution.engine mr IN [tez, spark] IS false Reduce Operator Tree: Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 12 Data size: 25632 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: string), _col1 (type: int) + expressions: _col0 (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 Statistics: Num rows: 12 Data size: 25632 Basic stats: COMPLETE Column stats: NONE File Output Operator @@ -374,24 +374,24 @@ STAGE PLANS: outputColumnNames: _col0 Statistics: Num rows: 12 Data size: 25632 Basic stats: COMPLETE Column stats: NONE Group By Operator - keys: _col0 (type: string), 0 (type: int) + keys: _col0 (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1 Statistics: Num rows: 24 Data size: 51264 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: int) + key expressions: _col0 (type: string), _col1 (type: bigint) sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint) Statistics: Num rows: 24 Data size: 51264 Basic stats: COMPLETE Column stats: NONE Execution mode: vectorized Reduce Operator Tree: Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 12 Data size: 25632 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: string), _col1 (type: int) + expressions: _col0 (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 Statistics: Num rows: 12 Data size: 25632 Basic stats: COMPLETE Column stats: NONE File Output Operator http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/view_cbo.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/view_cbo.q.out b/ql/src/test/results/clientpositive/view_cbo.q.out index e10cb7a..2fafa4f 100644 --- a/ql/src/test/results/clientpositive/view_cbo.q.out +++ b/ql/src/test/results/clientpositive/view_cbo.q.out @@ -26,20 +26,20 @@ STAGE PLANS: Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: avg(_col2) - keys: _col0 (type: string), _col1 (type: string), 0 (type: int) + keys: _col0 (type: string), _col1 (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Statistics: Num rows: 1500 Data size: 15936 Basic stats: COMPLETE Column stats: NONE value expressions: _col3 (type: struct<count:bigint,sum:double,input:double>) Reduce Operator Tree: Group By Operator aggregations: avg(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 750 Data size: 7968 Basic stats: COMPLETE Column stats: NONE