http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id2.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id2.q.out b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id2.q.out index e6075c7..74f6289 100644 --- a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id2.q.out +++ b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id2.q.out @@ -73,16 +73,16 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: rand() (type: double) Reduce Sink Vectorization: @@ -123,7 +123,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -133,16 +133,16 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: PARTIALS - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [0] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: int) Reduce Sink Vectorization: @@ -166,7 +166,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -176,16 +176,16 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: FINAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [0] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: bigint) + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), _col3 (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3 Select Vectorization: className: VectorSelectOperator @@ -275,16 +275,16 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: rand() (type: double) Reduce Sink Vectorization: @@ -325,7 +325,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -335,16 +335,16 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: PARTIALS - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [0] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: int) Reduce Sink Vectorization: @@ -368,7 +368,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -378,16 +378,16 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: FINAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [0] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: bigint) + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), _col3 (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3 Select Vectorization: className: VectorSelectOperator @@ -487,16 +487,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: rand() (type: double) Reduce Sink Vectorization: @@ -536,7 +536,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -544,16 +544,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: PARTIALS - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: int) Reduce Sink Vectorization: @@ -576,7 +576,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -584,16 +584,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: FINAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col2 (type: int) + expressions: _col2 (type: bigint) outputColumnNames: _col2 Select Vectorization: className: VectorSelectOperator @@ -606,16 +606,16 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 2:int + keyExpressions: col 2:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: _col2 (type: int) + keys: _col2 (type: bigint) mode: hash outputColumnNames: _col0, _col1 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int) + key expressions: _col0 (type: bigint) sort order: + Map-reduce partition columns: rand() (type: double) Reduce Sink Vectorization: @@ -639,7 +639,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 2 - dataColumns: KEY._col0:int, VALUE._col0:bigint + dataColumns: KEY._col0:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -649,18 +649,18 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 1:bigint) -> bigint className: VectorGroupByOperator groupByMode: PARTIALS - keyExpressions: col 0:int + keyExpressions: col 0:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [0] - keys: KEY._col0 (type: int) + keys: KEY._col0 (type: bigint) mode: partials outputColumnNames: _col0, _col1 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int) + key expressions: _col0 (type: bigint) sort order: + - Map-reduce partition columns: _col0 (type: int) + Map-reduce partition columns: _col0 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkLongOperator keyColumnNums: [0] @@ -681,7 +681,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 2 - dataColumns: KEY._col0:int, VALUE._col0:bigint + dataColumns: KEY._col0:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -691,11 +691,11 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 1:bigint) -> bigint className: VectorGroupByOperator groupByMode: FINAL - keyExpressions: col 0:int + keyExpressions: col 0:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [0] - keys: KEY._col0 (type: int) + keys: KEY._col0 (type: bigint) mode: final outputColumnNames: _col0, _col1 Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE @@ -794,16 +794,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: rand() (type: double) Reduce Sink Vectorization: @@ -843,7 +843,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -851,16 +851,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: PARTIALS - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: int) Reduce Sink Vectorization: @@ -883,7 +883,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -891,16 +891,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: FINAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col2 (type: int) + expressions: _col2 (type: bigint) outputColumnNames: _col2 Select Vectorization: className: VectorSelectOperator @@ -913,16 +913,16 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 2:int + keyExpressions: col 2:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: _col2 (type: int) + keys: _col2 (type: bigint) mode: hash outputColumnNames: _col0, _col1 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int) + key expressions: _col0 (type: bigint) sort order: + Map-reduce partition columns: rand() (type: double) Reduce Sink Vectorization: @@ -946,7 +946,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 2 - dataColumns: KEY._col0:int, VALUE._col0:bigint + dataColumns: KEY._col0:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -956,18 +956,18 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 1:bigint) -> bigint className: VectorGroupByOperator groupByMode: PARTIALS - keyExpressions: col 0:int + keyExpressions: col 0:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [0] - keys: KEY._col0 (type: int) + keys: KEY._col0 (type: bigint) mode: partials outputColumnNames: _col0, _col1 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int) + key expressions: _col0 (type: bigint) sort order: + - Map-reduce partition columns: _col0 (type: int) + Map-reduce partition columns: _col0 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkLongOperator keyColumnNums: [0] @@ -988,7 +988,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 2 - dataColumns: KEY._col0:int, VALUE._col0:bigint + dataColumns: KEY._col0:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -998,11 +998,11 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 1:bigint) -> bigint className: VectorGroupByOperator groupByMode: FINAL - keyExpressions: col 0:int + keyExpressions: col 0:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [0] - keys: KEY._col0 (type: int) + keys: KEY._col0 (type: bigint) mode: final outputColumnNames: _col0, _col1 Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE @@ -1097,16 +1097,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: rand() (type: double) Reduce Sink Vectorization: @@ -1146,7 +1146,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1154,16 +1154,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: PARTIALS - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: int) Reduce Sink Vectorization: @@ -1175,7 +1175,7 @@ STAGE PLANS: valueColumnNums: [] Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: int) Reduce Sink Vectorization: @@ -1198,7 +1198,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1206,16 +1206,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: FINAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col2 (type: int) + expressions: _col2 (type: bigint) outputColumnNames: _col0 Select Vectorization: className: VectorSelectOperator @@ -1223,9 +1223,9 @@ STAGE PLANS: projectedOutputColumnNums: [2] Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int) + key expressions: _col0 (type: bigint) sort order: + - Map-reduce partition columns: _col0 (type: int) + Map-reduce partition columns: _col0 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkLongOperator keyColumnNums: [2] @@ -1240,8 +1240,8 @@ STAGE PLANS: condition map: Inner Join 0 to 1 keys: - 0 _col0 (type: int) - 1 _col0 (type: int) + 0 _col0 (type: bigint) + 1 _col0 (type: bigint) outputColumnNames: _col0, _col1 Statistics: Num rows: 9 Data size: 79 Basic stats: COMPLETE Column stats: NONE File Output Operator @@ -1263,7 +1263,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1271,16 +1271,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: FINAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col2 (type: int) + expressions: _col2 (type: bigint) outputColumnNames: _col0 Select Vectorization: className: VectorSelectOperator @@ -1288,9 +1288,9 @@ STAGE PLANS: projectedOutputColumnNums: [2] Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int) + key expressions: _col0 (type: bigint) sort order: + - Map-reduce partition columns: _col0 (type: int) + Map-reduce partition columns: _col0 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkLongOperator keyColumnNums: [2] @@ -1425,16 +1425,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: rand() (type: double) Reduce Sink Vectorization: @@ -1474,7 +1474,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1482,16 +1482,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: PARTIALS - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: partials outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: int) Reduce Sink Vectorization: @@ -1503,7 +1503,7 @@ STAGE PLANS: valueColumnNums: [] Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ Map-reduce partition columns: _col0 (type: int), _col1 (type: int) Reduce Sink Vectorization: @@ -1526,7 +1526,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1534,16 +1534,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: FINAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col2 (type: int) + expressions: _col2 (type: bigint) outputColumnNames: _col0 Select Vectorization: className: VectorSelectOperator @@ -1551,9 +1551,9 @@ STAGE PLANS: projectedOutputColumnNums: [2] Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int) + key expressions: _col0 (type: bigint) sort order: + - Map-reduce partition columns: _col0 (type: int) + Map-reduce partition columns: _col0 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkLongOperator keyColumnNums: [2] @@ -1568,8 +1568,8 @@ STAGE PLANS: condition map: Inner Join 0 to 1 keys: - 0 _col0 (type: int) - 1 _col0 (type: int) + 0 _col0 (type: bigint) + 1 _col0 (type: bigint) outputColumnNames: _col0, _col1 Statistics: Num rows: 9 Data size: 79 Basic stats: COMPLETE Column stats: NONE File Output Operator @@ -1591,7 +1591,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1599,16 +1599,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: FINAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: STREAMING projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: final outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col2 (type: int) + expressions: _col2 (type: bigint) outputColumnNames: _col0 Select Vectorization: className: VectorSelectOperator @@ -1616,9 +1616,9 @@ STAGE PLANS: projectedOutputColumnNums: [2] Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int) + key expressions: _col0 (type: bigint) sort order: + - Map-reduce partition columns: _col0 (type: int) + Map-reduce partition columns: _col0 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkLongOperator keyColumnNums: [2] @@ -1746,18 +1746,18 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -1795,7 +1795,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1805,16 +1805,16 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [0] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: bigint) + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), _col3 (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3 Select Vectorization: className: VectorSelectOperator @@ -1912,18 +1912,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -1960,7 +1960,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1968,16 +1968,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col2 (type: int) + expressions: _col2 (type: bigint) outputColumnNames: _col2 Select Vectorization: className: VectorSelectOperator @@ -1990,18 +1990,18 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 2:int + keyExpressions: col 2:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: _col2 (type: int) + keys: _col2 (type: bigint) mode: hash outputColumnNames: _col0, _col1 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int) + key expressions: _col0 (type: bigint) sort order: + - Map-reduce partition columns: _col0 (type: int) + Map-reduce partition columns: _col0 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkLongOperator keyColumnNums: [0] @@ -2022,7 +2022,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 2 - dataColumns: KEY._col0:int, VALUE._col0:bigint + dataColumns: KEY._col0:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -2032,11 +2032,11 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 1:bigint) -> bigint className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int + keyExpressions: col 0:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [0] - keys: KEY._col0 (type: int) + keys: KEY._col0 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE @@ -2130,18 +2130,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -2150,9 +2150,9 @@ STAGE PLANS: valueColumnNums: [] Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -2189,7 +2189,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -2197,16 +2197,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col2 (type: int) + expressions: _col2 (type: bigint) outputColumnNames: _col0 Select Vectorization: className: VectorSelectOperator @@ -2214,9 +2214,9 @@ STAGE PLANS: projectedOutputColumnNums: [2] Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int) + key expressions: _col0 (type: bigint) sort order: + - Map-reduce partition columns: _col0 (type: int) + Map-reduce partition columns: _col0 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkLongOperator keyColumnNums: [2] @@ -2231,8 +2231,8 @@ STAGE PLANS: condition map: Inner Join 0 to 1 keys: - 0 _col0 (type: int) - 1 _col0 (type: int) + 0 _col0 (type: bigint) + 1 _col0 (type: bigint) outputColumnNames: _col0, _col1 Statistics: Num rows: 9 Data size: 79 Basic stats: COMPLETE Column stats: NONE File Output Operator @@ -2254,7 +2254,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -2262,16 +2262,16 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col2 (type: int) + expressions: _col2 (type: bigint) outputColumnNames: _col0 Select Vectorization: className: VectorSelectOperator @@ -2279,9 +2279,9 @@ STAGE PLANS: projectedOutputColumnNums: [2] Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int) + key expressions: _col0 (type: bigint) sort order: + - Map-reduce partition columns: _col0 (type: int) + Map-reduce partition columns: _col0 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkLongOperator keyColumnNums: [2]
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id3.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id3.q.out b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id3.q.out index da4b81f..da4ba84 100644 --- a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id3.q.out +++ b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_id3.q.out @@ -80,11 +80,11 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: key (type: int), value (type: int), 0 (type: int) + keys: key (type: int), value (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE @@ -92,17 +92,17 @@ STAGE PLANS: Filter Vectorization: className: VectorFilterOperator native: true - predicateExpression: FilterLongColEqualLongScalar(col 2:int, val 1) + predicateExpression: FilterLongColEqualLongScalar(col 2:bigint, val 1) predicate: (_col2 = 1) (type: boolean) Statistics: Num rows: 6 Data size: 48 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), 1 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), 1 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), 1 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), 1 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 4] - keyExpressions: ConstantVectorExpression(val 1) -> 4:int + keyExpressions: ConstantVectorExpression(val 1) -> 4:bigint native: true nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true valueColumnNums: [3] @@ -137,7 +137,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [bigint] Reduce Operator Tree: @@ -147,23 +147,23 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 1) -> 4:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 1) -> 4:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [0] - keys: KEY._col0 (type: int), KEY._col1 (type: int), 1 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), 1 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE pruneGroupingSetId: true Select Operator - expressions: _col0 (type: int), _col1 (type: int), 1 (type: int), _col3 (type: bigint) + expressions: _col0 (type: int), _col1 (type: int), 1 (type: bigint), _col3 (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3 Select Vectorization: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 1, 3, 2] - selectExpressions: ConstantVectorExpression(val 1) -> 3:int + selectExpressions: ConstantVectorExpression(val 1) -> 3:bigint Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false @@ -256,18 +256,18 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -305,7 +305,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -315,11 +315,11 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [0] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 6 Data size: 48 Basic stats: COMPLETE Column stats: NONE @@ -327,17 +327,17 @@ STAGE PLANS: Filter Vectorization: className: VectorFilterOperator native: true - predicateExpression: FilterLongColEqualLongScalar(col 2:int, val 1) + predicateExpression: FilterLongColEqualLongScalar(col 2:bigint, val 1) predicate: (_col2 = 1) (type: boolean) Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), 1 (type: int), _col3 (type: bigint) + expressions: _col0 (type: int), _col1 (type: int), 1 (type: bigint), _col3 (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3 Select Vectorization: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 1, 4, 3] - selectExpressions: ConstantVectorExpression(val 1) -> 4:int + selectExpressions: ConstantVectorExpression(val 1) -> 4:bigint Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets1.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets1.q.out b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets1.q.out index d2b738b..137e06a 100644 --- a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets1.q.out +++ b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets1.q.out @@ -88,18 +88,18 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int + keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: a (type: string), b (type: string), 0 (type: int) + keys: a (type: string), b (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 24 Data size: 8832 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -137,7 +137,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -147,11 +147,11 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:string, col 1:string, col 2:int + keyExpressions: col 0:string, col 1:string, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [0] - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE @@ -250,18 +250,18 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int + keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: a (type: string), b (type: string), 0 (type: int) + keys: a (type: string), b (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 24 Data size: 8832 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -299,7 +299,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -309,11 +309,11 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:string, col 1:string, col 2:int + keyExpressions: col 0:string, col 1:string, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [0] - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE @@ -412,18 +412,18 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int + keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: a (type: string), b (type: string), 0 (type: int) + keys: a (type: string), b (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 24 Data size: 8832 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -461,7 +461,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -471,11 +471,11 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:string, col 1:string, col 2:int + keyExpressions: col 0:string, col 1:string, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [0] - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE @@ -574,18 +574,18 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int + keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: a (type: string), b (type: string), 0 (type: int) + keys: a (type: string), b (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -623,7 +623,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -633,11 +633,11 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:string, col 1:string, col 2:int + keyExpressions: col 0:string, col 1:string, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [0] - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE @@ -730,18 +730,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:string, col 1:string, col 2:string, ConstantVectorExpression(val 0) -> 4:int + keyExpressions: col 0:string, col 1:string, col 2:string, ConstantVectorExpression(val 0) -> 4:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: a (type: string), b (type: string), c (type: string), 0 (type: int) + keys: a (type: string), b (type: string), c (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 18 Data size: 9936 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) sort order: ++++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2, 3] @@ -778,7 +778,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:string, KEY._col3:int + dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:string, KEY._col3:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -786,11 +786,11 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:string, col 1:string, col 2:string, col 3:int + keyExpressions: col 0:string, col 1:string, col 2:string, col 3:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 4968 Basic stats: COMPLETE Column stats: NONE