http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_grouping.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_grouping.q.out b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_grouping.q.out index b81a0d3..8dd5cf0 100644 --- a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_grouping.q.out +++ b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_grouping.q.out @@ -74,18 +74,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -122,7 +122,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -130,22 +130,22 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 1) (type: int), grouping(_col2, 0) (type: int) + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1) (type: bigint), grouping(_col2, 0) (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3, _col4 Select Vectorization: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 1, 2, 3, 4] - selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int + selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false @@ -235,18 +235,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -283,7 +283,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -291,22 +291,22 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 1) (type: int), grouping(_col2, 0) (type: int) + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1) (type: bigint), grouping(_col2, 0) (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3, _col4 Select Vectorization: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 1, 2, 3, 4] - selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int + selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false @@ -403,18 +403,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -451,7 +451,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -459,11 +459,11 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE @@ -471,7 +471,7 @@ STAGE PLANS: Filter Vectorization: className: VectorFilterOperator native: true - predicateExpression: FilterLongColEqualLongScalar(col 3:int, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int) + predicateExpression: FilterLongColEqualLongScalar(col 3:bigint, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint) predicate: (grouping(_col2, 1) = 1) (type: boolean) Statistics: Num rows: 6 Data size: 48 Basic stats: COMPLETE Column stats: NONE Select Operator @@ -572,18 +572,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: _col0 (type: int), _col1 (type: int), 0 (type: int) + keys: _col0 (type: int), _col1 (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -620,7 +620,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -628,11 +628,11 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE @@ -640,20 +640,20 @@ STAGE PLANS: Filter Vectorization: className: VectorFilterOperator native: true - predicateExpression: FilterExprOrExpr(children: FilterLongColEqualLongScalar(col 3:int, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int), FilterLongColEqualLongScalar(col 3:int, val 1)(children: VectorUDFAdaptor(grouping(_col2, 0)) -> 3:int)) + predicateExpression: FilterExprOrExpr(children: FilterLongColEqualLongScalar(col 3:bigint, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint), FilterLongColEqualLongScalar(col 3:bigint, val 1)(children: VectorUDFAdaptor(grouping(_col2, 0)) -> 3:bigint)) predicate: ((grouping(_col2, 0) = 1) or (grouping(_col2, 1) = 1)) (type: boolean) Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), (grouping(_col2, 1) + grouping(_col2, 0)) (type: int), CASE WHEN (((grouping(_col2, 1) + grouping(_col2, 0)) = 1)) THEN (_col0) ELSE (null) END (type: int) + expressions: _col0 (type: int), _col1 (type: int), (grouping(_col2, 1) + grouping(_col2, 0)) (type: bigint), CASE WHEN (((grouping(_col2, 1) + grouping(_col2, 0)) = 1)) THEN (_col0) ELSE (null) END (type: int) outputColumnNames: _col0, _col1, _col2, _col3 Select Vectorization: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 1, 5, 4] - selectExpressions: LongColAddLongColumn(col 3:int, col 4:int)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int) -> 5:int, IfExprColumnNull(col 3:boolean, col 0:int, null)(children: LongColEqualLongScalar(col 6:int, val 1)(children: LongColAddLongColumn(col 3:int, col 4:int)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int) -> 6:int) -> 3:boolean, col 0:int) -> 4:int + selectExpressions: LongColAddLongColumn(col 3:bigint, col 4:bigint)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint) -> 5:bigint, IfExprColumnNull(col 3:boolean, col 0:int, null)(children: LongColEqualLongScalar(col 6:bigint, val 1)(children: LongColAddLongColumn(col 3:bigint, col 4:bigint)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint) -> 6:bigint) -> 3:boolean, col 0:int) -> 4:int Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col2 (type: int), _col3 (type: int) + key expressions: _col2 (type: bigint), _col3 (type: int) sort order: -+ Reduce Sink Vectorization: className: VectorReduceSinkObjectHashOperator @@ -675,12 +675,12 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY.reducesinkkey0:int, KEY.reducesinkkey1:int, VALUE._col0:int, VALUE._col1:int + dataColumns: KEY.reducesinkkey0:bigint, KEY.reducesinkkey1:int, VALUE._col0:int, VALUE._col1:int partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: Select Operator - expressions: VALUE._col0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey0 (type: int) + expressions: VALUE._col0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey0 (type: bigint) outputColumnNames: _col0, _col1, _col2 Select Vectorization: className: VectorSelectOperator @@ -778,18 +778,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: key (type: int), value (type: int), 0 (type: int) + keys: key (type: int), value (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -826,7 +826,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -834,22 +834,22 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 1) (type: int), grouping(_col2, 0) (type: int) + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1) (type: bigint), grouping(_col2, 0) (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3, _col4 Select Vectorization: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 1, 2, 3, 4] - selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int + selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false @@ -939,18 +939,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: key (type: int), value (type: int), 0 (type: int) + keys: key (type: int), value (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -987,7 +987,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -995,22 +995,22 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 1) (type: int), grouping(_col2, 0) (type: int) + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1) (type: bigint), grouping(_col2, 0) (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3, _col4 Select Vectorization: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 1, 2, 3, 4] - selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int + selectExpressions: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false @@ -1107,11 +1107,11 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: key (type: int), value (type: int), 0 (type: int) + keys: key (type: int), value (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE @@ -1119,13 +1119,13 @@ STAGE PLANS: Filter Vectorization: className: VectorFilterOperator native: true - predicateExpression: FilterLongColEqualLongScalar(col 3:int, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int) + predicateExpression: FilterLongColEqualLongScalar(col 3:bigint, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint) predicate: (grouping(_col2, 1) = 1) (type: boolean) Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -1162,7 +1162,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1170,11 +1170,11 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 6 Data size: 48 Basic stats: COMPLETE Column stats: NONE @@ -1269,11 +1269,11 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: key (type: int), value (type: int), 0 (type: int) + keys: key (type: int), value (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE @@ -1281,13 +1281,13 @@ STAGE PLANS: Filter Vectorization: className: VectorFilterOperator native: true - predicateExpression: FilterExprOrExpr(children: FilterLongColEqualLongScalar(col 3:int, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int), FilterLongColEqualLongScalar(col 3:int, val 1)(children: VectorUDFAdaptor(grouping(_col2, 0)) -> 3:int)) + predicateExpression: FilterExprOrExpr(children: FilterLongColEqualLongScalar(col 3:bigint, val 1)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint), FilterLongColEqualLongScalar(col 3:bigint, val 1)(children: VectorUDFAdaptor(grouping(_col2, 0)) -> 3:bigint)) predicate: ((grouping(_col2, 0) = 1) or (grouping(_col2, 1) = 1)) (type: boolean) Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -1324,7 +1324,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1332,30 +1332,30 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), (grouping(_col2, 1) + grouping(_col2, 0)) (type: int) + expressions: _col0 (type: int), _col1 (type: int), (grouping(_col2, 1) + grouping(_col2, 0)) (type: bigint) outputColumnNames: _col0, _col1, _col2 Select Vectorization: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 1, 5] - selectExpressions: LongColAddLongColumn(col 3:int, col 4:int)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:int, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:int) -> 5:int + selectExpressions: LongColAddLongColumn(col 3:bigint, col 4:bigint)(children: VectorUDFAdaptor(grouping(_col2, 1)) -> 3:bigint, VectorUDFAdaptor(grouping(_col2, 0)) -> 4:bigint) -> 5:bigint Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col2 (type: int), CASE WHEN ((_col2 = 1)) THEN (_col0) END (type: int) + key expressions: _col2 (type: bigint), CASE WHEN ((_col2 = 1)) THEN (_col0) END (type: int) sort order: -+ Reduce Sink Vectorization: className: VectorReduceSinkObjectHashOperator keyColumnNums: [5, 4] - keyExpressions: IfExprColumnNull(col 3:boolean, col 0:int, null)(children: LongColEqualLongScalar(col 5:int, val 1) -> 3:boolean, col 0:int) -> 4:int + keyExpressions: IfExprColumnNull(col 3:boolean, col 0:int, null)(children: LongColEqualLongScalar(col 5:bigint, val 1) -> 3:boolean, col 0:int) -> 4:int native: true nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true valueColumnNums: [0, 1] @@ -1373,12 +1373,12 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY.reducesinkkey0:int, KEY.reducesinkkey1:int, VALUE._col0:int, VALUE._col1:int + dataColumns: KEY.reducesinkkey0:bigint, KEY.reducesinkkey1:int, VALUE._col0:int, VALUE._col1:int partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: Select Operator - expressions: VALUE._col0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey0 (type: int) + expressions: VALUE._col0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey0 (type: bigint) outputColumnNames: _col0, _col1, _col2 Select Vectorization: className: VectorSelectOperator @@ -1476,18 +1476,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: key (type: int), value (type: int), 0 (type: int) + keys: key (type: int), value (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -1524,7 +1524,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1532,22 +1532,22 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 1, 0) (type: int) + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1, 0) (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3 Select Vectorization: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 1, 2, 3] - selectExpressions: VectorUDFAdaptor(grouping(_col2, 1, 0)) -> 3:int + selectExpressions: VectorUDFAdaptor(grouping(_col2, 1, 0)) -> 3:bigint Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false @@ -1642,18 +1642,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: key (type: int), value (type: int), 0 (type: int) + keys: key (type: int), value (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 24 Data size: 192 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -1690,7 +1690,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1698,22 +1698,22 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 0, 1) (type: int) + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 0, 1) (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3 Select Vectorization: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 1, 2, 3] - selectExpressions: VectorUDFAdaptor(grouping(_col2, 0, 1)) -> 3:int + selectExpressions: VectorUDFAdaptor(grouping(_col2, 0, 1)) -> 3:bigint Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false @@ -1808,18 +1808,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: key (type: int), value (type: int), 0 (type: int) + keys: key (type: int), value (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -1856,7 +1856,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -1864,22 +1864,22 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 1, 0) (type: int) + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1, 0) (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3 Select Vectorization: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 1, 2, 3] - selectExpressions: VectorUDFAdaptor(grouping(_col2, 1, 0)) -> 3:int + selectExpressions: VectorUDFAdaptor(grouping(_col2, 1, 0)) -> 3:bigint Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false @@ -1969,18 +1969,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:int + keyExpressions: col 0:int, col 1:int, ConstantVectorExpression(val 0) -> 3:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: key (type: int), value (type: int), 0 (type: int) + keys: key (type: int), value (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 18 Data size: 144 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -2017,7 +2017,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 3 - dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:int + dataColumns: KEY._col0:int, KEY._col1:int, KEY._col2:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -2025,22 +2025,22 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int, col 2:int + keyExpressions: col 0:int, col 1:int, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), grouping(_col2, 0, 1) (type: int) + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 0, 1) (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3 Select Vectorization: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 1, 2, 3] - selectExpressions: VectorUDFAdaptor(grouping(_col2, 0, 1)) -> 3:int + selectExpressions: VectorUDFAdaptor(grouping(_col2, 0, 1)) -> 3:bigint Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false
http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_limit.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_limit.q.out b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_limit.q.out index e8ca06e..dee20d1 100644 --- a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_limit.q.out +++ b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_sets_limit.q.out @@ -74,18 +74,18 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int + keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: a (type: string), b (type: string), 0 (type: int) + keys: a (type: string), b (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 24 Data size: 8832 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -124,7 +124,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -134,11 +134,11 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:string, col 1:string, col 2:int + keyExpressions: col 0:string, col 1:string, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [0] - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE @@ -275,18 +275,18 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int + keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: a (type: string), b (type: string), 0 (type: int) + keys: a (type: string), b (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 24 Data size: 8832 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -325,7 +325,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -335,11 +335,11 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:string, col 1:string, col 2:int + keyExpressions: col 0:string, col 1:string, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [0] - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE @@ -476,18 +476,18 @@ STAGE PLANS: aggregators: VectorUDAFCountStar(*) -> bigint className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:int + keyExpressions: col 0:string, col 1:string, ConstantVectorExpression(val 0) -> 4:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0] - keys: a (type: string), b (type: string), 0 (type: int) + keys: a (type: string), b (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 12 Data size: 4416 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2] @@ -526,7 +526,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:int, VALUE._col0:bigint + dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:bigint, VALUE._col0:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -536,11 +536,11 @@ STAGE PLANS: aggregators: VectorUDAFCountMerge(col 3:bigint) -> bigint className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:string, col 1:string, col 2:int + keyExpressions: col 0:string, col 1:string, col 2:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [0] - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE @@ -675,18 +675,18 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:string, col 1:string, col 2:string, ConstantVectorExpression(val 0) -> 4:int + keyExpressions: col 0:string, col 1:string, col 2:string, ConstantVectorExpression(val 0) -> 4:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [] - keys: a (type: string), b (type: string), c (type: string), 0 (type: int) + keys: a (type: string), b (type: string), c (type: string), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 18 Data size: 9936 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: int) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) sort order: ++++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: int) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1, 2, 3] @@ -724,7 +724,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:string, KEY._col3:int + dataColumns: KEY._col0:string, KEY._col1:string, KEY._col2:string, KEY._col3:bigint partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -732,11 +732,11 @@ STAGE PLANS: Group By Vectorization: className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:string, col 1:string, col 2:string, col 3:int + keyExpressions: col 0:string, col 1:string, col 2:string, col 3:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [] - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: int) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 9 Data size: 4968 Basic stats: COMPLETE Column stats: NONE http://git-wip-us.apache.org/repos/asf/hive/blob/ddd4c9ae/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_window.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_window.q.out b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_window.q.out index 4de6ebb..9374bcc 100644 --- a/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_window.q.out +++ b/ql/src/test/results/clientpositive/llap/vector_groupby_grouping_window.q.out @@ -72,25 +72,25 @@ STAGE PLANS: aggregators: VectorUDAFMaxLong(col 1:int) -> int, VectorUDAFMaxLong(col 2:int) -> int className: VectorGroupByOperator groupByMode: HASH - keyExpressions: col 0:int, ConstantVectorExpression(val 0) -> 4:int + keyExpressions: col 0:int, ConstantVectorExpression(val 0) -> 4:bigint native: false vectorProcessingMode: HASH projectedOutputColumnNums: [0, 1] - keys: category (type: int), 0 (type: int) + keys: category (type: int), 0 (type: bigint) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int) + key expressions: _col0 (type: int), _col1 (type: bigint) sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int) + Map-reduce partition columns: _col0 (type: int), _col1 (type: bigint) Reduce Sink Vectorization: className: VectorReduceSinkMultiKeyOperator keyColumnNums: [0, 1] native: true nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true valueColumnNums: [2, 3] - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col2 (type: int), _col3 (type: int) Execution mode: vectorized, llap LLAP IO: all inputs @@ -121,7 +121,7 @@ STAGE PLANS: vectorized: true rowBatchContext: dataColumnCount: 4 - dataColumns: KEY._col0:int, KEY._col1:int, VALUE._col0:int, VALUE._col1:int + dataColumns: KEY._col0:int, KEY._col1:bigint, VALUE._col0:int, VALUE._col1:int partitionColumnCount: 0 scratchColumnTypeNames: [] Reduce Operator Tree: @@ -131,14 +131,14 @@ STAGE PLANS: aggregators: VectorUDAFMaxLong(col 2:int) -> int, VectorUDAFMaxLong(col 3:int) -> int className: VectorGroupByOperator groupByMode: MERGEPARTIAL - keyExpressions: col 0:int, col 1:int + keyExpressions: col 0:int, col 1:bigint native: false vectorProcessingMode: MERGE_PARTIAL projectedOutputColumnNums: [0, 1] - keys: KEY._col0 (type: int), KEY._col1 (type: int) + keys: KEY._col0 (type: int), KEY._col1 (type: bigint) mode: mergepartial outputColumnNames: _col0, _col2, _col3 - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE pruneGroupingSetId: true Filter Operator Filter Vectorization: @@ -146,7 +146,7 @@ STAGE PLANS: native: true predicateExpression: FilterLongColGreaterLongScalar(col 2:int, val 0) predicate: (_col3 > 0) (type: boolean) - Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: int), _col3 (type: int) sort order: ++ @@ -158,7 +158,7 @@ STAGE PLANS: nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true partitionColumnNums: [0] valueColumnNums: [1] - Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col2 (type: int) Reducer 3 Execution mode: vectorized, llap @@ -183,7 +183,7 @@ STAGE PLANS: className: VectorSelectOperator native: true projectedOutputColumnNums: [0, 2, 1] - Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE PTF Operator Function definitions: Input definition @@ -217,7 +217,7 @@ STAGE PLANS: outputTypes: [int, int, int, int] partitionExpressions: [col 0:int] streamingColumns: [3] - Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col0 (type: int), _col2 (type: int), _col3 (type: int), rank_window_0 (type: int) outputColumnNames: _col0, _col1, _col2, _col3