Repository: hive Updated Branches: refs/heads/branch-2 d0df902e4 -> 96c2a2c08
http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/spark/union17.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/union17.q.out b/ql/src/test/results/clientpositive/spark/union17.q.out index 9063e64..bcb95e4 100644 --- a/ql/src/test/results/clientpositive/spark/union17.q.out +++ b/ql/src/test/results/clientpositive/spark/union17.q.out @@ -109,14 +109,18 @@ STAGE PLANS: mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1 + Select Operator + expressions: _col0 (type: string), UDFToString(_col1) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1 Reducer 4 Reduce Operator Tree: Group By Operator @@ -125,14 +129,18 @@ STAGE PLANS: mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), UDFToString(_col2) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: PARTIAL + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2 Reducer 8 Reduce Operator Tree: Group By Operator http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/spark/union19.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/union19.q.out b/ql/src/test/results/clientpositive/spark/union19.q.out index f7168fe..31795b2 100644 --- a/ql/src/test/results/clientpositive/spark/union19.q.out +++ b/ql/src/test/results/clientpositive/spark/union19.q.out @@ -135,14 +135,18 @@ STAGE PLANS: mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1 + Select Operator + expressions: _col0 (type: string), UDFToString(_col1) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1 Stage: Stage-0 Move Operator http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/spark/union_remove_19.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/union_remove_19.q.out b/ql/src/test/results/clientpositive/spark/union_remove_19.q.out index bfa7c14..a5b5a40 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_19.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_19.q.out @@ -386,7 +386,7 @@ STAGE PLANS: outputColumnNames: _col0, _col1 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: (UDFToDouble(_col0) + UDFToDouble(_col0)) (type: double), _col1 (type: bigint) + expressions: UDFToString((UDFToDouble(_col0) + UDFToDouble(_col0))) (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE File Output Operator @@ -406,7 +406,7 @@ STAGE PLANS: outputColumnNames: _col0, _col1 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: (UDFToDouble(_col0) + UDFToDouble(_col0)) (type: double), _col1 (type: bigint) + expressions: UDFToString((UDFToDouble(_col0) + UDFToDouble(_col0))) (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE File Output Operator http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out b/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out index 1eca688..20aac92 100644 --- a/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out +++ b/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out @@ -3471,8 +3471,8 @@ POSTHOOK: Lineage: alltypesnull.cdouble EXPRESSION [] POSTHOOK: Lineage: alltypesnull.cfloat EXPRESSION [] POSTHOOK: Lineage: alltypesnull.cint EXPRESSION [] POSTHOOK: Lineage: alltypesnull.csmallint EXPRESSION [] -POSTHOOK: Lineage: alltypesnull.cstring1 SIMPLE [] -POSTHOOK: Lineage: alltypesnull.cstring2 SIMPLE [] +POSTHOOK: Lineage: alltypesnull.cstring1 EXPRESSION [] +POSTHOOK: Lineage: alltypesnull.cstring2 EXPRESSION [] POSTHOOK: Lineage: alltypesnull.ctimestamp1 EXPRESSION [] POSTHOOK: Lineage: alltypesnull.ctimestamp2 EXPRESSION [] POSTHOOK: Lineage: alltypesnull.ctinyint EXPRESSION [] http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/tez/explainanalyze_2.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/explainanalyze_2.q.out b/ql/src/test/results/clientpositive/tez/explainanalyze_2.q.out index 88c68d4..ea572cd 100644 --- a/ql/src/test/results/clientpositive/tez/explainanalyze_2.q.out +++ b/ql/src/test/results/clientpositive/tez/explainanalyze_2.q.out @@ -1778,40 +1778,44 @@ Stage-4 Dependency Collection{} Stage-2 Reducer 5 - File Output Operator [FS_18] + File Output Operator [FS_19] table:{"name:":"default.dest1"} - Group By Operator [GBY_16] (rows=205/310 width=96) - Output:["_col0","_col1"],aggregations:["count(DISTINCT KEY._col1:0._col0)"],keys:KEY._col0 - <-Reducer 4 [SIMPLE_EDGE] - SHUFFLE [RS_15] - PartitionCols:_col0 - Group By Operator [GBY_12] (rows=501/310 width=272) - Output:["_col0","_col1"],keys:KEY._col0, KEY._col1 - <-Union 3 [SIMPLE_EDGE] - <-Map 6 [CONTAINS] - Reduce Output Operator [RS_11] - PartitionCols:_col0, _col1 - Select Operator [SEL_7] (rows=500/500 width=178) - Output:["_col0","_col1"] - TableScan [TS_6] (rows=500/500 width=178) - Output:["key","value"] - <-Reducer 2 [CONTAINS] - Reduce Output Operator [RS_11] - PartitionCols:_col0, _col1 - Select Operator [SEL_5] (rows=1/1 width=272) - Output:["_col0","_col1"] - Group By Operator [GBY_4] (rows=1/1 width=8) - Output:["_col0"],aggregations:["count(1)"] - <-Map 1 [CUSTOM_SIMPLE_EDGE] - PARTITION_ONLY_SHUFFLE [RS_3] - Select Operator [SEL_1] (rows=500/500 width=10) - TableScan [TS_0] (rows=500/500 width=10) - default@src,s1,Tbl:COMPLETE,Col:COMPLETE - File Output Operator [FS_23] - table:{"name:":"default.dest2"} - Group By Operator [GBY_21] (rows=501/310 width=280) - Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT substr(_col1, 5))"],keys:_col0, _col1 - Please refer to the previous Group By Operator [GBY_12] + Select Operator [SEL_17] (rows=205/310 width=272) + Output:["_col0","_col1"] + Group By Operator [GBY_16] (rows=205/310 width=96) + Output:["_col0","_col1"],aggregations:["count(DISTINCT KEY._col1:0._col0)"],keys:KEY._col0 + <-Reducer 4 [SIMPLE_EDGE] + SHUFFLE [RS_15] + PartitionCols:_col0 + Group By Operator [GBY_12] (rows=501/310 width=272) + Output:["_col0","_col1"],keys:KEY._col0, KEY._col1 + <-Union 3 [SIMPLE_EDGE] + <-Map 6 [CONTAINS] + Reduce Output Operator [RS_11] + PartitionCols:_col0, _col1 + Select Operator [SEL_7] (rows=500/500 width=178) + Output:["_col0","_col1"] + TableScan [TS_6] (rows=500/500 width=178) + Output:["key","value"] + <-Reducer 2 [CONTAINS] + Reduce Output Operator [RS_11] + PartitionCols:_col0, _col1 + Select Operator [SEL_5] (rows=1/1 width=272) + Output:["_col0","_col1"] + Group By Operator [GBY_4] (rows=1/1 width=8) + Output:["_col0"],aggregations:["count(1)"] + <-Map 1 [CUSTOM_SIMPLE_EDGE] + PARTITION_ONLY_SHUFFLE [RS_3] + Select Operator [SEL_1] (rows=500/500 width=10) + TableScan [TS_0] (rows=500/500 width=10) + default@src,s1,Tbl:COMPLETE,Col:COMPLETE + File Output Operator [FS_25] + table:{"name:":"default.dest2"} + Select Operator [SEL_23] (rows=501/310 width=456) + Output:["_col0","_col1","_col2"] + Group By Operator [GBY_22] (rows=501/310 width=280) + Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT substr(_col1, 5))"],keys:_col0, _col1 + Please refer to the previous Group By Operator [GBY_12] Stage-5 Stats-Aggr Operator Stage-1 @@ -2006,52 +2010,56 @@ Stage-4 Dependency Collection{} Stage-2 Reducer 4 - File Output Operator [FS_17] + File Output Operator [FS_18] table:{"name:":"default.dest1"} - Group By Operator [GBY_15] (rows=205/310 width=96) - Output:["_col0","_col1"],aggregations:["count(DISTINCT KEY._col1:0._col0)"],keys:KEY._col0 - <-Union 3 [SIMPLE_EDGE] - <-Map 6 [CONTAINS] - Reduce Output Operator [RS_14] - PartitionCols:_col0 - Select Operator [SEL_7] (rows=500/500 width=178) - Output:["_col0","_col1"] - TableScan [TS_6] (rows=500/500 width=178) - Output:["key","value"] - Reduce Output Operator [RS_19] - PartitionCols:_col0, _col1 - Please refer to the previous Select Operator [SEL_7] - <-Map 7 [CONTAINS] - Reduce Output Operator [RS_14] - PartitionCols:_col0 - Select Operator [SEL_11] (rows=500/500 width=178) - Output:["_col0","_col1"] - TableScan [TS_10] (rows=500/500 width=178) - Output:["key","value"] - Reduce Output Operator [RS_19] - PartitionCols:_col0, _col1 - Please refer to the previous Select Operator [SEL_11] - <-Reducer 2 [CONTAINS] - Reduce Output Operator [RS_14] - PartitionCols:_col0 - Select Operator [SEL_5] (rows=1/1 width=272) - Output:["_col0","_col1"] - Group By Operator [GBY_4] (rows=1/1 width=8) - Output:["_col0"],aggregations:["count(1)"] - <-Map 1 [CUSTOM_SIMPLE_EDGE] - PARTITION_ONLY_SHUFFLE [RS_3] - Select Operator [SEL_1] (rows=500/500 width=10) - TableScan [TS_0] (rows=500/500 width=10) - default@src,s1,Tbl:COMPLETE,Col:COMPLETE - Reduce Output Operator [RS_19] - PartitionCols:_col0, _col1 - Please refer to the previous Select Operator [SEL_5] + Select Operator [SEL_16] (rows=205/310 width=272) + Output:["_col0","_col1"] + Group By Operator [GBY_15] (rows=205/310 width=96) + Output:["_col0","_col1"],aggregations:["count(DISTINCT KEY._col1:0._col0)"],keys:KEY._col0 + <-Union 3 [SIMPLE_EDGE] + <-Map 6 [CONTAINS] + Reduce Output Operator [RS_14] + PartitionCols:_col0 + Select Operator [SEL_7] (rows=500/500 width=178) + Output:["_col0","_col1"] + TableScan [TS_6] (rows=500/500 width=178) + Output:["key","value"] + Reduce Output Operator [RS_20] + PartitionCols:_col0, _col1 + Please refer to the previous Select Operator [SEL_7] + <-Map 7 [CONTAINS] + Reduce Output Operator [RS_14] + PartitionCols:_col0 + Select Operator [SEL_11] (rows=500/500 width=178) + Output:["_col0","_col1"] + TableScan [TS_10] (rows=500/500 width=178) + Output:["key","value"] + Reduce Output Operator [RS_20] + PartitionCols:_col0, _col1 + Please refer to the previous Select Operator [SEL_11] + <-Reducer 2 [CONTAINS] + Reduce Output Operator [RS_14] + PartitionCols:_col0 + Select Operator [SEL_5] (rows=1/1 width=272) + Output:["_col0","_col1"] + Group By Operator [GBY_4] (rows=1/1 width=8) + Output:["_col0"],aggregations:["count(1)"] + <-Map 1 [CUSTOM_SIMPLE_EDGE] + PARTITION_ONLY_SHUFFLE [RS_3] + Select Operator [SEL_1] (rows=500/500 width=10) + TableScan [TS_0] (rows=500/500 width=10) + default@src,s1,Tbl:COMPLETE,Col:COMPLETE + Reduce Output Operator [RS_20] + PartitionCols:_col0, _col1 + Please refer to the previous Select Operator [SEL_5] Reducer 5 - File Output Operator [FS_22] + File Output Operator [FS_24] table:{"name:":"default.dest2"} - Group By Operator [GBY_20] (rows=1001/310 width=280) - Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT KEY._col2:0._col0)"],keys:KEY._col0, KEY._col1 - <- Please refer to the previous Union 3 [SIMPLE_EDGE] + Select Operator [SEL_22] (rows=1001/310 width=456) + Output:["_col0","_col1","_col2"] + Group By Operator [GBY_21] (rows=1001/310 width=280) + Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT KEY._col2:0._col0)"],keys:KEY._col0, KEY._col1 + <- Please refer to the previous Union 3 [SIMPLE_EDGE] Stage-5 Stats-Aggr Operator Stage-1 @@ -2120,42 +2128,46 @@ Stage-4 Dependency Collection{} Stage-2 Reducer 4 - File Output Operator [FS_14] + File Output Operator [FS_15] table:{"name:":"default.dest1"} - Group By Operator [GBY_12] (rows=205/310 width=96) - Output:["_col0","_col1"],aggregations:["count(DISTINCT KEY._col1:0._col0)"],keys:KEY._col0 - <-Union 3 [SIMPLE_EDGE] - <-Map 6 [CONTAINS] - Reduce Output Operator [RS_11] - PartitionCols:_col0 - Select Operator [SEL_7] (rows=500/500 width=178) - Output:["_col0","_col1"] - TableScan [TS_6] (rows=500/500 width=178) - Output:["key","value"] - Reduce Output Operator [RS_16] - PartitionCols:_col0, _col1 - Please refer to the previous Select Operator [SEL_7] - <-Reducer 2 [CONTAINS] - Reduce Output Operator [RS_11] - PartitionCols:_col0 - Select Operator [SEL_5] (rows=1/1 width=272) - Output:["_col0","_col1"] - Group By Operator [GBY_4] (rows=1/1 width=8) - Output:["_col0"],aggregations:["count(1)"] - <-Map 1 [CUSTOM_SIMPLE_EDGE] - PARTITION_ONLY_SHUFFLE [RS_3] - Select Operator [SEL_1] (rows=500/500 width=10) - TableScan [TS_0] (rows=500/500 width=10) - default@src,s1,Tbl:COMPLETE,Col:COMPLETE - Reduce Output Operator [RS_16] - PartitionCols:_col0, _col1 - Please refer to the previous Select Operator [SEL_5] + Select Operator [SEL_13] (rows=205/310 width=272) + Output:["_col0","_col1"] + Group By Operator [GBY_12] (rows=205/310 width=96) + Output:["_col0","_col1"],aggregations:["count(DISTINCT KEY._col1:0._col0)"],keys:KEY._col0 + <-Union 3 [SIMPLE_EDGE] + <-Map 6 [CONTAINS] + Reduce Output Operator [RS_11] + PartitionCols:_col0 + Select Operator [SEL_7] (rows=500/500 width=178) + Output:["_col0","_col1"] + TableScan [TS_6] (rows=500/500 width=178) + Output:["key","value"] + Reduce Output Operator [RS_17] + PartitionCols:_col0, _col1 + Please refer to the previous Select Operator [SEL_7] + <-Reducer 2 [CONTAINS] + Reduce Output Operator [RS_11] + PartitionCols:_col0 + Select Operator [SEL_5] (rows=1/1 width=272) + Output:["_col0","_col1"] + Group By Operator [GBY_4] (rows=1/1 width=8) + Output:["_col0"],aggregations:["count(1)"] + <-Map 1 [CUSTOM_SIMPLE_EDGE] + PARTITION_ONLY_SHUFFLE [RS_3] + Select Operator [SEL_1] (rows=500/500 width=10) + TableScan [TS_0] (rows=500/500 width=10) + default@src,s1,Tbl:COMPLETE,Col:COMPLETE + Reduce Output Operator [RS_17] + PartitionCols:_col0, _col1 + Please refer to the previous Select Operator [SEL_5] Reducer 5 - File Output Operator [FS_19] + File Output Operator [FS_21] table:{"name:":"default.dest2"} - Group By Operator [GBY_17] (rows=501/310 width=280) - Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT KEY._col2:0._col0)"],keys:KEY._col0, KEY._col1 - <- Please refer to the previous Union 3 [SIMPLE_EDGE] + Select Operator [SEL_19] (rows=501/310 width=456) + Output:["_col0","_col1","_col2"] + Group By Operator [GBY_18] (rows=501/310 width=280) + Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT KEY._col2:0._col0)"],keys:KEY._col0, KEY._col1 + <- Please refer to the previous Union 3 [SIMPLE_EDGE] Stage-5 Stats-Aggr Operator Stage-1 http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/udf1.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/udf1.q.out b/ql/src/test/results/clientpositive/udf1.q.out index b3b694b..eebd90f 100644 --- a/ql/src/test/results/clientpositive/udf1.q.out +++ b/ql/src/test/results/clientpositive/udf1.q.out @@ -53,7 +53,7 @@ STAGE PLANS: predicate: (UDFToDouble(key) = 86.0) (type: boolean) Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: true (type: boolean), false (type: boolean), true (type: boolean), true (type: boolean), true (type: boolean), false (type: boolean), false (type: boolean), false (type: boolean), true (type: boolean), true (type: boolean), false (type: boolean), true (type: boolean), true (type: boolean), 'acc' (type: string), 'abc' (type: string), 'abb' (type: string), 'hive' (type: string), 'hadoop' (type: string), 'AaAbAcA' (type: string), false (type: boolean) + expressions: 'TRUE' (type: string), 'FALSE' (type: string), 'TRUE' (type: string), 'TRUE' (type: string), 'TRUE' (type: string), 'FALSE' (type: string), 'FALSE' (type: string), 'FALSE' (type: string), 'TRUE' (type: string), 'TRUE' (type: string), 'FALSE' (type: string), 'TRUE' (type: string), 'TRUE' (type: string), 'acc' (type: string), 'abc' (type: string), 'abb' (type: string), 'hive' (type: string), 'hadoop' (type: string), 'AaAbAcA' (type: string), 'FALSE' (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19 Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE File Output Operator @@ -137,26 +137,26 @@ POSTHOOK: query: FROM src INSERT OVERWRITE TABLE dest1 SELECT 'a' LIKE '%a%', 'b POSTHOOK: type: QUERY POSTHOOK: Input: default@src POSTHOOK: Output: default@dest1 -POSTHOOK: Lineage: dest1.c1 SIMPLE [] -POSTHOOK: Lineage: dest1.c10 SIMPLE [] -POSTHOOK: Lineage: dest1.c11 SIMPLE [] -POSTHOOK: Lineage: dest1.c12 SIMPLE [] -POSTHOOK: Lineage: dest1.c13 SIMPLE [] +POSTHOOK: Lineage: dest1.c1 EXPRESSION [] +POSTHOOK: Lineage: dest1.c10 EXPRESSION [] +POSTHOOK: Lineage: dest1.c11 EXPRESSION [] +POSTHOOK: Lineage: dest1.c12 EXPRESSION [] +POSTHOOK: Lineage: dest1.c13 EXPRESSION [] POSTHOOK: Lineage: dest1.c14 SIMPLE [] POSTHOOK: Lineage: dest1.c15 SIMPLE [] POSTHOOK: Lineage: dest1.c16 SIMPLE [] POSTHOOK: Lineage: dest1.c17 SIMPLE [] POSTHOOK: Lineage: dest1.c18 SIMPLE [] POSTHOOK: Lineage: dest1.c19 SIMPLE [] -POSTHOOK: Lineage: dest1.c2 SIMPLE [] -POSTHOOK: Lineage: dest1.c20 SIMPLE [] -POSTHOOK: Lineage: dest1.c3 SIMPLE [] -POSTHOOK: Lineage: dest1.c4 SIMPLE [] -POSTHOOK: Lineage: dest1.c5 SIMPLE [] -POSTHOOK: Lineage: dest1.c6 SIMPLE [] -POSTHOOK: Lineage: dest1.c7 SIMPLE [] -POSTHOOK: Lineage: dest1.c8 SIMPLE [] -POSTHOOK: Lineage: dest1.c9 SIMPLE [] +POSTHOOK: Lineage: dest1.c2 EXPRESSION [] +POSTHOOK: Lineage: dest1.c20 EXPRESSION [] +POSTHOOK: Lineage: dest1.c3 EXPRESSION [] +POSTHOOK: Lineage: dest1.c4 EXPRESSION [] +POSTHOOK: Lineage: dest1.c5 EXPRESSION [] +POSTHOOK: Lineage: dest1.c6 EXPRESSION [] +POSTHOOK: Lineage: dest1.c7 EXPRESSION [] +POSTHOOK: Lineage: dest1.c8 EXPRESSION [] +POSTHOOK: Lineage: dest1.c9 EXPRESSION [] PREHOOK: query: SELECT dest1.* FROM dest1 PREHOOK: type: QUERY PREHOOK: Input: default@dest1 @@ -165,4 +165,4 @@ POSTHOOK: query: SELECT dest1.* FROM dest1 POSTHOOK: type: QUERY POSTHOOK: Input: default@dest1 #### A masked pattern was here #### -true false true true true false false false true true false true true acc abc abb hive hadoop AaAbAcA false +TRUE FALSE TRUE TRUE TRUE FALSE FALSE FALSE TRUE TRUE FALSE TRUE TRUE acc abc abb hive hadoop AaAbAcA FALSE http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/udf3.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/udf3.q.out b/ql/src/test/results/clientpositive/udf3.q.out index 528b8c7..96038f1 100644 --- a/ql/src/test/results/clientpositive/udf3.q.out +++ b/ql/src/test/results/clientpositive/udf3.q.out @@ -43,14 +43,18 @@ STAGE PLANS: mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3, _col4 Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1 + Select Operator + expressions: UDFToString(_col0) (type: string), UDFToString(_col1) (type: string), UDFToString(_col2) (type: string), UDFToString(_col3) (type: string), UDFToString(_col4) (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1 Stage: Stage-0 Move Operator http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/udtf_stack.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/udtf_stack.q.out b/ql/src/test/results/clientpositive/udtf_stack.q.out index 97dcd3a..d1f880f 100644 --- a/ql/src/test/results/clientpositive/udtf_stack.q.out +++ b/ql/src/test/results/clientpositive/udtf_stack.q.out @@ -182,13 +182,17 @@ STAGE PLANS: UDTF Operator Statistics: Num rows: 1 Data size: 185 Basic stats: COMPLETE Column stats: COMPLETE function name: stack - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 185 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Select Operator + expressions: col0 (type: string), col1 (type: string), UDFToString(col2) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/union17.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/union17.q.out b/ql/src/test/results/clientpositive/union17.q.out index 5c2e6ac..18e06e1 100644 --- a/ql/src/test/results/clientpositive/union17.q.out +++ b/ql/src/test/results/clientpositive/union17.q.out @@ -140,14 +140,18 @@ STAGE PLANS: mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1 + Select Operator + expressions: _col0 (type: string), UDFToString(_col1) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1 Stage: Stage-0 Move Operator @@ -178,14 +182,18 @@ STAGE PLANS: mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), UDFToString(_col2) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: PARTIAL + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2 Stage: Stage-1 Move Operator http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/union19.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/union19.q.out b/ql/src/test/results/clientpositive/union19.q.out index 5781ae5..5ce5905 100644 --- a/ql/src/test/results/clientpositive/union19.q.out +++ b/ql/src/test/results/clientpositive/union19.q.out @@ -141,14 +141,18 @@ STAGE PLANS: mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1 + Select Operator + expressions: _col0 (type: string), UDFToString(_col1) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1 Stage: Stage-0 Move Operator http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/union_remove_19.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/union_remove_19.q.out b/ql/src/test/results/clientpositive/union_remove_19.q.out index 5918afd..8a3a73f 100644 --- a/ql/src/test/results/clientpositive/union_remove_19.q.out +++ b/ql/src/test/results/clientpositive/union_remove_19.q.out @@ -415,7 +415,7 @@ STAGE PLANS: outputColumnNames: _col0, _col1 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: (UDFToDouble(_col0) + UDFToDouble(_col0)) (type: double), _col1 (type: bigint) + expressions: UDFToString((UDFToDouble(_col0) + UDFToDouble(_col0))) (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE File Output Operator @@ -470,7 +470,7 @@ STAGE PLANS: outputColumnNames: _col0, _col1 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: (UDFToDouble(_col0) + UDFToDouble(_col0)) (type: double), _col1 (type: bigint) + expressions: UDFToString((UDFToDouble(_col0) + UDFToDouble(_col0))) (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE File Output Operator http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/varchar_1.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/varchar_1.q.out b/ql/src/test/results/clientpositive/varchar_1.q.out index 57d8ee6..6b35013 100644 --- a/ql/src/test/results/clientpositive/varchar_1.q.out +++ b/ql/src/test/results/clientpositive/varchar_1.q.out @@ -94,8 +94,8 @@ POSTHOOK: query: insert overwrite table varchar1_1 POSTHOOK: type: QUERY POSTHOOK: Input: default@varchar1 POSTHOOK: Output: default@varchar1_1 -POSTHOOK: Lineage: varchar1_1.key SIMPLE [(varchar1)varchar1.FieldSchema(name:key, type:varchar(10), comment:null), ] -POSTHOOK: Lineage: varchar1_1.value SIMPLE [(varchar1)varchar1.FieldSchema(name:value, type:varchar(20), comment:null), ] +POSTHOOK: Lineage: varchar1_1.key EXPRESSION [(varchar1)varchar1.FieldSchema(name:key, type:varchar(10), comment:null), ] +POSTHOOK: Lineage: varchar1_1.value EXPRESSION [(varchar1)varchar1.FieldSchema(name:value, type:varchar(20), comment:null), ] PREHOOK: query: select key, value from varchar1_1 order by key, value PREHOOK: type: QUERY PREHOOK: Input: default@varchar1_1 http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/vector_null_projection.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/vector_null_projection.q.out b/ql/src/test/results/clientpositive/vector_null_projection.q.out index 94aea2f..e9f7733 100644 --- a/ql/src/test/results/clientpositive/vector_null_projection.q.out +++ b/ql/src/test/results/clientpositive/vector_null_projection.q.out @@ -49,23 +49,37 @@ STAGE PLANS: TableScan alias: a Statistics: Num rows: 1 Data size: 87 Basic stats: COMPLETE Column stats: COMPLETE + TableScan Vectorization: + native: true + projectedOutputColumns: [0] Select Operator - expressions: null (type: void) + expressions: null (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Select Vectorization: + className: VectorSelectOperator + native: true + projectedOutputColumns: [1] + selectExpressions: ConstantVectorExpression(val null) -> 1:string + Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + File Sink Vectorization: + className: VectorFileSinkOperator + native: false + Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized Map Vectorization: enabled: true enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true + groupByVectorOutput: true inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - notVectorizedReason: Select expression for SELECT operator: Data type void of Const void null not supported - vectorized: false + allNative: false + usesVectorUDFAdaptor: false + vectorized: true Stage: Stage-0 Fetch Operator @@ -164,12 +178,12 @@ STAGE PLANS: outputColumnNames: _col0 Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: null (type: void) + expressions: null (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/vector_nvl.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/vector_nvl.q.out b/ql/src/test/results/clientpositive/vector_nvl.q.out index 08cc168..4a14288 100644 --- a/ql/src/test/results/clientpositive/vector_nvl.q.out +++ b/ql/src/test/results/clientpositive/vector_nvl.q.out @@ -306,13 +306,17 @@ STAGE PLANS: Limit Number of rows: 10 Statistics: Num rows: 10 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Select Operator + expressions: UDFToString(_col0) (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 10 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Map Vectorization: enabled: true enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true