Repository: hive
Updated Branches:
  refs/heads/branch-2 d0df902e4 -> 96c2a2c08


http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/spark/union17.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/union17.q.out 
b/ql/src/test/results/clientpositive/spark/union17.q.out
index 9063e64..bcb95e4 100644
--- a/ql/src/test/results/clientpositive/spark/union17.q.out
+++ b/ql/src/test/results/clientpositive/spark/union17.q.out
@@ -109,14 +109,18 @@ STAGE PLANS:
                 mode: mergepartial
                 outputColumnNames: _col0, _col1
                 Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE 
Column stats: PARTIAL
-                File Output Operator
-                  compressed: false
-                  Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE 
Column stats: PARTIAL
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      name: default.dest1
+                Select Operator
+                  expressions: _col0 (type: string), UDFToString(_col1) (type: 
string)
+                  outputColumnNames: _col0, _col1
+                  Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE 
Column stats: PARTIAL
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 1 Data size: 272 Basic stats: 
COMPLETE Column stats: PARTIAL
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.dest1
         Reducer 4 
             Reduce Operator Tree:
               Group By Operator
@@ -125,14 +129,18 @@ STAGE PLANS:
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2
                 Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE 
Column stats: PARTIAL
-                File Output Operator
-                  compressed: false
-                  Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE 
Column stats: PARTIAL
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      name: default.dest2
+                Select Operator
+                  expressions: _col0 (type: string), _col1 (type: string), 
UDFToString(_col2) (type: string)
+                  outputColumnNames: _col0, _col1, _col2
+                  Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE 
Column stats: PARTIAL
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 1 Data size: 456 Basic stats: 
COMPLETE Column stats: PARTIAL
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.dest2
         Reducer 8 
             Reduce Operator Tree:
               Group By Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/spark/union19.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/union19.q.out 
b/ql/src/test/results/clientpositive/spark/union19.q.out
index f7168fe..31795b2 100644
--- a/ql/src/test/results/clientpositive/spark/union19.q.out
+++ b/ql/src/test/results/clientpositive/spark/union19.q.out
@@ -135,14 +135,18 @@ STAGE PLANS:
                 mode: mergepartial
                 outputColumnNames: _col0, _col1
                 Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE 
Column stats: PARTIAL
-                File Output Operator
-                  compressed: false
-                  Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE 
Column stats: PARTIAL
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      name: default.dest1
+                Select Operator
+                  expressions: _col0 (type: string), UDFToString(_col1) (type: 
string)
+                  outputColumnNames: _col0, _col1
+                  Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE 
Column stats: PARTIAL
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 1 Data size: 272 Basic stats: 
COMPLETE Column stats: PARTIAL
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                        name: default.dest1
 
   Stage: Stage-0
     Move Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/spark/union_remove_19.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/union_remove_19.q.out 
b/ql/src/test/results/clientpositive/spark/union_remove_19.q.out
index bfa7c14..a5b5a40 100644
--- a/ql/src/test/results/clientpositive/spark/union_remove_19.q.out
+++ b/ql/src/test/results/clientpositive/spark/union_remove_19.q.out
@@ -386,7 +386,7 @@ STAGE PLANS:
                 outputColumnNames: _col0, _col1
                 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
-                  expressions: (UDFToDouble(_col0) + UDFToDouble(_col0)) 
(type: double), _col1 (type: bigint)
+                  expressions: UDFToString((UDFToDouble(_col0) + 
UDFToDouble(_col0))) (type: string), _col1 (type: bigint)
                   outputColumnNames: _col0, _col1
                   Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator
@@ -406,7 +406,7 @@ STAGE PLANS:
                 outputColumnNames: _col0, _col1
                 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
-                  expressions: (UDFToDouble(_col0) + UDFToDouble(_col0)) 
(type: double), _col1 (type: bigint)
+                  expressions: UDFToString((UDFToDouble(_col0) + 
UDFToDouble(_col0))) (type: string), _col1 (type: bigint)
                   outputColumnNames: _col0, _col1
                   Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out 
b/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out
index 1eca688..20aac92 100644
--- a/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out
+++ b/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out
@@ -3471,8 +3471,8 @@ POSTHOOK: Lineage: alltypesnull.cdouble EXPRESSION []
 POSTHOOK: Lineage: alltypesnull.cfloat EXPRESSION []
 POSTHOOK: Lineage: alltypesnull.cint EXPRESSION []
 POSTHOOK: Lineage: alltypesnull.csmallint EXPRESSION []
-POSTHOOK: Lineage: alltypesnull.cstring1 SIMPLE []
-POSTHOOK: Lineage: alltypesnull.cstring2 SIMPLE []
+POSTHOOK: Lineage: alltypesnull.cstring1 EXPRESSION []
+POSTHOOK: Lineage: alltypesnull.cstring2 EXPRESSION []
 POSTHOOK: Lineage: alltypesnull.ctimestamp1 EXPRESSION []
 POSTHOOK: Lineage: alltypesnull.ctimestamp2 EXPRESSION []
 POSTHOOK: Lineage: alltypesnull.ctinyint EXPRESSION []

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/tez/explainanalyze_2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/explainanalyze_2.q.out 
b/ql/src/test/results/clientpositive/tez/explainanalyze_2.q.out
index 88c68d4..ea572cd 100644
--- a/ql/src/test/results/clientpositive/tez/explainanalyze_2.q.out
+++ b/ql/src/test/results/clientpositive/tez/explainanalyze_2.q.out
@@ -1778,40 +1778,44 @@ Stage-4
           Dependency Collection{}
             Stage-2
               Reducer 5
-              File Output Operator [FS_18]
+              File Output Operator [FS_19]
                 table:{"name:":"default.dest1"}
-                Group By Operator [GBY_16] (rows=205/310 width=96)
-                  Output:["_col0","_col1"],aggregations:["count(DISTINCT 
KEY._col1:0._col0)"],keys:KEY._col0
-                <-Reducer 4 [SIMPLE_EDGE]
-                  SHUFFLE [RS_15]
-                    PartitionCols:_col0
-                    Group By Operator [GBY_12] (rows=501/310 width=272)
-                      Output:["_col0","_col1"],keys:KEY._col0, KEY._col1
-                    <-Union 3 [SIMPLE_EDGE]
-                      <-Map 6 [CONTAINS]
-                        Reduce Output Operator [RS_11]
-                          PartitionCols:_col0, _col1
-                          Select Operator [SEL_7] (rows=500/500 width=178)
-                            Output:["_col0","_col1"]
-                            TableScan [TS_6] (rows=500/500 width=178)
-                              Output:["key","value"]
-                      <-Reducer 2 [CONTAINS]
-                        Reduce Output Operator [RS_11]
-                          PartitionCols:_col0, _col1
-                          Select Operator [SEL_5] (rows=1/1 width=272)
-                            Output:["_col0","_col1"]
-                            Group By Operator [GBY_4] (rows=1/1 width=8)
-                              Output:["_col0"],aggregations:["count(1)"]
-                            <-Map 1 [CUSTOM_SIMPLE_EDGE]
-                              PARTITION_ONLY_SHUFFLE [RS_3]
-                                Select Operator [SEL_1] (rows=500/500 width=10)
-                                  TableScan [TS_0] (rows=500/500 width=10)
-                                    default@src,s1,Tbl:COMPLETE,Col:COMPLETE
-                  File Output Operator [FS_23]
-                    table:{"name:":"default.dest2"}
-                    Group By Operator [GBY_21] (rows=501/310 width=280)
-                      
Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT substr(_col1, 
5))"],keys:_col0, _col1
-                       Please refer to the previous Group By Operator [GBY_12]
+                Select Operator [SEL_17] (rows=205/310 width=272)
+                  Output:["_col0","_col1"]
+                  Group By Operator [GBY_16] (rows=205/310 width=96)
+                    Output:["_col0","_col1"],aggregations:["count(DISTINCT 
KEY._col1:0._col0)"],keys:KEY._col0
+                  <-Reducer 4 [SIMPLE_EDGE]
+                    SHUFFLE [RS_15]
+                      PartitionCols:_col0
+                      Group By Operator [GBY_12] (rows=501/310 width=272)
+                        Output:["_col0","_col1"],keys:KEY._col0, KEY._col1
+                      <-Union 3 [SIMPLE_EDGE]
+                        <-Map 6 [CONTAINS]
+                          Reduce Output Operator [RS_11]
+                            PartitionCols:_col0, _col1
+                            Select Operator [SEL_7] (rows=500/500 width=178)
+                              Output:["_col0","_col1"]
+                              TableScan [TS_6] (rows=500/500 width=178)
+                                Output:["key","value"]
+                        <-Reducer 2 [CONTAINS]
+                          Reduce Output Operator [RS_11]
+                            PartitionCols:_col0, _col1
+                            Select Operator [SEL_5] (rows=1/1 width=272)
+                              Output:["_col0","_col1"]
+                              Group By Operator [GBY_4] (rows=1/1 width=8)
+                                Output:["_col0"],aggregations:["count(1)"]
+                              <-Map 1 [CUSTOM_SIMPLE_EDGE]
+                                PARTITION_ONLY_SHUFFLE [RS_3]
+                                  Select Operator [SEL_1] (rows=500/500 
width=10)
+                                    TableScan [TS_0] (rows=500/500 width=10)
+                                      default@src,s1,Tbl:COMPLETE,Col:COMPLETE
+                    File Output Operator [FS_25]
+                      table:{"name:":"default.dest2"}
+                      Select Operator [SEL_23] (rows=501/310 width=456)
+                        Output:["_col0","_col1","_col2"]
+                        Group By Operator [GBY_22] (rows=501/310 width=280)
+                          
Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT substr(_col1, 
5))"],keys:_col0, _col1
+                           Please refer to the previous Group By Operator 
[GBY_12]
 Stage-5
   Stats-Aggr Operator
     Stage-1
@@ -2006,52 +2010,56 @@ Stage-4
           Dependency Collection{}
             Stage-2
               Reducer 4
-              File Output Operator [FS_17]
+              File Output Operator [FS_18]
                 table:{"name:":"default.dest1"}
-                Group By Operator [GBY_15] (rows=205/310 width=96)
-                  Output:["_col0","_col1"],aggregations:["count(DISTINCT 
KEY._col1:0._col0)"],keys:KEY._col0
-                <-Union 3 [SIMPLE_EDGE]
-                  <-Map 6 [CONTAINS]
-                    Reduce Output Operator [RS_14]
-                      PartitionCols:_col0
-                      Select Operator [SEL_7] (rows=500/500 width=178)
-                        Output:["_col0","_col1"]
-                        TableScan [TS_6] (rows=500/500 width=178)
-                          Output:["key","value"]
-                    Reduce Output Operator [RS_19]
-                      PartitionCols:_col0, _col1
-                       Please refer to the previous Select Operator [SEL_7]
-                  <-Map 7 [CONTAINS]
-                    Reduce Output Operator [RS_14]
-                      PartitionCols:_col0
-                      Select Operator [SEL_11] (rows=500/500 width=178)
-                        Output:["_col0","_col1"]
-                        TableScan [TS_10] (rows=500/500 width=178)
-                          Output:["key","value"]
-                    Reduce Output Operator [RS_19]
-                      PartitionCols:_col0, _col1
-                       Please refer to the previous Select Operator [SEL_11]
-                  <-Reducer 2 [CONTAINS]
-                    Reduce Output Operator [RS_14]
-                      PartitionCols:_col0
-                      Select Operator [SEL_5] (rows=1/1 width=272)
-                        Output:["_col0","_col1"]
-                        Group By Operator [GBY_4] (rows=1/1 width=8)
-                          Output:["_col0"],aggregations:["count(1)"]
-                        <-Map 1 [CUSTOM_SIMPLE_EDGE]
-                          PARTITION_ONLY_SHUFFLE [RS_3]
-                            Select Operator [SEL_1] (rows=500/500 width=10)
-                              TableScan [TS_0] (rows=500/500 width=10)
-                                default@src,s1,Tbl:COMPLETE,Col:COMPLETE
-                    Reduce Output Operator [RS_19]
-                      PartitionCols:_col0, _col1
-                       Please refer to the previous Select Operator [SEL_5]
+                Select Operator [SEL_16] (rows=205/310 width=272)
+                  Output:["_col0","_col1"]
+                  Group By Operator [GBY_15] (rows=205/310 width=96)
+                    Output:["_col0","_col1"],aggregations:["count(DISTINCT 
KEY._col1:0._col0)"],keys:KEY._col0
+                  <-Union 3 [SIMPLE_EDGE]
+                    <-Map 6 [CONTAINS]
+                      Reduce Output Operator [RS_14]
+                        PartitionCols:_col0
+                        Select Operator [SEL_7] (rows=500/500 width=178)
+                          Output:["_col0","_col1"]
+                          TableScan [TS_6] (rows=500/500 width=178)
+                            Output:["key","value"]
+                      Reduce Output Operator [RS_20]
+                        PartitionCols:_col0, _col1
+                         Please refer to the previous Select Operator [SEL_7]
+                    <-Map 7 [CONTAINS]
+                      Reduce Output Operator [RS_14]
+                        PartitionCols:_col0
+                        Select Operator [SEL_11] (rows=500/500 width=178)
+                          Output:["_col0","_col1"]
+                          TableScan [TS_10] (rows=500/500 width=178)
+                            Output:["key","value"]
+                      Reduce Output Operator [RS_20]
+                        PartitionCols:_col0, _col1
+                         Please refer to the previous Select Operator [SEL_11]
+                    <-Reducer 2 [CONTAINS]
+                      Reduce Output Operator [RS_14]
+                        PartitionCols:_col0
+                        Select Operator [SEL_5] (rows=1/1 width=272)
+                          Output:["_col0","_col1"]
+                          Group By Operator [GBY_4] (rows=1/1 width=8)
+                            Output:["_col0"],aggregations:["count(1)"]
+                          <-Map 1 [CUSTOM_SIMPLE_EDGE]
+                            PARTITION_ONLY_SHUFFLE [RS_3]
+                              Select Operator [SEL_1] (rows=500/500 width=10)
+                                TableScan [TS_0] (rows=500/500 width=10)
+                                  default@src,s1,Tbl:COMPLETE,Col:COMPLETE
+                      Reduce Output Operator [RS_20]
+                        PartitionCols:_col0, _col1
+                         Please refer to the previous Select Operator [SEL_5]
               Reducer 5
-              File Output Operator [FS_22]
+              File Output Operator [FS_24]
                 table:{"name:":"default.dest2"}
-                Group By Operator [GBY_20] (rows=1001/310 width=280)
-                  
Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT 
KEY._col2:0._col0)"],keys:KEY._col0, KEY._col1
-                <- Please refer to the previous Union 3 [SIMPLE_EDGE]
+                Select Operator [SEL_22] (rows=1001/310 width=456)
+                  Output:["_col0","_col1","_col2"]
+                  Group By Operator [GBY_21] (rows=1001/310 width=280)
+                    
Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT 
KEY._col2:0._col0)"],keys:KEY._col0, KEY._col1
+                  <- Please refer to the previous Union 3 [SIMPLE_EDGE]
 Stage-5
   Stats-Aggr Operator
     Stage-1
@@ -2120,42 +2128,46 @@ Stage-4
           Dependency Collection{}
             Stage-2
               Reducer 4
-              File Output Operator [FS_14]
+              File Output Operator [FS_15]
                 table:{"name:":"default.dest1"}
-                Group By Operator [GBY_12] (rows=205/310 width=96)
-                  Output:["_col0","_col1"],aggregations:["count(DISTINCT 
KEY._col1:0._col0)"],keys:KEY._col0
-                <-Union 3 [SIMPLE_EDGE]
-                  <-Map 6 [CONTAINS]
-                    Reduce Output Operator [RS_11]
-                      PartitionCols:_col0
-                      Select Operator [SEL_7] (rows=500/500 width=178)
-                        Output:["_col0","_col1"]
-                        TableScan [TS_6] (rows=500/500 width=178)
-                          Output:["key","value"]
-                    Reduce Output Operator [RS_16]
-                      PartitionCols:_col0, _col1
-                       Please refer to the previous Select Operator [SEL_7]
-                  <-Reducer 2 [CONTAINS]
-                    Reduce Output Operator [RS_11]
-                      PartitionCols:_col0
-                      Select Operator [SEL_5] (rows=1/1 width=272)
-                        Output:["_col0","_col1"]
-                        Group By Operator [GBY_4] (rows=1/1 width=8)
-                          Output:["_col0"],aggregations:["count(1)"]
-                        <-Map 1 [CUSTOM_SIMPLE_EDGE]
-                          PARTITION_ONLY_SHUFFLE [RS_3]
-                            Select Operator [SEL_1] (rows=500/500 width=10)
-                              TableScan [TS_0] (rows=500/500 width=10)
-                                default@src,s1,Tbl:COMPLETE,Col:COMPLETE
-                    Reduce Output Operator [RS_16]
-                      PartitionCols:_col0, _col1
-                       Please refer to the previous Select Operator [SEL_5]
+                Select Operator [SEL_13] (rows=205/310 width=272)
+                  Output:["_col0","_col1"]
+                  Group By Operator [GBY_12] (rows=205/310 width=96)
+                    Output:["_col0","_col1"],aggregations:["count(DISTINCT 
KEY._col1:0._col0)"],keys:KEY._col0
+                  <-Union 3 [SIMPLE_EDGE]
+                    <-Map 6 [CONTAINS]
+                      Reduce Output Operator [RS_11]
+                        PartitionCols:_col0
+                        Select Operator [SEL_7] (rows=500/500 width=178)
+                          Output:["_col0","_col1"]
+                          TableScan [TS_6] (rows=500/500 width=178)
+                            Output:["key","value"]
+                      Reduce Output Operator [RS_17]
+                        PartitionCols:_col0, _col1
+                         Please refer to the previous Select Operator [SEL_7]
+                    <-Reducer 2 [CONTAINS]
+                      Reduce Output Operator [RS_11]
+                        PartitionCols:_col0
+                        Select Operator [SEL_5] (rows=1/1 width=272)
+                          Output:["_col0","_col1"]
+                          Group By Operator [GBY_4] (rows=1/1 width=8)
+                            Output:["_col0"],aggregations:["count(1)"]
+                          <-Map 1 [CUSTOM_SIMPLE_EDGE]
+                            PARTITION_ONLY_SHUFFLE [RS_3]
+                              Select Operator [SEL_1] (rows=500/500 width=10)
+                                TableScan [TS_0] (rows=500/500 width=10)
+                                  default@src,s1,Tbl:COMPLETE,Col:COMPLETE
+                      Reduce Output Operator [RS_17]
+                        PartitionCols:_col0, _col1
+                         Please refer to the previous Select Operator [SEL_5]
               Reducer 5
-              File Output Operator [FS_19]
+              File Output Operator [FS_21]
                 table:{"name:":"default.dest2"}
-                Group By Operator [GBY_17] (rows=501/310 width=280)
-                  
Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT 
KEY._col2:0._col0)"],keys:KEY._col0, KEY._col1
-                <- Please refer to the previous Union 3 [SIMPLE_EDGE]
+                Select Operator [SEL_19] (rows=501/310 width=456)
+                  Output:["_col0","_col1","_col2"]
+                  Group By Operator [GBY_18] (rows=501/310 width=280)
+                    
Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT 
KEY._col2:0._col0)"],keys:KEY._col0, KEY._col1
+                  <- Please refer to the previous Union 3 [SIMPLE_EDGE]
 Stage-5
   Stats-Aggr Operator
     Stage-1

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/udf1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/udf1.q.out 
b/ql/src/test/results/clientpositive/udf1.q.out
index b3b694b..eebd90f 100644
--- a/ql/src/test/results/clientpositive/udf1.q.out
+++ b/ql/src/test/results/clientpositive/udf1.q.out
@@ -53,7 +53,7 @@ STAGE PLANS:
               predicate: (UDFToDouble(key) = 86.0) (type: boolean)
               Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE 
Column stats: NONE
               Select Operator
-                expressions: true (type: boolean), false (type: boolean), true 
(type: boolean), true (type: boolean), true (type: boolean), false (type: 
boolean), false (type: boolean), false (type: boolean), true (type: boolean), 
true (type: boolean), false (type: boolean), true (type: boolean), true (type: 
boolean), 'acc' (type: string), 'abc' (type: string), 'abb' (type: string), 
'hive' (type: string), 'hadoop' (type: string), 'AaAbAcA' (type: string), false 
(type: boolean)
+                expressions: 'TRUE' (type: string), 'FALSE' (type: string), 
'TRUE' (type: string), 'TRUE' (type: string), 'TRUE' (type: string), 'FALSE' 
(type: string), 'FALSE' (type: string), 'FALSE' (type: string), 'TRUE' (type: 
string), 'TRUE' (type: string), 'FALSE' (type: string), 'TRUE' (type: string), 
'TRUE' (type: string), 'acc' (type: string), 'abc' (type: string), 'abb' (type: 
string), 'hive' (type: string), 'hadoop' (type: string), 'AaAbAcA' (type: 
string), 'FALSE' (type: string)
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, 
_col16, _col17, _col18, _col19
                 Statistics: Num rows: 250 Data size: 2656 Basic stats: 
COMPLETE Column stats: NONE
                 File Output Operator
@@ -137,26 +137,26 @@ POSTHOOK: query: FROM src INSERT OVERWRITE TABLE dest1 
SELECT 'a' LIKE '%a%', 'b
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 POSTHOOK: Output: default@dest1
-POSTHOOK: Lineage: dest1.c1 SIMPLE []
-POSTHOOK: Lineage: dest1.c10 SIMPLE []
-POSTHOOK: Lineage: dest1.c11 SIMPLE []
-POSTHOOK: Lineage: dest1.c12 SIMPLE []
-POSTHOOK: Lineage: dest1.c13 SIMPLE []
+POSTHOOK: Lineage: dest1.c1 EXPRESSION []
+POSTHOOK: Lineage: dest1.c10 EXPRESSION []
+POSTHOOK: Lineage: dest1.c11 EXPRESSION []
+POSTHOOK: Lineage: dest1.c12 EXPRESSION []
+POSTHOOK: Lineage: dest1.c13 EXPRESSION []
 POSTHOOK: Lineage: dest1.c14 SIMPLE []
 POSTHOOK: Lineage: dest1.c15 SIMPLE []
 POSTHOOK: Lineage: dest1.c16 SIMPLE []
 POSTHOOK: Lineage: dest1.c17 SIMPLE []
 POSTHOOK: Lineage: dest1.c18 SIMPLE []
 POSTHOOK: Lineage: dest1.c19 SIMPLE []
-POSTHOOK: Lineage: dest1.c2 SIMPLE []
-POSTHOOK: Lineage: dest1.c20 SIMPLE []
-POSTHOOK: Lineage: dest1.c3 SIMPLE []
-POSTHOOK: Lineage: dest1.c4 SIMPLE []
-POSTHOOK: Lineage: dest1.c5 SIMPLE []
-POSTHOOK: Lineage: dest1.c6 SIMPLE []
-POSTHOOK: Lineage: dest1.c7 SIMPLE []
-POSTHOOK: Lineage: dest1.c8 SIMPLE []
-POSTHOOK: Lineage: dest1.c9 SIMPLE []
+POSTHOOK: Lineage: dest1.c2 EXPRESSION []
+POSTHOOK: Lineage: dest1.c20 EXPRESSION []
+POSTHOOK: Lineage: dest1.c3 EXPRESSION []
+POSTHOOK: Lineage: dest1.c4 EXPRESSION []
+POSTHOOK: Lineage: dest1.c5 EXPRESSION []
+POSTHOOK: Lineage: dest1.c6 EXPRESSION []
+POSTHOOK: Lineage: dest1.c7 EXPRESSION []
+POSTHOOK: Lineage: dest1.c8 EXPRESSION []
+POSTHOOK: Lineage: dest1.c9 EXPRESSION []
 PREHOOK: query: SELECT dest1.* FROM dest1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@dest1
@@ -165,4 +165,4 @@ POSTHOOK: query: SELECT dest1.* FROM dest1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
 #### A masked pattern was here ####
-true   false   true    true    true    false   false   false   true    true    
false   true    true    acc     abc     abb     hive    hadoop  AaAbAcA false
+TRUE   FALSE   TRUE    TRUE    TRUE    FALSE   FALSE   FALSE   TRUE    TRUE    
FALSE   TRUE    TRUE    acc     abc     abb     hive    hadoop  AaAbAcA FALSE

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/udf3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/udf3.q.out 
b/ql/src/test/results/clientpositive/udf3.q.out
index 528b8c7..96038f1 100644
--- a/ql/src/test/results/clientpositive/udf3.q.out
+++ b/ql/src/test/results/clientpositive/udf3.q.out
@@ -43,14 +43,18 @@ STAGE PLANS:
           mode: mergepartial
           outputColumnNames: _col0, _col1, _col2, _col3, _col4
           Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column 
stats: COMPLETE
-          File Output Operator
-            compressed: false
-            Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column 
stats: COMPLETE
-            table:
-                input format: org.apache.hadoop.mapred.TextInputFormat
-                output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                name: default.dest1
+          Select Operator
+            expressions: UDFToString(_col0) (type: string), UDFToString(_col1) 
(type: string), UDFToString(_col2) (type: string), UDFToString(_col3) (type: 
string), UDFToString(_col4) (type: string)
+            outputColumnNames: _col0, _col1, _col2, _col3, _col4
+            Statistics: Num rows: 1 Data size: 920 Basic stats: COMPLETE 
Column stats: COMPLETE
+            File Output Operator
+              compressed: false
+              Statistics: Num rows: 1 Data size: 920 Basic stats: COMPLETE 
Column stats: COMPLETE
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.dest1
 
   Stage: Stage-0
     Move Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/udtf_stack.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/udtf_stack.q.out 
b/ql/src/test/results/clientpositive/udtf_stack.q.out
index 97dcd3a..d1f880f 100644
--- a/ql/src/test/results/clientpositive/udtf_stack.q.out
+++ b/ql/src/test/results/clientpositive/udtf_stack.q.out
@@ -182,13 +182,17 @@ STAGE PLANS:
               UDTF Operator
                 Statistics: Num rows: 1 Data size: 185 Basic stats: COMPLETE 
Column stats: COMPLETE
                 function name: stack
-                File Output Operator
-                  compressed: false
-                  Statistics: Num rows: 1 Data size: 185 Basic stats: COMPLETE 
Column stats: COMPLETE
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                Select Operator
+                  expressions: col0 (type: string), col1 (type: string), 
UDFToString(col2) (type: string)
+                  outputColumnNames: _col0, _col1, _col2
+                  Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE 
Column stats: COMPLETE
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 1 Data size: 84 Basic stats: 
COMPLETE Column stats: COMPLETE
+                    table:
+                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
 
   Stage: Stage-0
     Fetch Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/union17.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/union17.q.out 
b/ql/src/test/results/clientpositive/union17.q.out
index 5c2e6ac..18e06e1 100644
--- a/ql/src/test/results/clientpositive/union17.q.out
+++ b/ql/src/test/results/clientpositive/union17.q.out
@@ -140,14 +140,18 @@ STAGE PLANS:
           mode: mergepartial
           outputColumnNames: _col0, _col1
           Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column 
stats: PARTIAL
-          File Output Operator
-            compressed: false
-            Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column 
stats: PARTIAL
-            table:
-                input format: org.apache.hadoop.mapred.TextInputFormat
-                output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                name: default.dest1
+          Select Operator
+            expressions: _col0 (type: string), UDFToString(_col1) (type: 
string)
+            outputColumnNames: _col0, _col1
+            Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE 
Column stats: PARTIAL
+            File Output Operator
+              compressed: false
+              Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE 
Column stats: PARTIAL
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.dest1
 
   Stage: Stage-0
     Move Operator
@@ -178,14 +182,18 @@ STAGE PLANS:
           mode: mergepartial
           outputColumnNames: _col0, _col1, _col2
           Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column 
stats: PARTIAL
-          File Output Operator
-            compressed: false
-            Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE 
Column stats: PARTIAL
-            table:
-                input format: org.apache.hadoop.mapred.TextInputFormat
-                output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                name: default.dest2
+          Select Operator
+            expressions: _col0 (type: string), _col1 (type: string), 
UDFToString(_col2) (type: string)
+            outputColumnNames: _col0, _col1, _col2
+            Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE 
Column stats: PARTIAL
+            File Output Operator
+              compressed: false
+              Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE 
Column stats: PARTIAL
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.dest2
 
   Stage: Stage-1
     Move Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/union19.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/union19.q.out 
b/ql/src/test/results/clientpositive/union19.q.out
index 5781ae5..5ce5905 100644
--- a/ql/src/test/results/clientpositive/union19.q.out
+++ b/ql/src/test/results/clientpositive/union19.q.out
@@ -141,14 +141,18 @@ STAGE PLANS:
           mode: mergepartial
           outputColumnNames: _col0, _col1
           Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column 
stats: PARTIAL
-          File Output Operator
-            compressed: false
-            Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column 
stats: PARTIAL
-            table:
-                input format: org.apache.hadoop.mapred.TextInputFormat
-                output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                name: default.dest1
+          Select Operator
+            expressions: _col0 (type: string), UDFToString(_col1) (type: 
string)
+            outputColumnNames: _col0, _col1
+            Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE 
Column stats: PARTIAL
+            File Output Operator
+              compressed: false
+              Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE 
Column stats: PARTIAL
+              table:
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                  name: default.dest1
 
   Stage: Stage-0
     Move Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/union_remove_19.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/union_remove_19.q.out 
b/ql/src/test/results/clientpositive/union_remove_19.q.out
index 5918afd..8a3a73f 100644
--- a/ql/src/test/results/clientpositive/union_remove_19.q.out
+++ b/ql/src/test/results/clientpositive/union_remove_19.q.out
@@ -415,7 +415,7 @@ STAGE PLANS:
           outputColumnNames: _col0, _col1
           Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column 
stats: NONE
           Select Operator
-            expressions: (UDFToDouble(_col0) + UDFToDouble(_col0)) (type: 
double), _col1 (type: bigint)
+            expressions: UDFToString((UDFToDouble(_col0) + 
UDFToDouble(_col0))) (type: string), _col1 (type: bigint)
             outputColumnNames: _col0, _col1
             Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column 
stats: NONE
             File Output Operator
@@ -470,7 +470,7 @@ STAGE PLANS:
           outputColumnNames: _col0, _col1
           Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column 
stats: NONE
           Select Operator
-            expressions: (UDFToDouble(_col0) + UDFToDouble(_col0)) (type: 
double), _col1 (type: bigint)
+            expressions: UDFToString((UDFToDouble(_col0) + 
UDFToDouble(_col0))) (type: string), _col1 (type: bigint)
             outputColumnNames: _col0, _col1
             Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column 
stats: NONE
             File Output Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/varchar_1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/varchar_1.q.out 
b/ql/src/test/results/clientpositive/varchar_1.q.out
index 57d8ee6..6b35013 100644
--- a/ql/src/test/results/clientpositive/varchar_1.q.out
+++ b/ql/src/test/results/clientpositive/varchar_1.q.out
@@ -94,8 +94,8 @@ POSTHOOK: query: insert overwrite table varchar1_1
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@varchar1
 POSTHOOK: Output: default@varchar1_1
-POSTHOOK: Lineage: varchar1_1.key SIMPLE 
[(varchar1)varchar1.FieldSchema(name:key, type:varchar(10), comment:null), ]
-POSTHOOK: Lineage: varchar1_1.value SIMPLE 
[(varchar1)varchar1.FieldSchema(name:value, type:varchar(20), comment:null), ]
+POSTHOOK: Lineage: varchar1_1.key EXPRESSION 
[(varchar1)varchar1.FieldSchema(name:key, type:varchar(10), comment:null), ]
+POSTHOOK: Lineage: varchar1_1.value EXPRESSION 
[(varchar1)varchar1.FieldSchema(name:value, type:varchar(20), comment:null), ]
 PREHOOK: query: select key, value from varchar1_1 order by key, value
 PREHOOK: type: QUERY
 PREHOOK: Input: default@varchar1_1

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/vector_null_projection.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/vector_null_projection.q.out 
b/ql/src/test/results/clientpositive/vector_null_projection.q.out
index 94aea2f..e9f7733 100644
--- a/ql/src/test/results/clientpositive/vector_null_projection.q.out
+++ b/ql/src/test/results/clientpositive/vector_null_projection.q.out
@@ -49,23 +49,37 @@ STAGE PLANS:
           TableScan
             alias: a
             Statistics: Num rows: 1 Data size: 87 Basic stats: COMPLETE Column 
stats: COMPLETE
+            TableScan Vectorization:
+                native: true
+                projectedOutputColumns: [0]
             Select Operator
-              expressions: null (type: void)
+              expressions: null (type: string)
               outputColumnNames: _col0
-              Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE 
Column stats: COMPLETE
+              Select Vectorization:
+                  className: VectorSelectOperator
+                  native: true
+                  projectedOutputColumns: [1]
+                  selectExpressions: ConstantVectorExpression(val null) -> 
1:string
+              Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE 
Column stats: COMPLETE
               File Output Operator
                 compressed: false
-                Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE 
Column stats: COMPLETE
+                File Sink Vectorization:
+                    className: VectorFileSinkOperator
+                    native: false
+                Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE 
Column stats: COMPLETE
                 table:
                     input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                     output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
                     serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+      Execution mode: vectorized
       Map Vectorization:
           enabled: true
           enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS 
true
+          groupByVectorOutput: true
           inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-          notVectorizedReason: Select expression for SELECT operator: Data 
type void of Const void null not supported
-          vectorized: false
+          allNative: false
+          usesVectorUDFAdaptor: false
+          vectorized: true
 
   Stage: Stage-0
     Fetch Operator
@@ -164,12 +178,12 @@ STAGE PLANS:
           outputColumnNames: _col0
           Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column 
stats: COMPLETE
           Select Operator
-            expressions: null (type: void)
+            expressions: null (type: string)
             outputColumnNames: _col0
-            Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column 
stats: COMPLETE
+            Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column 
stats: COMPLETE
             File Output Operator
               compressed: false
-              Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE 
Column stats: COMPLETE
+              Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE 
Column stats: COMPLETE
               table:
                   input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                   output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/96c2a2c0/ql/src/test/results/clientpositive/vector_nvl.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/vector_nvl.q.out 
b/ql/src/test/results/clientpositive/vector_nvl.q.out
index 08cc168..4a14288 100644
--- a/ql/src/test/results/clientpositive/vector_nvl.q.out
+++ b/ql/src/test/results/clientpositive/vector_nvl.q.out
@@ -306,13 +306,17 @@ STAGE PLANS:
               Limit
                 Number of rows: 10
                 Statistics: Num rows: 10 Data size: 4 Basic stats: COMPLETE 
Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  Statistics: Num rows: 10 Data size: 4 Basic stats: COMPLETE 
Column stats: COMPLETE
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                Select Operator
+                  expressions: UDFToString(_col0) (type: string)
+                  outputColumnNames: _col0
+                  Statistics: Num rows: 10 Data size: 84 Basic stats: COMPLETE 
Column stats: COMPLETE
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 10 Data size: 84 Basic stats: 
COMPLETE Column stats: COMPLETE
+                    table:
+                        input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
+                        output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
       Map Vectorization:
           enabled: true
           enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS 
true

Reply via email to