http://git-wip-us.apache.org/repos/asf/hive/blob/ad6ce078/ql/src/test/results/clientpositive/llap/vector_reduce1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_reduce1.q.out 
b/ql/src/test/results/clientpositive/llap/vector_reduce1.q.out
index 0bab7bd..4c252c7 100644
--- a/ql/src/test/results/clientpositive/llap/vector_reduce1.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_reduce1.q.out
@@ -105,16 +105,12 @@ POSTHOOK: Lineage: vectortab2korc.si SIMPLE 
[(vectortab2k)vectortab2k.FieldSchem
 POSTHOOK: Lineage: vectortab2korc.t SIMPLE 
[(vectortab2k)vectortab2k.FieldSchema(name:t, type:tinyint, comment:null), ]
 POSTHOOK: Lineage: vectortab2korc.ts SIMPLE 
[(vectortab2k)vectortab2k.FieldSchema(name:ts, type:timestamp, comment:null), ]
 POSTHOOK: Lineage: vectortab2korc.ts2 SIMPLE 
[(vectortab2k)vectortab2k.FieldSchema(name:ts2, type:timestamp, comment:null), ]
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select b from vectortab2korc order by b
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select b from vectortab2korc order by b
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -132,59 +128,25 @@ STAGE PLANS:
                 TableScan
                   alias: vectortab2korc
                   Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 
10, 11, 12]
                   Select Operator
                     expressions: b (type: bigint)
                     outputColumnNames: _col0
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [3]
                     Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       key expressions: _col0 (type: bigint)
                       sort order: +
-                      Reduce Sink Vectorization:
-                          className: VectorReduceSinkOperator
-                          native: false
-                          nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for 
keys IS true, LazyBinarySerDe for values IS true
-                          nativeConditionsNotMet: Uniform Hash IS false
                       Statistics: Num rows: 2000 Data size: 918712 Basic 
stats: COMPLETE Column stats: NONE
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: bigint)
                 outputColumnNames: _col0
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [0]
                 Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
                 File Output Operator
                   compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
                   Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/ad6ce078/ql/src/test/results/clientpositive/llap/vector_reduce2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_reduce2.q.out 
b/ql/src/test/results/clientpositive/llap/vector_reduce2.q.out
index 30c4f8e..a4ce890 100644
--- a/ql/src/test/results/clientpositive/llap/vector_reduce2.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_reduce2.q.out
@@ -105,16 +105,12 @@ POSTHOOK: Lineage: vectortab2korc.si SIMPLE 
[(vectortab2k)vectortab2k.FieldSchem
 POSTHOOK: Lineage: vectortab2korc.t SIMPLE 
[(vectortab2k)vectortab2k.FieldSchema(name:t, type:tinyint, comment:null), ]
 POSTHOOK: Lineage: vectortab2korc.ts SIMPLE 
[(vectortab2k)vectortab2k.FieldSchema(name:ts, type:timestamp, comment:null), ]
 POSTHOOK: Lineage: vectortab2korc.ts2 SIMPLE 
[(vectortab2k)vectortab2k.FieldSchema(name:ts2, type:timestamp, comment:null), ]
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select s, i, s2 from vectortab2korc order by s, i, s2
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select s, i, s2 from vectortab2korc order by s, i, s2
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -132,59 +128,25 @@ STAGE PLANS:
                 TableScan
                   alias: vectortab2korc
                   Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 
10, 11, 12]
                   Select Operator
                     expressions: s (type: string), i (type: int), s2 (type: 
string)
                     outputColumnNames: _col0, _col1, _col2
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [8, 2, 9]
                     Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       key expressions: _col0 (type: string), _col1 (type: 
int), _col2 (type: string)
                       sort order: +++
-                      Reduce Sink Vectorization:
-                          className: VectorReduceSinkOperator
-                          native: false
-                          nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for 
keys IS true, LazyBinarySerDe for values IS true
-                          nativeConditionsNotMet: Uniform Hash IS false
                       Statistics: Num rows: 2000 Data size: 918712 Basic 
stats: COMPLETE Column stats: NONE
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: string), 
KEY.reducesinkkey1 (type: int), KEY.reducesinkkey2 (type: string)
                 outputColumnNames: _col0, _col1, _col2
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [0, 1, 2]
                 Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
                 File Output Operator
                   compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
                   Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/ad6ce078/ql/src/test/results/clientpositive/llap/vector_reduce3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_reduce3.q.out 
b/ql/src/test/results/clientpositive/llap/vector_reduce3.q.out
index 21b9844..d34113c 100644
--- a/ql/src/test/results/clientpositive/llap/vector_reduce3.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_reduce3.q.out
@@ -105,16 +105,12 @@ POSTHOOK: Lineage: vectortab2korc.si SIMPLE 
[(vectortab2k)vectortab2k.FieldSchem
 POSTHOOK: Lineage: vectortab2korc.t SIMPLE 
[(vectortab2k)vectortab2k.FieldSchema(name:t, type:tinyint, comment:null), ]
 POSTHOOK: Lineage: vectortab2korc.ts SIMPLE 
[(vectortab2k)vectortab2k.FieldSchema(name:ts, type:timestamp, comment:null), ]
 POSTHOOK: Lineage: vectortab2korc.ts2 SIMPLE 
[(vectortab2k)vectortab2k.FieldSchema(name:ts2, type:timestamp, comment:null), ]
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select s from vectortab2korc order by s
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select s from vectortab2korc order by s
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -132,59 +128,25 @@ STAGE PLANS:
                 TableScan
                   alias: vectortab2korc
                   Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 
10, 11, 12]
                   Select Operator
                     expressions: s (type: string)
                     outputColumnNames: _col0
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [8]
                     Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       key expressions: _col0 (type: string)
                       sort order: +
-                      Reduce Sink Vectorization:
-                          className: VectorReduceSinkOperator
-                          native: false
-                          nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for 
keys IS true, LazyBinarySerDe for values IS true
-                          nativeConditionsNotMet: Uniform Hash IS false
                       Statistics: Num rows: 2000 Data size: 918712 Basic 
stats: COMPLETE Column stats: NONE
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: string)
                 outputColumnNames: _col0
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [0]
                 Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
                 File Output Operator
                   compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
                   Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/ad6ce078/ql/src/test/results/clientpositive/llap/vector_reduce_groupby_decimal.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_reduce_groupby_decimal.q.out 
b/ql/src/test/results/clientpositive/llap/vector_reduce_groupby_decimal.q.out
index 09efd32..9571b5b 100644
--- 
a/ql/src/test/results/clientpositive/llap/vector_reduce_groupby_decimal.q.out
+++ 
b/ql/src/test/results/clientpositive/llap/vector_reduce_groupby_decimal.q.out
@@ -14,24 +14,20 @@ POSTHOOK: Lineage: decimal_test.cdecimal1 EXPRESSION 
[(alltypesorc)alltypesorc.F
 POSTHOOK: Lineage: decimal_test.cdecimal2 EXPRESSION 
[(alltypesorc)alltypesorc.FieldSchema(name:cdouble, type:double, comment:null), 
]
 POSTHOOK: Lineage: decimal_test.cdouble SIMPLE 
[(alltypesorc)alltypesorc.FieldSchema(name:cdouble, type:double, comment:null), 
]
 POSTHOOK: Lineage: decimal_test.cint SIMPLE 
[(alltypesorc)alltypesorc.FieldSchema(name:cint, type:int, comment:null), ]
-PREHOOK: query: EXPLAIN VECTORIZATION EXPRESSION
+PREHOOK: query: EXPLAIN
 SELECT cint, cdouble, cdecimal1, cdecimal2, min(cdecimal1) as min_decimal1 
FROM decimal_test
 WHERE cdecimal1 is not null and cdecimal2 is not null
 GROUP BY cint, cdouble, cdecimal1, cdecimal2
 ORDER BY cint, cdouble, cdecimal1, cdecimal2
 LIMIT 50
 PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN VECTORIZATION EXPRESSION
+POSTHOOK: query: EXPLAIN
 SELECT cint, cdouble, cdecimal1, cdecimal2, min(cdecimal1) as min_decimal1 
FROM decimal_test
 WHERE cdecimal1 is not null and cdecimal2 is not null
 GROUP BY cint, cdouble, cdecimal1, cdecimal2
 ORDER BY cint, cdouble, cdecimal1, cdecimal2
 LIMIT 50
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -50,25 +46,11 @@ STAGE PLANS:
                 TableScan
                   alias: decimal_test
                   Statistics: Num rows: 6102 Data size: 1440072 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3]
                   Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: FilterExprAndExpr(children: 
SelectColumnIsNotNull(col 4)(children: CastDecimalToBoolean(col 2) -> 4:long) 
-> boolean, SelectColumnIsNotNull(col 4)(children: CastDecimalToBoolean(col 3) 
-> 4:long) -> boolean) -> boolean
                     predicate: (cdecimal1 is not null and cdecimal2 is not 
null) (type: boolean)
                     Statistics: Num rows: 6102 Data size: 1440072 Basic stats: 
COMPLETE Column stats: NONE
                     Group By Operator
                       aggregations: min(cdecimal1)
-                      Group By Vectorization:
-                          aggregators: VectorUDAFMinDecimal(col 2) -> 
decimal(20,10)
-                          className: VectorGroupByOperator
-                          vectorOutput: true
-                          keyExpressions: col 0, col 1, col 2, col 3
-                          native: false
-                          projectedOutputColumns: [0]
                       keys: cint (type: int), cdouble (type: double), 
cdecimal1 (type: decimal(20,10)), cdecimal2 (type: decimal(23,14))
                       mode: hash
                       outputColumnNames: _col0, _col1, _col2, _col3, _col4
@@ -77,43 +59,16 @@ STAGE PLANS:
                         key expressions: _col0 (type: int), _col1 (type: 
double), _col2 (type: decimal(20,10)), _col3 (type: decimal(23,14))
                         sort order: ++++
                         Map-reduce partition columns: _col0 (type: int), _col1 
(type: double), _col2 (type: decimal(20,10)), _col3 (type: decimal(23,14))
-                        Reduce Sink Vectorization:
-                            className: VectorReduceSinkOperator
-                            native: false
-                            nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, Uniform Hash IS true, No DISTINCT columns IS true, BinarySortableSerDe 
for keys IS true, LazyBinarySerDe for values IS true
-                            nativeConditionsNotMet: No TopN IS false
                         Statistics: Num rows: 6102 Data size: 1440072 Basic 
stats: COMPLETE Column stats: NONE
                         TopN Hash Memory Usage: 0.1
                         value expressions: _col4 (type: decimal(20,10))
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Group By Operator
                 aggregations: min(VALUE._col0)
-                Group By Vectorization:
-                    aggregators: VectorUDAFMinDecimal(col 4) -> decimal(20,10)
-                    className: VectorGroupByOperator
-                    vectorOutput: true
-                    keyExpressions: col 0, col 1, col 2, col 3
-                    native: false
-                    projectedOutputColumns: [0]
                 keys: KEY._col0 (type: int), KEY._col1 (type: double), 
KEY._col2 (type: decimal(20,10)), KEY._col3 (type: decimal(23,14))
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4
@@ -121,43 +76,21 @@ STAGE PLANS:
                 Reduce Output Operator
                   key expressions: _col0 (type: int), _col1 (type: double), 
_col2 (type: decimal(20,10)), _col3 (type: decimal(23,14))
                   sort order: ++++
-                  Reduce Sink Vectorization:
-                      className: VectorReduceSinkOperator
-                      native: false
-                      nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, 
LazyBinarySerDe for values IS true
-                      nativeConditionsNotMet: No TopN IS false, Uniform Hash 
IS false
                   Statistics: Num rows: 3051 Data size: 720036 Basic stats: 
COMPLETE Column stats: NONE
                   TopN Hash Memory Usage: 0.1
                   value expressions: _col4 (type: decimal(20,10))
         Reducer 3 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: int), 
KEY.reducesinkkey1 (type: double), KEY.reducesinkkey2 (type: decimal(20,10)), 
KEY.reducesinkkey3 (type: decimal(23,14)), VALUE._col0 (type: decimal(20,10))
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [0, 1, 2, 3, 4]
                 Statistics: Num rows: 3051 Data size: 720036 Basic stats: 
COMPLETE Column stats: NONE
                 Limit
                   Number of rows: 50
-                  Limit Vectorization:
-                      className: VectorLimitOperator
-                      native: true
                   Statistics: Num rows: 50 Data size: 11800 Basic stats: 
COMPLETE Column stats: NONE
                   File Output Operator
                     compressed: false
-                    File Sink Vectorization:
-                        className: VectorFileSinkOperator
-                        native: false
                     Statistics: Num rows: 50 Data size: 11800 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/ad6ce078/ql/src/test/results/clientpositive/llap/vector_string_concat.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_string_concat.q.out 
b/ql/src/test/results/clientpositive/llap/vector_string_concat.q.out
index 4927f17..ca3d2fa 100644
--- a/ql/src/test/results/clientpositive/llap/vector_string_concat.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_string_concat.q.out
@@ -97,78 +97,32 @@ POSTHOOK: Lineage: over1korc.s SIMPLE 
[(over1k)over1k.FieldSchema(name:s, type:s
 POSTHOOK: Lineage: over1korc.si SIMPLE [(over1k)over1k.FieldSchema(name:si, 
type:smallint, comment:null), ]
 POSTHOOK: Lineage: over1korc.t SIMPLE [(over1k)over1k.FieldSchema(name:t, 
type:tinyint, comment:null), ]
 POSTHOOK: Lineage: over1korc.ts SIMPLE [(over1k)over1k.FieldSchema(name:ts, 
type:timestamp, comment:null), ]
-PREHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT s AS `string`,
+PREHOOK: query: EXPLAIN SELECT s AS `string`,
        CONCAT(CONCAT('      ',s),'      ') AS `none_padded_str`,
        CONCAT(CONCAT('|',RTRIM(CONCAT(CONCAT('      ',s),'      '))),'|') AS 
`none_z_rtrim_str`
        FROM over1korc LIMIT 20
 PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT s AS `string`,
+POSTHOOK: query: EXPLAIN SELECT s AS `string`,
        CONCAT(CONCAT('      ',s),'      ') AS `none_padded_str`,
        CONCAT(CONCAT('|',RTRIM(CONCAT(CONCAT('      ',s),'      '))),'|') AS 
`none_z_rtrim_str`
        FROM over1korc LIMIT 20
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
+  Stage-0 is a root stage
 
 STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: over1korc
-                  Statistics: Num rows: 1049 Data size: 311170 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 
10]
-                  Select Operator
-                    expressions: s (type: string), concat(concat('      ', s), 
'      ') (type: string), concat(concat('|', rtrim(concat(concat('      ', s), 
'      '))), '|') (type: string)
-                    outputColumnNames: _col0, _col1, _col2
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [7, 12, 11]
-                        selectExpressions: 
StringGroupColConcatStringScalar(col 11, val       )(children: 
StringScalarConcatStringGroupCol(val       , col 7) -> 11:String_Family) -> 
12:String_Family, StringGroupColConcatStringScalar(col 13, val |)(children: 
StringScalarConcatStringGroupCol(val |, col 11)(children: StringRTrim(col 
13)(children: StringGroupColConcatStringScalar(col 11, val       )(children: 
StringScalarConcatStringGroupCol(val       , col 7) -> 11:String_Family) -> 
13:String_Family) -> 11:String) -> 13:String_Family) -> 11:String_Family
-                    Statistics: Num rows: 1049 Data size: 311170 Basic stats: 
COMPLETE Column stats: NONE
-                    Limit
-                      Number of rows: 20
-                      Limit Vectorization:
-                          className: VectorLimitOperator
-                          native: true
-                      Statistics: Num rows: 20 Data size: 5920 Basic stats: 
COMPLETE Column stats: NONE
-                      File Output Operator
-                        compressed: false
-                        File Sink Vectorization:
-                            className: VectorFileSinkOperator
-                            native: false
-                        Statistics: Num rows: 20 Data size: 5920 Basic stats: 
COMPLETE Column stats: NONE
-                        table:
-                            input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                            output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                            serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-
   Stage: Stage-0
     Fetch Operator
       limit: 20
       Processor Tree:
-        ListSink
+        TableScan
+          alias: over1korc
+          Select Operator
+            expressions: s (type: string), concat(concat('      ', s), '      
') (type: string), concat(concat('|', rtrim(concat(concat('      ', s), '      
'))), '|') (type: string)
+            outputColumnNames: _col0, _col1, _col2
+            Limit
+              Number of rows: 20
+              ListSink
 
 PREHOOK: query: SELECT s AS `string`,
        CONCAT(CONCAT('      ',s),'      ') AS `none_padded_str`,
@@ -311,24 +265,20 @@ POSTHOOK: Lineage: vectortab2korc.si SIMPLE 
[(vectortab2k)vectortab2k.FieldSchem
 POSTHOOK: Lineage: vectortab2korc.t SIMPLE 
[(vectortab2k)vectortab2k.FieldSchema(name:t, type:tinyint, comment:null), ]
 POSTHOOK: Lineage: vectortab2korc.ts SIMPLE 
[(vectortab2k)vectortab2k.FieldSchema(name:ts, type:timestamp, comment:null), ]
 POSTHOOK: Lineage: vectortab2korc.ts2 SIMPLE 
[(vectortab2k)vectortab2k.FieldSchema(name:ts2, type:timestamp, comment:null), ]
-PREHOOK: query: EXPLAIN VECTORIZATION EXPRESSION
+PREHOOK: query: EXPLAIN
 SELECT CONCAT(CONCAT(CONCAT('Quarter ',CAST(CAST((MONTH(dt) - 1) / 3 + 1 AS 
INT) AS STRING)),'-'),CAST(YEAR(dt) AS STRING)) AS `field`
     FROM vectortab2korc 
     GROUP BY CONCAT(CONCAT(CONCAT('Quarter ',CAST(CAST((MONTH(dt) - 1) / 3 + 1 
AS INT) AS STRING)),'-'),CAST(YEAR(dt) AS STRING))
     ORDER BY `field`
     LIMIT 50
 PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN VECTORIZATION EXPRESSION
+POSTHOOK: query: EXPLAIN
 SELECT CONCAT(CONCAT(CONCAT('Quarter ',CAST(CAST((MONTH(dt) - 1) / 3 + 1 AS 
INT) AS STRING)),'-'),CAST(YEAR(dt) AS STRING)) AS `field`
     FROM vectortab2korc 
     GROUP BY CONCAT(CONCAT(CONCAT('Quarter ',CAST(CAST((MONTH(dt) - 1) / 3 + 1 
AS INT) AS STRING)),'-'),CAST(YEAR(dt) AS STRING))
     ORDER BY `field`
     LIMIT 50
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -347,25 +297,11 @@ STAGE PLANS:
                 TableScan
                   alias: vectortab2korc
                   Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 
10, 11, 12]
                   Select Operator
                     expressions: concat(concat(concat('Quarter ', 
UDFToString(UDFToInteger(((UDFToDouble((month(dt) - 1)) / 3.0) + 1.0)))), '-'), 
UDFToString(year(dt))) (type: string)
                     outputColumnNames: _col0
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [19]
-                        selectExpressions: StringGroupConcatColCol(col 17, col 
18)(children: StringGroupColConcatStringScalar(col 18, val -)(children: 
StringScalarConcatStringGroupCol(val Quarter , col 17)(children: 
CastLongToString(col 13)(children: CastDoubleToLong(col 15)(children: 
DoubleColAddDoubleScalar(col 16, val 1.0)(children: 
DoubleColDivideDoubleScalar(col 15, val 3.0)(children: CastLongToDouble(col 
14)(children: LongColSubtractLongScalar(col 13, val 1)(children: 
VectorUDFMonthDate(col 12, field MONTH) -> 13:long) -> 14:long) -> 15:double) 
-> 16:double) -> 15:double) -> 13:long) -> 17:String) -> 18:String_Family) -> 
17:String_Family, CastLongToString(col 13)(children: VectorUDFYearDate(col 12, 
field YEAR) -> 13:long) -> 18:String) -> 19:String_Family
                     Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
                     Group By Operator
-                      Group By Vectorization:
-                          className: VectorGroupByOperator
-                          vectorOutput: true
-                          keyExpressions: col 19
-                          native: false
-                          projectedOutputColumns: []
                       keys: _col0 (type: string)
                       mode: hash
                       outputColumnNames: _col0
@@ -374,40 +310,14 @@ STAGE PLANS:
                         key expressions: _col0 (type: string)
                         sort order: +
                         Map-reduce partition columns: _col0 (type: string)
-                        Reduce Sink Vectorization:
-                            className: VectorReduceSinkOperator
-                            native: false
-                            nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, Uniform Hash IS true, No DISTINCT columns IS true, BinarySortableSerDe 
for keys IS true, LazyBinarySerDe for values IS true
-                            nativeConditionsNotMet: No TopN IS false
                         Statistics: Num rows: 2000 Data size: 918712 Basic 
stats: COMPLETE Column stats: NONE
                         TopN Hash Memory Usage: 0.1
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Group By Operator
-                Group By Vectorization:
-                    className: VectorGroupByOperator
-                    vectorOutput: true
-                    keyExpressions: col 0
-                    native: false
-                    projectedOutputColumns: []
                 keys: KEY._col0 (type: string)
                 mode: mergepartial
                 outputColumnNames: _col0
@@ -415,42 +325,20 @@ STAGE PLANS:
                 Reduce Output Operator
                   key expressions: _col0 (type: string)
                   sort order: +
-                  Reduce Sink Vectorization:
-                      className: VectorReduceSinkOperator
-                      native: false
-                      nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, 
LazyBinarySerDe for values IS true
-                      nativeConditionsNotMet: No TopN IS false, Uniform Hash 
IS false
                   Statistics: Num rows: 1000 Data size: 459356 Basic stats: 
COMPLETE Column stats: NONE
                   TopN Hash Memory Usage: 0.1
         Reducer 3 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: string)
                 outputColumnNames: _col0
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [0]
                 Statistics: Num rows: 1000 Data size: 459356 Basic stats: 
COMPLETE Column stats: NONE
                 Limit
                   Number of rows: 50
-                  Limit Vectorization:
-                      className: VectorLimitOperator
-                      native: true
                   Statistics: Num rows: 50 Data size: 22950 Basic stats: 
COMPLETE Column stats: NONE
                   File Output Operator
                     compressed: false
-                    File Sink Vectorization:
-                        className: VectorFileSinkOperator
-                        native: false
                     Statistics: Num rows: 50 Data size: 22950 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/ad6ce078/ql/src/test/results/clientpositive/llap/vector_struct_in.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_struct_in.q.out 
b/ql/src/test/results/clientpositive/llap/vector_struct_in.q.out
index dedc4e6..0aa1e70 100644
--- a/ql/src/test/results/clientpositive/llap/vector_struct_in.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_struct_in.q.out
@@ -22,7 +22,7 @@ POSTHOOK: Input: default@values__tmp__table__1
 POSTHOOK: Output: default@test_1
 POSTHOOK: Lineage: test_1.id SIMPLE 
[(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, 
type:string, comment:), ]
 POSTHOOK: Lineage: test_1.lineid SIMPLE 
[(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, 
type:string, comment:), ]
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select * from test_1 where struct(`id`, `lineid`)
 IN (
 struct('two','3'),
@@ -36,7 +36,7 @@ struct('nine','1'),
 struct('ten','1')
 )
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select * from test_1 where struct(`id`, `lineid`)
 IN (
 struct('two','3'),
@@ -50,68 +50,15 @@ struct('nine','1'),
 struct('ten','1')
 )
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test_1
-                  Statistics: Num rows: 2 Data size: 346 Basic stats: COMPLETE 
Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1]
-                  Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: 
FilterStructColumnInList(structExpressions [col 0, col 1], 
fieldVectorColumnTypes [BYTES, BYTES], structColumnMap [0, 1]) -> boolean
-                    predicate: (struct(id,lineid)) IN (const 
struct('two','3'), const struct('three','1'), const struct('one','1'), const 
struct('five','2'), const struct('six','1'), const struct('eight','1'), const 
struct('seven','1'), const struct('nine','1'), const struct('ten','1')) (type: 
boolean)
-                    Statistics: Num rows: 1 Data size: 173 Basic stats: 
COMPLETE Column stats: NONE
-                    Select Operator
-                      expressions: id (type: string), lineid (type: string)
-                      outputColumnNames: _col0, _col1
-                      Select Vectorization:
-                          className: VectorSelectOperator
-                          native: true
-                          projectedOutputColumns: [0, 1]
-                      Statistics: Num rows: 1 Data size: 173 Basic stats: 
COMPLETE Column stats: NONE
-                      File Output Operator
-                        compressed: false
-                        File Sink Vectorization:
-                            className: VectorFileSinkOperator
-                            native: false
-                        Statistics: Num rows: 1 Data size: 173 Basic stats: 
COMPLETE Column stats: NONE
-                        table:
-                            input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                            output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                            serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
+Stage-0
+  Fetch Operator
+    limit:-1
+    Select Operator [SEL_2]
+      Output:["_col0","_col1"]
+      Filter Operator [FIL_4]
+        predicate:(struct(id,lineid)) IN (const struct('two','3'), const 
struct('three','1'), const struct('one','1'), const struct('five','2'), const 
struct('six','1'), const struct('eight','1'), const struct('seven','1'), const 
struct('nine','1'), const struct('ten','1'))
+        TableScan [TS_0]
+          Output:["id","lineid"]
 
 PREHOOK: query: select * from test_1 where struct(`id`, `lineid`)
 IN (
@@ -145,7 +92,7 @@ POSTHOOK: Input: default@test_1
 #### A masked pattern was here ####
 one    1
 seven  1
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select `id`, `lineid`, struct(`id`, `lineid`)
 IN (
 struct('two','3'),
@@ -159,7 +106,7 @@ struct('nine','1'),
 struct('ten','1')
 ) as b from test_1
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select `id`, `lineid`, struct(`id`, `lineid`)
 IN (
 struct('two','3'),
@@ -173,62 +120,13 @@ struct('nine','1'),
 struct('ten','1')
 ) as b from test_1
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test_1
-                  Statistics: Num rows: 2 Data size: 346 Basic stats: COMPLETE 
Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1]
-                  Select Operator
-                    expressions: id (type: string), lineid (type: string), 
(struct(id,lineid)) IN (const struct('two','3'), const struct('three','1'), 
const struct('one','1'), const struct('five','2'), const struct('six','1'), 
const struct('eight','1'), const struct('seven','1'), const struct('nine','1'), 
const struct('ten','1')) (type: boolean)
-                    outputColumnNames: _col0, _col1, _col2
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [0, 1, 3]
-                        selectExpressions: 
StructColumnInList(structExpressions [col 0, col 1], fieldVectorColumnTypes 
[BYTES, BYTES], structColumnMap [0, 1]) -> 3:boolean
-                    Statistics: Num rows: 2 Data size: 346 Basic stats: 
COMPLETE Column stats: NONE
-                    File Output Operator
-                      compressed: false
-                      File Sink Vectorization:
-                          className: VectorFileSinkOperator
-                          native: false
-                      Statistics: Num rows: 2 Data size: 346 Basic stats: 
COMPLETE Column stats: NONE
-                      table:
-                          input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                          output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                          serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
+Stage-0
+  Fetch Operator
+    limit:-1
+    Select Operator [SEL_1]
+      Output:["_col0","_col1","_col2"]
+      TableScan [TS_0]
+        Output:["id","lineid"]
 
 PREHOOK: query: select `id`, `lineid`, struct(`id`, `lineid`)
 IN (
@@ -282,7 +180,7 @@ POSTHOOK: Input: default@values__tmp__table__2
 POSTHOOK: Output: default@test_2
 POSTHOOK: Lineage: test_2.id EXPRESSION 
[(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col1, 
type:string, comment:), ]
 POSTHOOK: Lineage: test_2.lineid EXPRESSION 
[(values__tmp__table__2)values__tmp__table__2.FieldSchema(name:tmp_values_col2, 
type:string, comment:), ]
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select * from test_2 where struct(`id`, `lineid`)
 IN (
 struct(2,3),
@@ -296,7 +194,7 @@ struct(9,1),
 struct(10,1)
 )
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select * from test_2 where struct(`id`, `lineid`)
 IN (
 struct(2,3),
@@ -310,68 +208,15 @@ struct(9,1),
 struct(10,1)
 )
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test_2
-                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE 
Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1]
-                  Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: 
FilterStructColumnInList(structExpressions [col 0, col 1], 
fieldVectorColumnTypes [LONG, LONG], structColumnMap [0, 1]) -> boolean
-                    predicate: (struct(id,lineid)) IN (const struct(2,3), 
const struct(3,1), const struct(1,1), const struct(5,2), const struct(6,1), 
const struct(8,1), const struct(7,1), const struct(9,1), const struct(10,1)) 
(type: boolean)
-                    Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
-                    Select Operator
-                      expressions: id (type: int), lineid (type: int)
-                      outputColumnNames: _col0, _col1
-                      Select Vectorization:
-                          className: VectorSelectOperator
-                          native: true
-                          projectedOutputColumns: [0, 1]
-                      Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
-                      File Output Operator
-                        compressed: false
-                        File Sink Vectorization:
-                            className: VectorFileSinkOperator
-                            native: false
-                        Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
-                        table:
-                            input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                            output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                            serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
+Stage-0
+  Fetch Operator
+    limit:-1
+    Select Operator [SEL_2]
+      Output:["_col0","_col1"]
+      Filter Operator [FIL_4]
+        predicate:(struct(id,lineid)) IN (const struct(2,3), const 
struct(3,1), const struct(1,1), const struct(5,2), const struct(6,1), const 
struct(8,1), const struct(7,1), const struct(9,1), const struct(10,1))
+        TableScan [TS_0]
+          Output:["id","lineid"]
 
 PREHOOK: query: select * from test_2 where struct(`id`, `lineid`)
 IN (
@@ -405,7 +250,7 @@ POSTHOOK: Input: default@test_2
 #### A masked pattern was here ####
 1      1
 7      1
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select `id`, `lineid`, struct(`id`, `lineid`)
 IN (
 struct(2,3),
@@ -419,7 +264,7 @@ struct(9,1),
 struct(10,1)
 ) as b from test_2
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select `id`, `lineid`, struct(`id`, `lineid`)
 IN (
 struct(2,3),
@@ -433,62 +278,13 @@ struct(9,1),
 struct(10,1)
 ) as b from test_2
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test_2
-                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE 
Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1]
-                  Select Operator
-                    expressions: id (type: int), lineid (type: int), 
(struct(id,lineid)) IN (const struct(2,3), const struct(3,1), const 
struct(1,1), const struct(5,2), const struct(6,1), const struct(8,1), const 
struct(7,1), const struct(9,1), const struct(10,1)) (type: boolean)
-                    outputColumnNames: _col0, _col1, _col2
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [0, 1, 3]
-                        selectExpressions: 
StructColumnInList(structExpressions [col 0, col 1], fieldVectorColumnTypes 
[LONG, LONG], structColumnMap [0, 1]) -> 3:boolean
-                    Statistics: Num rows: 2 Data size: 16 Basic stats: 
COMPLETE Column stats: NONE
-                    File Output Operator
-                      compressed: false
-                      File Sink Vectorization:
-                          className: VectorFileSinkOperator
-                          native: false
-                      Statistics: Num rows: 2 Data size: 16 Basic stats: 
COMPLETE Column stats: NONE
-                      table:
-                          input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                          output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                          serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
+Stage-0
+  Fetch Operator
+    limit:-1
+    Select Operator [SEL_1]
+      Output:["_col0","_col1","_col2"]
+      TableScan [TS_0]
+        Output:["id","lineid"]
 
 PREHOOK: query: select `id`, `lineid`, struct(`id`, `lineid`)
 IN (
@@ -542,7 +338,7 @@ POSTHOOK: Input: default@values__tmp__table__3
 POSTHOOK: Output: default@test_3
 POSTHOOK: Lineage: test_3.id SIMPLE 
[(values__tmp__table__3)values__tmp__table__3.FieldSchema(name:tmp_values_col1, 
type:string, comment:), ]
 POSTHOOK: Lineage: test_3.lineid EXPRESSION 
[(values__tmp__table__3)values__tmp__table__3.FieldSchema(name:tmp_values_col2, 
type:string, comment:), ]
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select * from test_3 where struct(`id`, `lineid`)
 IN (
 struct('two',3),
@@ -556,7 +352,7 @@ struct('nine',1),
 struct('ten',1)
 )
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select * from test_3 where struct(`id`, `lineid`)
 IN (
 struct('two',3),
@@ -570,68 +366,15 @@ struct('nine',1),
 struct('ten',1)
 )
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test_3
-                  Statistics: Num rows: 2 Data size: 184 Basic stats: COMPLETE 
Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1]
-                  Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: 
FilterStructColumnInList(structExpressions [col 0, col 1], 
fieldVectorColumnTypes [BYTES, LONG], structColumnMap [0, 1]) -> boolean
-                    predicate: (struct(id,lineid)) IN (const struct('two',3), 
const struct('three',1), const struct('one',1), const struct('five',2), const 
struct('six',1), const struct('eight',1), const struct('seven',1), const 
struct('nine',1), const struct('ten',1)) (type: boolean)
-                    Statistics: Num rows: 1 Data size: 92 Basic stats: 
COMPLETE Column stats: NONE
-                    Select Operator
-                      expressions: id (type: string), lineid (type: int)
-                      outputColumnNames: _col0, _col1
-                      Select Vectorization:
-                          className: VectorSelectOperator
-                          native: true
-                          projectedOutputColumns: [0, 1]
-                      Statistics: Num rows: 1 Data size: 92 Basic stats: 
COMPLETE Column stats: NONE
-                      File Output Operator
-                        compressed: false
-                        File Sink Vectorization:
-                            className: VectorFileSinkOperator
-                            native: false
-                        Statistics: Num rows: 1 Data size: 92 Basic stats: 
COMPLETE Column stats: NONE
-                        table:
-                            input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                            output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                            serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
+Stage-0
+  Fetch Operator
+    limit:-1
+    Select Operator [SEL_2]
+      Output:["_col0","_col1"]
+      Filter Operator [FIL_4]
+        predicate:(struct(id,lineid)) IN (const struct('two',3), const 
struct('three',1), const struct('one',1), const struct('five',2), const 
struct('six',1), const struct('eight',1), const struct('seven',1), const 
struct('nine',1), const struct('ten',1))
+        TableScan [TS_0]
+          Output:["id","lineid"]
 
 PREHOOK: query: select * from test_3 where struct(`id`, `lineid`)
 IN (
@@ -665,7 +408,7 @@ POSTHOOK: Input: default@test_3
 #### A masked pattern was here ####
 one    1
 seven  1
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select `id`, `lineid`, struct(`id`, `lineid`)
 IN (
 struct('two',3),
@@ -679,7 +422,7 @@ struct('nine',1),
 struct('ten',1)
 ) as b from test_3
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select `id`, `lineid`, struct(`id`, `lineid`)
 IN (
 struct('two',3),
@@ -693,62 +436,13 @@ struct('nine',1),
 struct('ten',1)
 ) as b from test_3
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test_3
-                  Statistics: Num rows: 2 Data size: 184 Basic stats: COMPLETE 
Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1]
-                  Select Operator
-                    expressions: id (type: string), lineid (type: int), 
(struct(id,lineid)) IN (const struct('two',3), const struct('three',1), const 
struct('one',1), const struct('five',2), const struct('six',1), const 
struct('eight',1), const struct('seven',1), const struct('nine',1), const 
struct('ten',1)) (type: boolean)
-                    outputColumnNames: _col0, _col1, _col2
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [0, 1, 3]
-                        selectExpressions: 
StructColumnInList(structExpressions [col 0, col 1], fieldVectorColumnTypes 
[BYTES, LONG], structColumnMap [0, 1]) -> 3:boolean
-                    Statistics: Num rows: 2 Data size: 184 Basic stats: 
COMPLETE Column stats: NONE
-                    File Output Operator
-                      compressed: false
-                      File Sink Vectorization:
-                          className: VectorFileSinkOperator
-                          native: false
-                      Statistics: Num rows: 2 Data size: 184 Basic stats: 
COMPLETE Column stats: NONE
-                      table:
-                          input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                          output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                          serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
+Stage-0
+  Fetch Operator
+    limit:-1
+    Select Operator [SEL_1]
+      Output:["_col0","_col1","_col2"]
+      TableScan [TS_0]
+        Output:["id","lineid"]
 
 PREHOOK: query: select `id`, `lineid`, struct(`id`, `lineid`)
 IN (
@@ -803,7 +497,7 @@ POSTHOOK: Output: default@test_4
 POSTHOOK: Lineage: test_4.my_bigint EXPRESSION 
[(values__tmp__table__4)values__tmp__table__4.FieldSchema(name:tmp_values_col1, 
type:string, comment:), ]
 POSTHOOK: Lineage: test_4.my_double EXPRESSION 
[(values__tmp__table__4)values__tmp__table__4.FieldSchema(name:tmp_values_col3, 
type:string, comment:), ]
 POSTHOOK: Lineage: test_4.my_string SIMPLE 
[(values__tmp__table__4)values__tmp__table__4.FieldSchema(name:tmp_values_col2, 
type:string, comment:), ]
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select * from test_4 where struct(`my_bigint`, `my_string`, `my_double`)
 IN (
 struct(1L, "a", 1.5D),
@@ -818,7 +512,7 @@ struct(1L, "a", 0.5D),
 struct(3L, "b", 1.5D)
 )
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select * from test_4 where struct(`my_bigint`, `my_string`, `my_double`)
 IN (
 struct(1L, "a", 1.5D),
@@ -833,68 +527,15 @@ struct(1L, "a", 0.5D),
 struct(3L, "b", 1.5D)
 )
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test_4
-                  Statistics: Num rows: 3 Data size: 303 Basic stats: COMPLETE 
Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2]
-                  Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: 
FilterStructColumnInList(structExpressions [col 0, col 1, col 2], 
fieldVectorColumnTypes [LONG, BYTES, DOUBLE], structColumnMap [0, 1, 2]) -> 
boolean
-                    predicate: (struct(my_bigint,my_string,my_double)) IN 
(const struct(1,'a',1.5), const struct(1,'b',-0.5), const struct(3,'b',1.5), 
const struct(1,'d',1.5), const struct(1,'c',1.5), const struct(1,'b',2.5), 
const struct(1,'b',0.5), const struct(5,'b',1.5), const struct(1,'a',0.5), 
const struct(3,'b',1.5)) (type: boolean)
-                    Statistics: Num rows: 1 Data size: 101 Basic stats: 
COMPLETE Column stats: NONE
-                    Select Operator
-                      expressions: my_bigint (type: bigint), my_string (type: 
string), my_double (type: double)
-                      outputColumnNames: _col0, _col1, _col2
-                      Select Vectorization:
-                          className: VectorSelectOperator
-                          native: true
-                          projectedOutputColumns: [0, 1, 2]
-                      Statistics: Num rows: 1 Data size: 101 Basic stats: 
COMPLETE Column stats: NONE
-                      File Output Operator
-                        compressed: false
-                        File Sink Vectorization:
-                            className: VectorFileSinkOperator
-                            native: false
-                        Statistics: Num rows: 1 Data size: 101 Basic stats: 
COMPLETE Column stats: NONE
-                        table:
-                            input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                            output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                            serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
+Stage-0
+  Fetch Operator
+    limit:-1
+    Select Operator [SEL_2]
+      Output:["_col0","_col1","_col2"]
+      Filter Operator [FIL_4]
+        predicate:(struct(my_bigint,my_string,my_double)) IN (const 
struct(1,'a',1.5), const struct(1,'b',-0.5), const struct(3,'b',1.5), const 
struct(1,'d',1.5), const struct(1,'c',1.5), const struct(1,'b',2.5), const 
struct(1,'b',0.5), const struct(5,'b',1.5), const struct(1,'a',0.5), const 
struct(3,'b',1.5))
+        TableScan [TS_0]
+          Output:["my_bigint","my_string","my_double"]
 
 PREHOOK: query: select * from test_4 where struct(`my_bigint`, `my_string`, 
`my_double`)
 IN (
@@ -929,7 +570,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_4
 #### A masked pattern was here ####
 1      a       0.5
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select `my_bigint`, `my_string`, `my_double`, struct(`my_bigint`, `my_string`, 
`my_double`)
 IN (
 struct(1L, "a", 1.5D),
@@ -944,7 +585,7 @@ struct(1L, "a", 0.5D),
 struct(3L, "b", 1.5D)
 ) as b from test_4
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select `my_bigint`, `my_string`, `my_double`, struct(`my_bigint`, `my_string`, 
`my_double`)
 IN (
 struct(1L, "a", 1.5D),
@@ -959,62 +600,13 @@ struct(1L, "a", 0.5D),
 struct(3L, "b", 1.5D)
 ) as b from test_4
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: test_4
-                  Statistics: Num rows: 3 Data size: 303 Basic stats: COMPLETE 
Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2]
-                  Select Operator
-                    expressions: my_bigint (type: bigint), my_string (type: 
string), my_double (type: double), (struct(my_bigint,my_string,my_double)) IN 
(const struct(1,'a',1.5), const struct(1,'b',-0.5), const struct(3,'b',1.5), 
const struct(1,'d',1.5), const struct(1,'c',1.5), const struct(1,'b',2.5), 
const struct(1,'b',0.5), const struct(5,'b',1.5), const struct(1,'a',0.5), 
const struct(3,'b',1.5)) (type: boolean)
-                    outputColumnNames: _col0, _col1, _col2, _col3
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [0, 1, 2, 4]
-                        selectExpressions: 
StructColumnInList(structExpressions [col 0, col 1, col 2], 
fieldVectorColumnTypes [LONG, BYTES, DOUBLE], structColumnMap [0, 1, 2]) -> 
4:boolean
-                    Statistics: Num rows: 3 Data size: 303 Basic stats: 
COMPLETE Column stats: NONE
-                    File Output Operator
-                      compressed: false
-                      File Sink Vectorization:
-                          className: VectorFileSinkOperator
-                          native: false
-                      Statistics: Num rows: 3 Data size: 303 Basic stats: 
COMPLETE Column stats: NONE
-                      table:
-                          input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                          output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                          serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
+Stage-0
+  Fetch Operator
+    limit:-1
+    Select Operator [SEL_1]
+      Output:["_col0","_col1","_col2","_col3"]
+      TableScan [TS_0]
+        Output:["my_bigint","my_string","my_double"]
 
 PREHOOK: query: select `my_bigint`, `my_string`, `my_double`, 
struct(`my_bigint`, `my_string`, `my_double`)
 IN (

http://git-wip-us.apache.org/repos/asf/hive/blob/ad6ce078/ql/src/test/results/clientpositive/llap/vector_varchar_4.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_varchar_4.q.out 
b/ql/src/test/results/clientpositive/llap/vector_varchar_4.q.out
index 5979f8b..7d14256 100644
--- a/ql/src/test/results/clientpositive/llap/vector_varchar_4.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_varchar_4.q.out
@@ -121,16 +121,12 @@ POSTHOOK: query: create table 
varchar_lazy_binary_columnar(vt varchar(10), vsi v
 POSTHOOK: type: CREATETABLE
 POSTHOOK: Output: database:default
 POSTHOOK: Output: default@varchar_lazy_binary_columnar
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 insert overwrite table varchar_lazy_binary_columnar select t, si, i, b, f, d, 
s from vectortab2korc
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 insert overwrite table varchar_lazy_binary_columnar select t, si, i, b, f, d, 
s from vectortab2korc
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-2 depends on stages: Stage-1
@@ -147,23 +143,12 @@ STAGE PLANS:
                 TableScan
                   alias: vectortab2korc
                   Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 
10, 11, 12]
                   Select Operator
                     expressions: CAST( t AS varchar(10)) (type: varchar(10)), 
CAST( si AS varchar(10)) (type: varchar(10)), CAST( i AS varchar(20)) (type: 
varchar(20)), CAST( b AS varchar(30)) (type: varchar(30)), CAST( f AS 
varchar(20)) (type: varchar(20)), CAST( d AS varchar(20)) (type: varchar(20)), 
CAST( s AS varchar(50)) (type: varchar(50))
                     outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [13, 14, 15, 16, 17, 18, 19]
-                        selectExpressions: CastLongToVarChar(col 0, maxLength 
10) -> 13:VarChar, CastLongToVarChar(col 1, maxLength 10) -> 14:VarChar, 
CastLongToVarChar(col 2, maxLength 20) -> 15:VarChar, CastLongToVarChar(col 3, 
maxLength 30) -> 16:VarChar, VectorUDFAdaptor(CAST( f AS varchar(20))) -> 
17:varchar(20), VectorUDFAdaptor(CAST( d AS varchar(20))) -> 18:varchar(20), 
CastStringGroupToVarChar(col 8, maxLength 50) -> 19:VarChar
                     Statistics: Num rows: 2000 Data size: 918712 Basic stats: 
COMPLETE Column stats: NONE
                     File Output Operator
                       compressed: false
-                      File Sink Vectorization:
-                          className: VectorFileSinkOperator
-                          native: false
                       Statistics: Num rows: 2000 Data size: 918712 Basic 
stats: COMPLETE Column stats: NONE
                       table:
                           input format: 
org.apache.hadoop.hive.ql.io.RCFileInputFormat
@@ -172,14 +157,6 @@ STAGE PLANS:
                           name: default.varchar_lazy_binary_columnar
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: true
-                vectorized: true
 
   Stage: Stage-2
     Dependency Collection

http://git-wip-us.apache.org/repos/asf/hive/blob/ad6ce078/ql/src/test/results/clientpositive/llap/vector_varchar_mapjoin1.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_varchar_mapjoin1.q.out 
b/ql/src/test/results/clientpositive/llap/vector_varchar_mapjoin1.q.out
index 09259c8..9c2c536 100644
--- a/ql/src/test/results/clientpositive/llap/vector_varchar_mapjoin1.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_varchar_mapjoin1.q.out
@@ -125,15 +125,11 @@ POSTHOOK: Output: default@varchar_join1_str_orc
 POSTHOOK: Lineage: varchar_join1_str_orc.c1 SIMPLE 
[(varchar_join1_str)varchar_join1_str.FieldSchema(name:c1, type:int, 
comment:null), ]
 POSTHOOK: Lineage: varchar_join1_str_orc.c2 SIMPLE 
[(varchar_join1_str)varchar_join1_str.FieldSchema(name:c2, type:string, 
comment:null), ]
 PREHOOK: query: -- Join varchar with same length varchar
-explain vectorization select * from varchar_join1_vc1_orc a join 
varchar_join1_vc1_orc b on (a.c2 = b.c2) order by a.c1
+explain select * from varchar_join1_vc1_orc a join varchar_join1_vc1_orc b on 
(a.c2 = b.c2) order by a.c1
 PREHOOK: type: QUERY
 POSTHOOK: query: -- Join varchar with same length varchar
-explain vectorization select * from varchar_join1_vc1_orc a join 
varchar_join1_vc1_orc b on (a.c2 = b.c2) order by a.c1
+explain select * from varchar_join1_vc1_orc a join varchar_join1_vc1_orc b on 
(a.c2 = b.c2) order by a.c1
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -176,14 +172,6 @@ STAGE PLANS:
                           value expressions: _col1 (type: varchar(10)), _col2 
(type: int), _col3 (type: varchar(10))
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Map 3 
             Map Operator Tree:
                 TableScan
@@ -204,23 +192,8 @@ STAGE PLANS:
                         value expressions: _col0 (type: int)
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: true
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 
(type: varchar(10)), VALUE._col1 (type: int), VALUE._col2 (type: varchar(10))
@@ -252,15 +225,11 @@ POSTHOOK: Input: default@varchar_join1_vc1_orc
 2      abc     2       abc 
 3       abc    3        abc
 PREHOOK: query: -- Join varchar with different length varchar
-explain vectorization select * from varchar_join1_vc1_orc a join 
varchar_join1_vc2_orc b on (a.c2 = b.c2) order by a.c1
+explain select * from varchar_join1_vc1_orc a join varchar_join1_vc2_orc b on 
(a.c2 = b.c2) order by a.c1
 PREHOOK: type: QUERY
 POSTHOOK: query: -- Join varchar with different length varchar
-explain vectorization select * from varchar_join1_vc1_orc a join 
varchar_join1_vc2_orc b on (a.c2 = b.c2) order by a.c1
+explain select * from varchar_join1_vc1_orc a join varchar_join1_vc2_orc b on 
(a.c2 = b.c2) order by a.c1
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -303,14 +272,6 @@ STAGE PLANS:
                           value expressions: _col1 (type: varchar(10)), _col2 
(type: int), _col3 (type: varchar(20))
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Map 3 
             Map Operator Tree:
                 TableScan
@@ -331,23 +292,8 @@ STAGE PLANS:
                         value expressions: _col0 (type: int)
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: true
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 
(type: varchar(10)), VALUE._col1 (type: int), VALUE._col2 (type: varchar(20))
@@ -381,15 +327,11 @@ POSTHOOK: Input: default@varchar_join1_vc2_orc
 2      abc     2       abc 
 3       abc    3        abc
 PREHOOK: query: -- Join varchar with string
-explain vectorization select * from varchar_join1_vc1_orc a join 
varchar_join1_str_orc b on (a.c2 = b.c2) order by a.c1
+explain select * from varchar_join1_vc1_orc a join varchar_join1_str_orc b on 
(a.c2 = b.c2) order by a.c1
 PREHOOK: type: QUERY
 POSTHOOK: query: -- Join varchar with string
-explain vectorization select * from varchar_join1_vc1_orc a join 
varchar_join1_str_orc b on (a.c2 = b.c2) order by a.c1
+explain select * from varchar_join1_vc1_orc a join varchar_join1_str_orc b on 
(a.c2 = b.c2) order by a.c1
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -432,14 +374,6 @@ STAGE PLANS:
                           value expressions: _col1 (type: varchar(10)), _col2 
(type: int), _col3 (type: string)
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Map 3 
             Map Operator Tree:
                 TableScan
@@ -460,23 +394,8 @@ STAGE PLANS:
                         value expressions: _col0 (type: int)
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: true
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 
(type: varchar(10)), VALUE._col1 (type: int), VALUE._col2 (type: string)

Reply via email to