http://git-wip-us.apache.org/repos/asf/hive/blob/aed21d0b/ql/src/test/results/clientpositive/llap/vector_count_distinct.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_count_distinct.q.out 
b/ql/src/test/results/clientpositive/llap/vector_count_distinct.q.out
index 9e185c6..3d67664 100644
--- a/ql/src/test/results/clientpositive/llap/vector_count_distinct.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_count_distinct.q.out
@@ -1231,18 +1231,14 @@ POSTHOOK: Lineage: web_sales 
PARTITION(ws_web_site_sk=9).ws_web_page_sk SIMPLE [
 POSTHOOK: Lineage: web_sales PARTITION(ws_web_site_sk=9).ws_wholesale_cost 
SIMPLE [(web_sales_txt)web_sales_txt.FieldSchema(name:ws_wholesale_cost, 
type:decimal(7,2), comment:null), ]
 PREHOOK: query: 
------------------------------------------------------------------------------------------
 
-explain vectorization expression
+explain
 select count(distinct ws_order_number) from web_sales
 PREHOOK: type: QUERY
 POSTHOOK: query: 
------------------------------------------------------------------------------------------
 
-explain vectorization expression
+explain
 select count(distinct ws_order_number) from web_sales
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -1261,24 +1257,11 @@ STAGE PLANS:
                 TableScan
                   alias: web_sales
                   Statistics: Num rows: 2000 Data size: 3520000 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 
10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 
30, 31, 32, 33]
                   Select Operator
                     expressions: ws_order_number (type: int)
                     outputColumnNames: ws_order_number
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [16]
                     Statistics: Num rows: 2000 Data size: 3520000 Basic stats: 
COMPLETE Column stats: NONE
                     Group By Operator
-                      Group By Vectorization:
-                          className: VectorGroupByOperator
-                          vectorOutput: true
-                          keyExpressions: col 16
-                          native: false
-                          projectedOutputColumns: []
                       keys: ws_order_number (type: int)
                       mode: hash
                       outputColumnNames: _col0
@@ -1287,88 +1270,36 @@ STAGE PLANS:
                         key expressions: _col0 (type: int)
                         sort order: +
                         Map-reduce partition columns: _col0 (type: int)
-                        Reduce Sink Vectorization:
-                            className: VectorReduceSinkLongOperator
-                            native: true
-                            nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No TopN IS true, Uniform Hash IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
                         Statistics: Num rows: 2000 Data size: 3520000 Basic 
stats: COMPLETE Column stats: NONE
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Group By Operator
-                Group By Vectorization:
-                    className: VectorGroupByOperator
-                    vectorOutput: true
-                    keyExpressions: col 0
-                    native: false
-                    projectedOutputColumns: []
                 keys: KEY._col0 (type: int)
                 mode: mergepartial
                 outputColumnNames: _col0
                 Statistics: Num rows: 1000 Data size: 1760000 Basic stats: 
COMPLETE Column stats: NONE
                 Group By Operator
                   aggregations: count(_col0)
-                  Group By Vectorization:
-                      aggregators: VectorUDAFCount(col 0) -> bigint
-                      className: VectorGroupByOperator
-                      vectorOutput: true
-                      native: false
-                      projectedOutputColumns: [0]
                   mode: hash
                   outputColumnNames: _col0
                   Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
                   Reduce Output Operator
                     sort order: 
-                    Reduce Sink Vectorization:
-                        className: VectorReduceSinkOperator
-                        native: false
-                        nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for 
keys IS true, LazyBinarySerDe for values IS true
-                        nativeConditionsNotMet: Uniform Hash IS false
                     Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
                     value expressions: _col0 (type: bigint)
         Reducer 3 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Group By Operator
                 aggregations: count(VALUE._col0)
-                Group By Vectorization:
-                    aggregators: VectorUDAFCountMerge(col 0) -> bigint
-                    className: VectorGroupByOperator
-                    vectorOutput: true
-                    native: false
-                    projectedOutputColumns: [0]
                 mode: mergepartial
                 outputColumnNames: _col0
                 Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
                   Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/aed21d0b/ql/src/test/results/clientpositive/llap/vector_data_types.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_data_types.q.out 
b/ql/src/test/results/clientpositive/llap/vector_data_types.q.out
index fc82926..a7a74c3 100644
--- a/ql/src/test/results/clientpositive/llap/vector_data_types.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_data_types.q.out
@@ -97,14 +97,10 @@ POSTHOOK: Lineage: over1korc.s SIMPLE 
[(over1k)over1k.FieldSchema(name:s, type:s
 POSTHOOK: Lineage: over1korc.si SIMPLE [(over1k)over1k.FieldSchema(name:si, 
type:smallint, comment:null), ]
 POSTHOOK: Lineage: over1korc.t SIMPLE [(over1k)over1k.FieldSchema(name:t, 
type:tinyint, comment:null), ]
 POSTHOOK: Lineage: over1korc.ts SIMPLE [(over1k)over1k.FieldSchema(name:ts, 
type:timestamp, comment:null), ]
-PREHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT t, si, i, b, f, d, bo, 
s, ts, dec, bin FROM over1korc ORDER BY t, si, i LIMIT 20
+PREHOOK: query: EXPLAIN SELECT t, si, i, b, f, d, bo, s, ts, dec, bin FROM 
over1korc ORDER BY t, si, i LIMIT 20
 PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT t, si, i, b, f, d, 
bo, s, ts, dec, bin FROM over1korc ORDER BY t, si, i LIMIT 20
+POSTHOOK: query: EXPLAIN SELECT t, si, i, b, f, d, bo, s, ts, dec, bin FROM 
over1korc ORDER BY t, si, i LIMIT 20
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: false
-  enabledConditionsNotMet: [hive.vectorized.execution.enabled IS false]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -197,14 +193,10 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@over1korc
 #### A masked pattern was here ####
 -17045922556
-PREHOOK: query: EXPLAIN VECTORIZATION EXPRESSION select t, si, i, b, f, d, bo, 
s, ts, dec, bin FROM over1korc ORDER BY t, si, i LIMIT 20
+PREHOOK: query: EXPLAIN select t, si, i, b, f, d, bo, s, ts, dec, bin FROM 
over1korc ORDER BY t, si, i LIMIT 20
 PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN VECTORIZATION EXPRESSION select t, si, i, b, f, d, 
bo, s, ts, dec, bin FROM over1korc ORDER BY t, si, i LIMIT 20
+POSTHOOK: query: EXPLAIN select t, si, i, b, f, d, bo, s, ts, dec, bin FROM 
over1korc ORDER BY t, si, i LIMIT 20
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -222,67 +214,30 @@ STAGE PLANS:
                 TableScan
                   alias: over1korc
                   Statistics: Num rows: 1049 Data size: 311170 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 
10]
                   Select Operator
                     expressions: t (type: tinyint), si (type: smallint), i 
(type: int), b (type: bigint), f (type: float), d (type: double), bo (type: 
boolean), s (type: string), ts (type: timestamp), dec (type: decimal(4,2)), bin 
(type: binary)
                     outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7, _col8, _col9, _col10
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 
10]
                     Statistics: Num rows: 1049 Data size: 311170 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       key expressions: _col0 (type: tinyint), _col1 (type: 
smallint), _col2 (type: int)
                       sort order: +++
-                      Reduce Sink Vectorization:
-                          className: VectorReduceSinkOperator
-                          native: false
-                          nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, 
LazyBinarySerDe for values IS true
-                          nativeConditionsNotMet: No TopN IS false, Uniform 
Hash IS false
                       Statistics: Num rows: 1049 Data size: 311170 Basic 
stats: COMPLETE Column stats: NONE
                       TopN Hash Memory Usage: 0.1
                       value expressions: _col3 (type: bigint), _col4 (type: 
float), _col5 (type: double), _col6 (type: boolean), _col7 (type: string), 
_col8 (type: timestamp), _col9 (type: decimal(4,2)), _col10 (type: binary)
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: tinyint), 
KEY.reducesinkkey1 (type: smallint), KEY.reducesinkkey2 (type: int), 
VALUE._col0 (type: bigint), VALUE._col1 (type: float), VALUE._col2 (type: 
double), VALUE._col3 (type: boolean), VALUE._col4 (type: string), VALUE._col5 
(type: timestamp), VALUE._col6 (type: decimal(4,2)), VALUE._col7 (type: binary)
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10]
                 Statistics: Num rows: 1049 Data size: 311170 Basic stats: 
COMPLETE Column stats: NONE
                 Limit
                   Number of rows: 20
-                  Limit Vectorization:
-                      className: VectorLimitOperator
-                      native: true
                   Statistics: Num rows: 20 Data size: 5920 Basic stats: 
COMPLETE Column stats: NONE
                   File Output Operator
                     compressed: false
-                    File Sink Vectorization:
-                        className: VectorFileSinkOperator
-                        native: false
                     Statistics: Num rows: 20 Data size: 5920 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/aed21d0b/ql/src/test/results/clientpositive/llap/vector_decimal_aggregate.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_decimal_aggregate.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_aggregate.q.out
index 1a31067..6faf453 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_aggregate.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_aggregate.q.out
@@ -23,7 +23,7 @@ POSTHOOK: Lineage: decimal_vgby.cint SIMPLE 
[(alltypesorc)alltypesorc.FieldSchem
 PREHOOK: query: -- SORT_QUERY_RESULTS
 
 -- First only do simple aggregations that output primitives only
-EXPLAIN VECTORIZATION EXPRESSION SELECT cint,
+EXPLAIN SELECT cint,
     COUNT(cdecimal1), MAX(cdecimal1), MIN(cdecimal1), SUM(cdecimal1),
     COUNT(cdecimal2), MAX(cdecimal2), MIN(cdecimal2), SUM(cdecimal2)
     FROM decimal_vgby
@@ -33,17 +33,13 @@ PREHOOK: type: QUERY
 POSTHOOK: query: -- SORT_QUERY_RESULTS
 
 -- First only do simple aggregations that output primitives only
-EXPLAIN VECTORIZATION EXPRESSION SELECT cint,
+EXPLAIN SELECT cint,
     COUNT(cdecimal1), MAX(cdecimal1), MIN(cdecimal1), SUM(cdecimal1),
     COUNT(cdecimal2), MAX(cdecimal2), MIN(cdecimal2), SUM(cdecimal2)
     FROM decimal_vgby
     GROUP BY cint
     HAVING COUNT(*) > 1
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -61,26 +57,12 @@ STAGE PLANS:
                 TableScan
                   alias: decimal_vgby
                   Statistics: Num rows: 12288 Data size: 2165060 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3]
                   Select Operator
                     expressions: cint (type: int), cdecimal1 (type: 
decimal(20,10)), cdecimal2 (type: decimal(23,14))
                     outputColumnNames: cint, cdecimal1, cdecimal2
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [3, 1, 2]
                     Statistics: Num rows: 12288 Data size: 2165060 Basic 
stats: COMPLETE Column stats: NONE
                     Group By Operator
                       aggregations: count(cdecimal1), max(cdecimal1), 
min(cdecimal1), sum(cdecimal1), count(cdecimal2), max(cdecimal2), 
min(cdecimal2), sum(cdecimal2), count()
-                      Group By Vectorization:
-                          aggregators: VectorUDAFCount(col 1) -> bigint, 
VectorUDAFMaxDecimal(col 1) -> decimal(20,10), VectorUDAFMinDecimal(col 1) -> 
decimal(20,10), VectorUDAFSumDecimal(col 1) -> decimal(38,18), 
VectorUDAFCount(col 2) -> bigint, VectorUDAFMaxDecimal(col 2) -> 
decimal(23,14), VectorUDAFMinDecimal(col 2) -> decimal(23,14), 
VectorUDAFSumDecimal(col 2) -> decimal(38,18), VectorUDAFCountStar(*) -> bigint
-                          className: VectorGroupByOperator
-                          vectorOutput: true
-                          keyExpressions: col 3
-                          native: false
-                          projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8]
                       keys: cint (type: int)
                       mode: hash
                       outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7, _col8, _col9
@@ -89,65 +71,28 @@ STAGE PLANS:
                         key expressions: _col0 (type: int)
                         sort order: +
                         Map-reduce partition columns: _col0 (type: int)
-                        Reduce Sink Vectorization:
-                            className: VectorReduceSinkLongOperator
-                            native: true
-                            nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No TopN IS true, Uniform Hash IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
                         Statistics: Num rows: 12288 Data size: 2165060 Basic 
stats: COMPLETE Column stats: NONE
                         value expressions: _col1 (type: bigint), _col2 (type: 
decimal(20,10)), _col3 (type: decimal(20,10)), _col4 (type: decimal(30,10)), 
_col5 (type: bigint), _col6 (type: decimal(23,14)), _col7 (type: 
decimal(23,14)), _col8 (type: decimal(33,14)), _col9 (type: bigint)
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Group By Operator
                 aggregations: count(VALUE._col0), max(VALUE._col1), 
min(VALUE._col2), sum(VALUE._col3), count(VALUE._col4), max(VALUE._col5), 
min(VALUE._col6), sum(VALUE._col7), count(VALUE._col8)
-                Group By Vectorization:
-                    aggregators: VectorUDAFCountMerge(col 1) -> bigint, 
VectorUDAFMaxDecimal(col 2) -> decimal(20,10), VectorUDAFMinDecimal(col 3) -> 
decimal(20,10), VectorUDAFSumDecimal(col 4) -> decimal(38,18), 
VectorUDAFCountMerge(col 5) -> bigint, VectorUDAFMaxDecimal(col 6) -> 
decimal(23,14), VectorUDAFMinDecimal(col 7) -> decimal(23,14), 
VectorUDAFSumDecimal(col 8) -> decimal(38,18), VectorUDAFCountMerge(col 9) -> 
bigint
-                    className: VectorGroupByOperator
-                    vectorOutput: true
-                    keyExpressions: col 0
-                    native: false
-                    projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8]
                 keys: KEY._col0 (type: int)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9
                 Statistics: Num rows: 6144 Data size: 1082530 Basic stats: 
COMPLETE Column stats: NONE
                 Filter Operator
-                  Filter Vectorization:
-                      className: VectorFilterOperator
-                      native: true
-                      predicateExpression: FilterLongColGreaterLongScalar(col 
9, val 1) -> boolean
                   predicate: (_col9 > 1) (type: boolean)
                   Statistics: Num rows: 2048 Data size: 360843 Basic stats: 
COMPLETE Column stats: NONE
                   Select Operator
                     expressions: _col0 (type: int), _col1 (type: bigint), 
_col2 (type: decimal(20,10)), _col3 (type: decimal(20,10)), _col4 (type: 
decimal(30,10)), _col5 (type: bigint), _col6 (type: decimal(23,14)), _col7 
(type: decimal(23,14)), _col8 (type: decimal(33,14))
                     outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7, _col8
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8]
                     Statistics: Num rows: 2048 Data size: 360843 Basic stats: 
COMPLETE Column stats: NONE
                     File Output Operator
                       compressed: false
-                      File Sink Vectorization:
-                          className: VectorFileSinkOperator
-                          native: false
                       Statistics: Num rows: 2048 Data size: 360843 Basic 
stats: COMPLETE Column stats: NONE
                       table:
                           input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -187,7 +132,7 @@ POSTHOOK: Input: default@decimal_vgby
 762    2       5831542.2692483780      1531.2194054054 5833073.4886537834      
2       6984454.21109769200000  1833.94569230769250     6986288.15678999969250
 NULL   3072    9318.4351351351 -4298.1513513514        5018444.1081079808      
3072    11160.71538461538500    -5147.90769230769300    6010604.30769230735360
 PREHOOK: query: -- Now add the others...
-EXPLAIN VECTORIZATION EXPRESSION SELECT cint,
+EXPLAIN SELECT cint,
     COUNT(cdecimal1), MAX(cdecimal1), MIN(cdecimal1), SUM(cdecimal1), 
AVG(cdecimal1), STDDEV_POP(cdecimal1), STDDEV_SAMP(cdecimal1),
     COUNT(cdecimal2), MAX(cdecimal2), MIN(cdecimal2), SUM(cdecimal2), 
AVG(cdecimal2), STDDEV_POP(cdecimal2), STDDEV_SAMP(cdecimal2)
     FROM decimal_vgby
@@ -195,17 +140,13 @@ EXPLAIN VECTORIZATION EXPRESSION SELECT cint,
     HAVING COUNT(*) > 1
 PREHOOK: type: QUERY
 POSTHOOK: query: -- Now add the others...
-EXPLAIN VECTORIZATION EXPRESSION SELECT cint,
+EXPLAIN SELECT cint,
     COUNT(cdecimal1), MAX(cdecimal1), MIN(cdecimal1), SUM(cdecimal1), 
AVG(cdecimal1), STDDEV_POP(cdecimal1), STDDEV_SAMP(cdecimal1),
     COUNT(cdecimal2), MAX(cdecimal2), MIN(cdecimal2), SUM(cdecimal2), 
AVG(cdecimal2), STDDEV_POP(cdecimal2), STDDEV_SAMP(cdecimal2)
     FROM decimal_vgby
     GROUP BY cint
     HAVING COUNT(*) > 1
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -223,27 +164,12 @@ STAGE PLANS:
                 TableScan
                   alias: decimal_vgby
                   Statistics: Num rows: 12288 Data size: 2165060 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3]
                   Select Operator
                     expressions: cint (type: int), cdecimal1 (type: 
decimal(20,10)), cdecimal2 (type: decimal(23,14))
                     outputColumnNames: cint, cdecimal1, cdecimal2
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [3, 1, 2]
                     Statistics: Num rows: 12288 Data size: 2165060 Basic 
stats: COMPLETE Column stats: NONE
                     Group By Operator
                       aggregations: count(cdecimal1), max(cdecimal1), 
min(cdecimal1), sum(cdecimal1), avg(cdecimal1), stddev_pop(cdecimal1), 
stddev_samp(cdecimal1), count(cdecimal2), max(cdecimal2), min(cdecimal2), 
sum(cdecimal2), avg(cdecimal2), stddev_pop(cdecimal2), stddev_samp(cdecimal2), 
count()
-                      Group By Vectorization:
-                          aggregators: VectorUDAFCount(col 1) -> bigint, 
VectorUDAFMaxDecimal(col 1) -> decimal(20,10), VectorUDAFMinDecimal(col 1) -> 
decimal(20,10), VectorUDAFSumDecimal(col 1) -> decimal(38,18), 
VectorUDAFAvgDecimal(col 1) -> struct<count:bigint,sum:decimal(30,10)>, 
VectorUDAFStdPopDecimal(col 1) -> 
struct<count:bigint,sum:double,variance:double>, VectorUDAFStdSampDecimal(col 
1) -> struct<count:bigint,sum:double,variance:double>, VectorUDAFCount(col 2) 
-> bigint, VectorUDAFMaxDecimal(col 2) -> decimal(23,14), 
VectorUDAFMinDecimal(col 2) -> decimal(23,14), VectorUDAFSumDecimal(col 2) -> 
decimal(38,18), VectorUDAFAvgDecimal(col 2) -> 
struct<count:bigint,sum:decimal(33,14)>, VectorUDAFStdPopDecimal(col 2) -> 
struct<count:bigint,sum:double,variance:double>, VectorUDAFStdSampDecimal(col 
2) -> struct<count:bigint,sum:double,variance:double>, VectorUDAFCountStar(*) 
-> bigint
-                          className: VectorGroupByOperator
-                          vectorOutput: false
-                          keyExpressions: col 3
-                          native: false
-                          projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 
9, 10, 11, 12, 13, 14]
-                          vectorOutputConditionsNotMet: Vector output of 
VectorUDAFAvgDecimal(col 1) -> struct<count:bigint,sum:decimal(30,10)> output 
type STRUCT requires PRIMITIVE IS false, Vector output of 
VectorUDAFStdPopDecimal(col 1) -> 
struct<count:bigint,sum:double,variance:double> output type STRUCT requires 
PRIMITIVE IS false, Vector output of VectorUDAFStdSampDecimal(col 1) -> 
struct<count:bigint,sum:double,variance:double> output type STRUCT requires 
PRIMITIVE IS false, Vector output of VectorUDAFAvgDecimal(col 2) -> 
struct<count:bigint,sum:decimal(33,14)> output type STRUCT requires PRIMITIVE 
IS false, Vector output of VectorUDAFStdPopDecimal(col 2) -> 
struct<count:bigint,sum:double,variance:double> output type STRUCT requires 
PRIMITIVE IS false, Vector output of VectorUDAFStdSampDecimal(col 2) -> 
struct<count:bigint,sum:double,variance:double> output type STRUCT requires 
PRIMITIVE IS false
                       keys: cint (type: int)
                       mode: hash
                       outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, 
_col15
@@ -256,21 +182,8 @@ STAGE PLANS:
                         value expressions: _col1 (type: bigint), _col2 (type: 
decimal(20,10)), _col3 (type: decimal(20,10)), _col4 (type: decimal(30,10)), 
_col5 (type: struct<count:bigint,sum:decimal(30,10),input:decimal(20,10)>), 
_col6 (type: struct<count:bigint,sum:double,variance:double>), _col7 (type: 
struct<count:bigint,sum:double,variance:double>), _col8 (type: bigint), _col9 
(type: decimal(23,14)), _col10 (type: decimal(23,14)), _col11 (type: 
decimal(33,14)), _col12 (type: 
struct<count:bigint,sum:decimal(33,14),input:decimal(23,14)>), _col13 (type: 
struct<count:bigint,sum:double,variance:double>), _col14 (type: 
struct<count:bigint,sum:double,variance:double>), _col15 (type: bigint)
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: false
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                notVectorizedReason: Aggregation Function UDF avg parameter 
expression for GROUPBY operator: Data type 
struct<count:bigint,sum:decimal(30,10),input:decimal(20,10)> of 
Column[VALUE._col4] not supported
-                vectorized: false
             Reduce Operator Tree:
               Group By Operator
                 aggregations: count(VALUE._col0), max(VALUE._col1), 
min(VALUE._col2), sum(VALUE._col3), avg(VALUE._col4), stddev_pop(VALUE._col5), 
stddev_samp(VALUE._col6), count(VALUE._col7), max(VALUE._col8), 
min(VALUE._col9), sum(VALUE._col10), avg(VALUE._col11), 
stddev_pop(VALUE._col12), stddev_samp(VALUE._col13), count(VALUE._col14)

http://git-wip-us.apache.org/repos/asf/hive/blob/aed21d0b/ql/src/test/results/clientpositive/llap/vector_decimal_cast.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_decimal_cast.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_cast.q.out
index 5cce027..16d9929 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_cast.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_cast.q.out
@@ -1,76 +1,25 @@
-PREHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT cdouble, cint, 
cboolean1, ctimestamp1, CAST(cdouble AS DECIMAL(20,10)), CAST(cint AS 
DECIMAL(23,14)), CAST(cboolean1 AS DECIMAL(5,2)), CAST(ctimestamp1 AS 
DECIMAL(15,0)) FROM alltypesorc WHERE cdouble IS NOT NULL AND cint IS NOT NULL 
AND cboolean1 IS NOT NULL AND ctimestamp1 IS NOT NULL LIMIT 10
+PREHOOK: query: EXPLAIN SELECT cdouble, cint, cboolean1, ctimestamp1, 
CAST(cdouble AS DECIMAL(20,10)), CAST(cint AS DECIMAL(23,14)), CAST(cboolean1 
AS DECIMAL(5,2)), CAST(ctimestamp1 AS DECIMAL(15,0)) FROM alltypesorc WHERE 
cdouble IS NOT NULL AND cint IS NOT NULL AND cboolean1 IS NOT NULL AND 
ctimestamp1 IS NOT NULL LIMIT 10
 PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT cdouble, cint, 
cboolean1, ctimestamp1, CAST(cdouble AS DECIMAL(20,10)), CAST(cint AS 
DECIMAL(23,14)), CAST(cboolean1 AS DECIMAL(5,2)), CAST(ctimestamp1 AS 
DECIMAL(15,0)) FROM alltypesorc WHERE cdouble IS NOT NULL AND cint IS NOT NULL 
AND cboolean1 IS NOT NULL AND ctimestamp1 IS NOT NULL LIMIT 10
+POSTHOOK: query: EXPLAIN SELECT cdouble, cint, cboolean1, ctimestamp1, 
CAST(cdouble AS DECIMAL(20,10)), CAST(cint AS DECIMAL(23,14)), CAST(cboolean1 
AS DECIMAL(5,2)), CAST(ctimestamp1 AS DECIMAL(15,0)) FROM alltypesorc WHERE 
cdouble IS NOT NULL AND cint IS NOT NULL AND cboolean1 IS NOT NULL AND 
ctimestamp1 IS NOT NULL LIMIT 10
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
+  Stage-0 is a root stage
 
 STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: alltypesorc
-                  Statistics: Num rows: 12288 Data size: 638316 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 
10, 11]
-                  Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: FilterExprAndExpr(children: 
SelectColumnIsNotNull(col 5) -> boolean, SelectColumnIsNotNull(col 2) -> 
boolean, SelectColumnIsNotNull(col 10) -> boolean, SelectColumnIsNotNull(col 8) 
-> boolean) -> boolean
-                    predicate: (cdouble is not null and cint is not null and 
cboolean1 is not null and ctimestamp1 is not null) (type: boolean)
-                    Statistics: Num rows: 5112 Data size: 265564 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    Select Operator
-                      expressions: cdouble (type: double), cint (type: int), 
cboolean1 (type: boolean), ctimestamp1 (type: timestamp), CAST( cdouble AS 
decimal(20,10)) (type: decimal(20,10)), CAST( cint AS decimal(23,14)) (type: 
decimal(23,14)), CAST( cboolean1 AS decimal(5,2)) (type: decimal(5,2)), CAST( 
ctimestamp1 AS decimal(15,0)) (type: decimal(15,0))
-                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7
-                      Select Vectorization:
-                          className: VectorSelectOperator
-                          native: true
-                          projectedOutputColumns: [5, 2, 10, 8, 12, 13, 14, 15]
-                          selectExpressions: CastDoubleToDecimal(col 5) -> 
12:decimal(20,10), CastLongToDecimal(col 2) -> 13:decimal(23,14), 
CastLongToDecimal(col 10) -> 14:decimal(5,2), CastTimestampToDecimal(col 8) -> 
15:decimal(15,0)
-                      Statistics: Num rows: 5112 Data size: 2555740 Basic 
stats: COMPLETE Column stats: COMPLETE
-                      Limit
-                        Number of rows: 10
-                        Limit Vectorization:
-                            className: VectorLimitOperator
-                            native: true
-                        Statistics: Num rows: 10 Data size: 5008 Basic stats: 
COMPLETE Column stats: COMPLETE
-                        File Output Operator
-                          compressed: false
-                          File Sink Vectorization:
-                              className: VectorFileSinkOperator
-                              native: false
-                          Statistics: Num rows: 10 Data size: 5008 Basic 
stats: COMPLETE Column stats: COMPLETE
-                          table:
-                              input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                              output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                              serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-
   Stage: Stage-0
     Fetch Operator
       limit: 10
       Processor Tree:
-        ListSink
+        TableScan
+          alias: alltypesorc
+          Filter Operator
+            predicate: (cdouble is not null and cint is not null and cboolean1 
is not null and ctimestamp1 is not null) (type: boolean)
+            Select Operator
+              expressions: cdouble (type: double), cint (type: int), cboolean1 
(type: boolean), ctimestamp1 (type: timestamp), CAST( cdouble AS 
decimal(20,10)) (type: decimal(20,10)), CAST( cint AS decimal(23,14)) (type: 
decimal(23,14)), CAST( cboolean1 AS decimal(5,2)) (type: decimal(5,2)), CAST( 
ctimestamp1 AS decimal(15,0)) (type: decimal(15,0))
+              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7
+              Limit
+                Number of rows: 10
+                ListSink
 
 PREHOOK: query: SELECT cdouble, cint, cboolean1, ctimestamp1, CAST(cdouble AS 
DECIMAL(20,10)), CAST(cint AS DECIMAL(23,14)), CAST(cboolean1 AS DECIMAL(5,2)), 
CAST(ctimestamp1 AS DECIMAL(15,0)) FROM alltypesorc WHERE cdouble IS NOT NULL 
AND cint IS NOT NULL AND cboolean1 IS NOT NULL AND ctimestamp1 IS NOT NULL 
LIMIT 10
 PREHOOK: type: QUERY

http://git-wip-us.apache.org/repos/asf/hive/blob/aed21d0b/ql/src/test/results/clientpositive/llap/vector_decimal_expressions.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_decimal_expressions.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_expressions.q.out
index 16232c2..d37b973 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_expressions.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_expressions.q.out
@@ -15,18 +15,14 @@ POSTHOOK: Output: default@decimal_test
 POSTHOOK: Lineage: decimal_test.cdecimal1 EXPRESSION 
[(alltypesorc)alltypesorc.FieldSchema(name:cdouble, type:double, comment:null), 
]
 POSTHOOK: Lineage: decimal_test.cdecimal2 EXPRESSION 
[(alltypesorc)alltypesorc.FieldSchema(name:cdouble, type:double, comment:null), 
]
 POSTHOOK: Lineage: decimal_test.cdouble SIMPLE 
[(alltypesorc)alltypesorc.FieldSchema(name:cdouble, type:double, comment:null), 
]
-PREHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT cdecimal1 + cdecimal2 
as c1, cdecimal1 - (2*cdecimal2) as c2, ((cdecimal1+2.34)/cdecimal2) as c3, 
(cdecimal1 * (cdecimal2/3.4)) as c4, cdecimal1 % 10 as c5, CAST(cdecimal1 AS 
INT) as c6, CAST(cdecimal2 AS SMALLINT) as c7, CAST(cdecimal2 AS TINYINT) as 
c8, CAST(cdecimal1 AS BIGINT) as c9, CAST (cdecimal1 AS BOOLEAN) as c10, 
CAST(cdecimal2 AS DOUBLE) as c11, CAST(cdecimal1 AS FLOAT) as c12, 
CAST(cdecimal2 AS STRING) as c13, CAST(cdecimal1 AS TIMESTAMP) as c14 FROM 
decimal_test WHERE cdecimal1 > 0 AND cdecimal1 < 12345.5678 AND cdecimal2 != 0 
AND cdecimal2 > 1000 AND cdouble IS NOT NULL
+PREHOOK: query: EXPLAIN SELECT cdecimal1 + cdecimal2 as c1, cdecimal1 - 
(2*cdecimal2) as c2, ((cdecimal1+2.34)/cdecimal2) as c3, (cdecimal1 * 
(cdecimal2/3.4)) as c4, cdecimal1 % 10 as c5, CAST(cdecimal1 AS INT) as c6, 
CAST(cdecimal2 AS SMALLINT) as c7, CAST(cdecimal2 AS TINYINT) as c8, 
CAST(cdecimal1 AS BIGINT) as c9, CAST (cdecimal1 AS BOOLEAN) as c10, 
CAST(cdecimal2 AS DOUBLE) as c11, CAST(cdecimal1 AS FLOAT) as c12, 
CAST(cdecimal2 AS STRING) as c13, CAST(cdecimal1 AS TIMESTAMP) as c14 FROM 
decimal_test WHERE cdecimal1 > 0 AND cdecimal1 < 12345.5678 AND cdecimal2 != 0 
AND cdecimal2 > 1000 AND cdouble IS NOT NULL
 ORDER BY c1, c2, c3, c4, c5, c6, c7, c8, c9, c10, c11, c12, c13, c14
 LIMIT 10
 PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT cdecimal1 + cdecimal2 
as c1, cdecimal1 - (2*cdecimal2) as c2, ((cdecimal1+2.34)/cdecimal2) as c3, 
(cdecimal1 * (cdecimal2/3.4)) as c4, cdecimal1 % 10 as c5, CAST(cdecimal1 AS 
INT) as c6, CAST(cdecimal2 AS SMALLINT) as c7, CAST(cdecimal2 AS TINYINT) as 
c8, CAST(cdecimal1 AS BIGINT) as c9, CAST (cdecimal1 AS BOOLEAN) as c10, 
CAST(cdecimal2 AS DOUBLE) as c11, CAST(cdecimal1 AS FLOAT) as c12, 
CAST(cdecimal2 AS STRING) as c13, CAST(cdecimal1 AS TIMESTAMP) as c14 FROM 
decimal_test WHERE cdecimal1 > 0 AND cdecimal1 < 12345.5678 AND cdecimal2 != 0 
AND cdecimal2 > 1000 AND cdouble IS NOT NULL
+POSTHOOK: query: EXPLAIN SELECT cdecimal1 + cdecimal2 as c1, cdecimal1 - 
(2*cdecimal2) as c2, ((cdecimal1+2.34)/cdecimal2) as c3, (cdecimal1 * 
(cdecimal2/3.4)) as c4, cdecimal1 % 10 as c5, CAST(cdecimal1 AS INT) as c6, 
CAST(cdecimal2 AS SMALLINT) as c7, CAST(cdecimal2 AS TINYINT) as c8, 
CAST(cdecimal1 AS BIGINT) as c9, CAST (cdecimal1 AS BOOLEAN) as c10, 
CAST(cdecimal2 AS DOUBLE) as c11, CAST(cdecimal1 AS FLOAT) as c12, 
CAST(cdecimal2 AS STRING) as c13, CAST(cdecimal1 AS TIMESTAMP) as c14 FROM 
decimal_test WHERE cdecimal1 > 0 AND cdecimal1 < 12345.5678 AND cdecimal2 != 0 
AND cdecimal2 > 1000 AND cdouble IS NOT NULL
 ORDER BY c1, c2, c3, c4, c5, c6, c7, c8, c9, c10, c11, c12, c13, c14
 LIMIT 10
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -44,74 +40,32 @@ STAGE PLANS:
                 TableScan
                   alias: decimal_test
                   Statistics: Num rows: 12288 Data size: 2128368 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2]
                   Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: FilterExprAndExpr(children: 
FilterDecimalColGreaterDecimalScalar(col 1, val 0) -> boolean, 
FilterDecimalColLessDecimalScalar(col 1, val 12345.5678) -> boolean, 
FilterDecimalColNotEqualDecimalScalar(col 2, val 0) -> boolean, 
FilterDecimalColGreaterDecimalScalar(col 2, val 1000) -> boolean, 
SelectColumnIsNotNull(col 0) -> boolean) -> boolean
                     predicate: ((cdecimal1 > 0) and (cdecimal1 < 12345.5678) 
and (cdecimal2 <> 0) and (cdecimal2 > 1000) and cdouble is not null) (type: 
boolean)
                     Statistics: Num rows: 455 Data size: 78809 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
                       expressions: (cdecimal1 + cdecimal2) (type: 
decimal(25,14)), (cdecimal1 - (2 * cdecimal2)) (type: decimal(26,14)), 
((cdecimal1 + 2.34) / cdecimal2) (type: decimal(38,23)), (cdecimal1 * 
(cdecimal2 / 3.4)) (type: decimal(38,28)), (cdecimal1 % 10) (type: 
decimal(12,10)), UDFToInteger(cdecimal1) (type: int), UDFToShort(cdecimal2) 
(type: smallint), UDFToByte(cdecimal2) (type: tinyint), UDFToLong(cdecimal1) 
(type: bigint), UDFToBoolean(cdecimal1) (type: boolean), UDFToDouble(cdecimal2) 
(type: double), UDFToFloat(cdecimal1) (type: float), UDFToString(cdecimal2) 
(type: string), CAST( cdecimal1 AS TIMESTAMP) (type: timestamp)
                       outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13
-                      Select Vectorization:
-                          className: VectorSelectOperator
-                          native: true
-                          projectedOutputColumns: [3, 5, 7, 9, 10, 11, 12, 13, 
14, 15, 16, 17, 18, 19]
-                          selectExpressions: DecimalColAddDecimalColumn(col 1, 
col 2) -> 3:decimal(25,14), DecimalColSubtractDecimalColumn(col 1, col 
4)(children: DecimalScalarMultiplyDecimalColumn(val 2, col 2) -> 
4:decimal(25,14)) -> 5:decimal(26,14), DecimalColDivideDecimalColumn(col 6, col 
2)(children: DecimalColAddDecimalScalar(col 1, val 2.34) -> 6:decimal(21,10)) 
-> 7:decimal(38,23), DecimalColMultiplyDecimalColumn(col 1, col 8)(children: 
DecimalColDivideDecimalScalar(col 2, val 3.4) -> 8:decimal(28,18)) -> 
9:decimal(38,28), DecimalColModuloDecimalScalar(col 1, val 10) -> 
10:decimal(12,10), CastDecimalToLong(col 1) -> 11:long, CastDecimalToLong(col 
2) -> 12:long, CastDecimalToLong(col 2) -> 13:long, CastDecimalToLong(col 1) -> 
14:long, CastDecimalToBoolean(col 1) -> 15:long, CastDecimalToDouble(col 2) -> 
16:double, CastDecimalToDouble(col 1) -> 17:double, CastDecimalToString(col 2) 
-> 18:String, CastDecimalToTimestamp(col 1) -> 19:timestamp
                       Statistics: Num rows: 455 Data size: 78809 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
                         key expressions: _col0 (type: decimal(25,14)), _col1 
(type: decimal(26,14)), _col2 (type: decimal(38,23)), _col3 (type: 
decimal(38,28)), _col4 (type: decimal(12,10)), _col5 (type: int), _col6 (type: 
smallint), _col7 (type: tinyint), _col8 (type: bigint), _col9 (type: boolean), 
_col10 (type: double), _col11 (type: float), _col12 (type: string), _col13 
(type: timestamp)
                         sort order: ++++++++++++++
-                        Reduce Sink Vectorization:
-                            className: VectorReduceSinkOperator
-                            native: false
-                            nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, 
LazyBinarySerDe for values IS true
-                            nativeConditionsNotMet: No TopN IS false, Uniform 
Hash IS false
                         Statistics: Num rows: 455 Data size: 78809 Basic 
stats: COMPLETE Column stats: NONE
                         TopN Hash Memory Usage: 0.1
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: decimal(25,14)), 
KEY.reducesinkkey1 (type: decimal(26,14)), KEY.reducesinkkey2 (type: 
decimal(38,23)), KEY.reducesinkkey3 (type: decimal(38,28)), KEY.reducesinkkey4 
(type: decimal(12,10)), KEY.reducesinkkey5 (type: int), KEY.reducesinkkey6 
(type: smallint), KEY.reducesinkkey7 (type: tinyint), KEY.reducesinkkey8 (type: 
bigint), KEY.reducesinkkey9 (type: boolean), KEY.reducesinkkey10 (type: 
double), KEY.reducesinkkey11 (type: float), KEY.reducesinkkey12 (type: string), 
KEY.reducesinkkey13 (type: timestamp)
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 
11, 12, 13]
                 Statistics: Num rows: 455 Data size: 78809 Basic stats: 
COMPLETE Column stats: NONE
                 Limit
                   Number of rows: 10
-                  Limit Vectorization:
-                      className: VectorLimitOperator
-                      native: true
                   Statistics: Num rows: 10 Data size: 1730 Basic stats: 
COMPLETE Column stats: NONE
                   File Output Operator
                     compressed: false
-                    File Sink Vectorization:
-                        className: VectorFileSinkOperator
-                        native: false
                     Statistics: Num rows: 10 Data size: 1730 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/aed21d0b/ql/src/test/results/clientpositive/llap/vector_decimal_mapjoin.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_decimal_mapjoin.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_mapjoin.q.out
index 6666923..361c46b 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_mapjoin.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_mapjoin.q.out
@@ -72,16 +72,12 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@over1k
 POSTHOOK: Output: default@t2
 POSTHOOK: Lineage: t2.dec EXPRESSION [(over1k)over1k.FieldSchema(name:dec, 
type:decimal(4,2), comment:null), ]
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select t1.dec, t2.dec from t1 join t2 on (t1.dec=t2.dec)
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select t1.dec, t2.dec from t1 join t2 on (t1.dec=t2.dec)
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -99,23 +95,12 @@ STAGE PLANS:
                 TableScan
                   alias: t1
                   Statistics: Num rows: 1049 Data size: 117488 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0]
                   Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: SelectColumnIsNotNull(col 
1)(children: CastDecimalToBoolean(col 0) -> 1:long) -> boolean
                     predicate: dec is not null (type: boolean)
                     Statistics: Num rows: 1049 Data size: 117488 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
                       expressions: dec (type: decimal(4,2))
                       outputColumnNames: _col0
-                      Select Vectorization:
-                          className: VectorSelectOperator
-                          native: true
-                          projectedOutputColumns: [0]
                       Statistics: Num rows: 1049 Data size: 117488 Basic 
stats: COMPLETE Column stats: NONE
                       Map Join Operator
                         condition map:
@@ -123,21 +108,12 @@ STAGE PLANS:
                         keys:
                           0 _col0 (type: decimal(6,2))
                           1 _col0 (type: decimal(6,2))
-                        Map Join Vectorization:
-                            className: VectorMapJoinOperator
-                            native: false
-                            nativeConditionsMet: 
hive.vectorized.execution.mapjoin.native.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, One MapJoin Condition IS true, No nullsafe IS 
true, Not empty key IS true, When Fast Hash Table, then requires no Hybrid Hash 
Join IS true, Small table vectorizes IS true
-                            nativeConditionsNotMet: Supports Key Types IS false
-                            nativeNotSupportedKeyTypes: DECIMAL
                         outputColumnNames: _col0, _col1
                         input vertices:
                           1 Map 2
                         Statistics: Num rows: 1153 Data size: 129236 Basic 
stats: COMPLETE Column stats: NONE
                         File Output Operator
                           compressed: false
-                          File Sink Vectorization:
-                              className: VectorFileSinkOperator
-                              native: false
                           Statistics: Num rows: 1153 Data size: 129236 Basic 
stats: COMPLETE Column stats: NONE
                           table:
                               input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -145,56 +121,25 @@ STAGE PLANS:
                               serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Map 2 
             Map Operator Tree:
                 TableScan
                   alias: t2
                   Statistics: Num rows: 1049 Data size: 117488 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0]
                   Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: SelectColumnIsNotNull(col 
1)(children: CastDecimalToBoolean(col 0) -> 1:long) -> boolean
                     predicate: dec is not null (type: boolean)
                     Statistics: Num rows: 1049 Data size: 117488 Basic stats: 
COMPLETE Column stats: NONE
                     Select Operator
                       expressions: dec (type: decimal(4,0))
                       outputColumnNames: _col0
-                      Select Vectorization:
-                          className: VectorSelectOperator
-                          native: true
-                          projectedOutputColumns: [0]
                       Statistics: Num rows: 1049 Data size: 117488 Basic 
stats: COMPLETE Column stats: NONE
                       Reduce Output Operator
                         key expressions: _col0 (type: decimal(6,2))
                         sort order: +
                         Map-reduce partition columns: _col0 (type: 
decimal(6,2))
-                        Reduce Sink Vectorization:
-                            className: VectorReduceSinkMultiKeyOperator
-                            native: true
-                            nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No TopN IS true, Uniform Hash IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
                         Statistics: Num rows: 1049 Data size: 117488 Basic 
stats: COMPLETE Column stats: NONE
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: true
-                usesVectorUDFAdaptor: false
-                vectorized: true
 
   Stage: Stage-0
     Fetch Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/aed21d0b/ql/src/test/results/clientpositive/llap/vector_decimal_math_funcs.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_decimal_math_funcs.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_math_funcs.q.out
index 4a2ae48..6d5b578 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_math_funcs.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_math_funcs.q.out
@@ -14,7 +14,7 @@ POSTHOOK: Lineage: decimal_test.cdecimal2 EXPRESSION 
[(alltypesorc)alltypesorc.F
 POSTHOOK: Lineage: decimal_test.cdouble SIMPLE 
[(alltypesorc)alltypesorc.FieldSchema(name:cdouble, type:double, comment:null), 
]
 PREHOOK: query: -- Test math functions in vectorized mode to verify they run 
correctly end-to-end.
 
-explain vectorization expression 
+explain 
 select
    cdecimal1
   ,Round(cdecimal1, 2)
@@ -53,7 +53,7 @@ and sin(cdecimal1) >= -1.0
 PREHOOK: type: QUERY
 POSTHOOK: query: -- Test math functions in vectorized mode to verify they run 
correctly end-to-end.
 
-explain vectorization expression 
+explain 
 select
    cdecimal1
   ,Round(cdecimal1, 2)
@@ -90,69 +90,22 @@ where cbigint % 500 = 0
 -- test use of a math function in the WHERE clause
 and sin(cdecimal1) >= -1.0
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
+  Stage-0 is a root stage
 
 STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: decimal_test
-                  Statistics: Num rows: 12288 Data size: 2201752 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0, 1, 2, 3]
-                  Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: FilterExprAndExpr(children: 
FilterLongColEqualLongScalar(col 4, val 0)(children: 
LongColModuloLongScalar(col 0, val 500) -> 4:long) -> boolean, 
FilterDoubleColGreaterEqualDoubleScalar(col 6, val -1.0)(children: 
FuncSinDoubleToDouble(col 5)(children: CastDecimalToDouble(col 2) -> 5:double) 
-> 6:double) -> boolean) -> boolean
-                    predicate: (((cbigint % 500) = 0) and (sin(cdecimal1) >= 
-1.0)) (type: boolean)
-                    Statistics: Num rows: 2048 Data size: 366958 Basic stats: 
COMPLETE Column stats: NONE
-                    Select Operator
-                      expressions: cdecimal1 (type: decimal(20,10)), 
round(cdecimal1, 2) (type: decimal(13,2)), round(cdecimal1) (type: 
decimal(11,0)), floor(cdecimal1) (type: decimal(11,0)), ceil(cdecimal1) (type: 
decimal(11,0)), round(exp(cdecimal1), 58) (type: double), ln(cdecimal1) (type: 
double), log10(cdecimal1) (type: double), log2(cdecimal1) (type: double), 
log2((cdecimal1 - 15601)) (type: double), log(2, cdecimal1) (type: double), 
power(log2(cdecimal1), 2) (type: double), power(log2(cdecimal1), 2) (type: 
double), sqrt(cdecimal1) (type: double), abs(cdecimal1) (type: decimal(20,10)), 
sin(cdecimal1) (type: double), asin(cdecimal1) (type: double), cos(cdecimal1) 
(type: double), acos(cdecimal1) (type: double), atan(cdecimal1) (type: double), 
degrees(cdecimal1) (type: double), radians(cdecimal1) (type: double), cdecimal1 
(type: decimal(20,10)), (- cdecimal1) (type: decimal(20,10)), sign(cdecimal1) 
(type: int), cos(((- sin(log(cdecimal1))) + 3.14159)) (type: double)
-                      outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, 
_col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, 
_col25
-                      Select Vectorization:
-                          className: VectorSelectOperator
-                          native: true
-                          projectedOutputColumns: [2, 7, 8, 9, 10, 5, 11, 12, 
13, 15, 6, 16, 17, 19, 20, 21, 22, 23, 24, 25, 26, 27, 2, 28, 4, 29]
-                          selectExpressions: 
FuncRoundWithNumDigitsDecimalToDecimal(col 2, decimalPlaces 2) -> 
7:decimal(13,2), FuncRoundDecimalToDecimal(col 2) -> 8:decimal(11,0), 
FuncFloorDecimalToDecimal(col 2) -> 9:decimal(11,0), 
FuncCeilDecimalToDecimal(col 2) -> 10:decimal(11,0), 
RoundWithNumDigitsDoubleToDouble(col 6, decimalPlaces 58)(children: 
FuncExpDoubleToDouble(col 5)(children: CastDecimalToDouble(col 2) -> 5:double) 
-> 6:double) -> 5:double, FuncLnDoubleToDouble(col 6)(children: 
CastDecimalToDouble(col 2) -> 6:double) -> 11:double, 
FuncLog10DoubleToDouble(col 6)(children: CastDecimalToDouble(col 2) -> 
6:double) -> 12:double, FuncLog2DoubleToDouble(col 6)(children: 
CastDecimalToDouble(col 2) -> 6:double) -> 13:double, 
FuncLog2DoubleToDouble(col 6)(children: CastDecimalToDouble(col 14)(children: 
DecimalColSubtractDecimalScalar(col 2, val 15601) -> 14:decimal(21,10)) -> 
6:double) -> 15:double, VectorUDFAdaptor(log(2, cdecimal1)) -> 6:Double, 
VectorUDFAdaptor(power(log2(cd
 ecimal1), 2))(children: FuncLog2DoubleToDouble(col 16)(children: 
CastDecimalToDouble(col 2) -> 16:double) -> 17:double) -> 16:Double, 
VectorUDFAdaptor(power(log2(cdecimal1), 2))(children: 
FuncLog2DoubleToDouble(col 17)(children: CastDecimalToDouble(col 2) -> 
17:double) -> 18:double) -> 17:Double, FuncSqrtDoubleToDouble(col 18)(children: 
CastDecimalToDouble(col 2) -> 18:double) -> 19:double, 
FuncAbsDecimalToDecimal(col 2) -> 20:decimal(20,10), FuncSinDoubleToDouble(col 
18)(children: CastDecimalToDouble(col 2) -> 18:double) -> 21:double, 
FuncASinDoubleToDouble(col 18)(children: CastDecimalToDouble(col 2) -> 
18:double) -> 22:double, FuncCosDoubleToDouble(col 18)(children: 
CastDecimalToDouble(col 2) -> 18:double) -> 23:double, 
FuncACosDoubleToDouble(col 18)(children: CastDecimalToDouble(col 2) -> 
18:double) -> 24:double, FuncATanDoubleToDouble(col 18)(children: 
CastDecimalToDouble(col 2) -> 18:double) -> 25:double, 
FuncDegreesDoubleToDouble(col 18)(children: CastDecimalToDouble(col 2) -
 > 18:double) -> 26:double, FuncRadiansDoubleToDouble(col 18)(children: 
 > CastDecimalToDouble(col 2) -> 18:double) -> 27:double, 
 > FuncNegateDecimalToDecimal(col 2) -> 28:decimal(20,10), 
 > FuncSignDecimalToLong(col 2) -> 4:long, FuncCosDoubleToDouble(col 
 > 18)(children: DoubleColAddDoubleScalar(col 29, val 3.14159)(children: 
 > DoubleColUnaryMinus(col 18)(children: FuncSinDoubleToDouble(col 
 > 29)(children: FuncLnDoubleToDouble(col 18)(children: CastDecimalToDouble(col 
 > 2) -> 18:double) -> 29:double) -> 18:double) -> 29:double) -> 18:double) -> 
 > 29:double
-                      Statistics: Num rows: 2048 Data size: 366958 Basic 
stats: COMPLETE Column stats: NONE
-                      File Output Operator
-                        compressed: false
-                        File Sink Vectorization:
-                            className: VectorFileSinkOperator
-                            native: false
-                        Statistics: Num rows: 2048 Data size: 366958 Basic 
stats: COMPLETE Column stats: NONE
-                        table:
-                            input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                            output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                            serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: true
-                vectorized: true
-
   Stage: Stage-0
     Fetch Operator
       limit: -1
       Processor Tree:
-        ListSink
+        TableScan
+          alias: decimal_test
+          Filter Operator
+            predicate: (((cbigint % 500) = 0) and (sin(cdecimal1) >= -1.0)) 
(type: boolean)
+            Select Operator
+              expressions: cdecimal1 (type: decimal(20,10)), round(cdecimal1, 
2) (type: decimal(13,2)), round(cdecimal1) (type: decimal(11,0)), 
floor(cdecimal1) (type: decimal(11,0)), ceil(cdecimal1) (type: decimal(11,0)), 
round(exp(cdecimal1), 58) (type: double), ln(cdecimal1) (type: double), 
log10(cdecimal1) (type: double), log2(cdecimal1) (type: double), 
log2((cdecimal1 - 15601)) (type: double), log(2, cdecimal1) (type: double), 
power(log2(cdecimal1), 2) (type: double), power(log2(cdecimal1), 2) (type: 
double), sqrt(cdecimal1) (type: double), abs(cdecimal1) (type: decimal(20,10)), 
sin(cdecimal1) (type: double), asin(cdecimal1) (type: double), cos(cdecimal1) 
(type: double), acos(cdecimal1) (type: double), atan(cdecimal1) (type: double), 
degrees(cdecimal1) (type: double), radians(cdecimal1) (type: double), cdecimal1 
(type: decimal(20,10)), (- cdecimal1) (type: decimal(20,10)), sign(cdecimal1) 
(type: int), cos(((- sin(log(cdecimal1))) + 3.14159)) (type: double)
+              outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, 
_col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25
+              ListSink
 
 PREHOOK: query: select
    cdecimal1

http://git-wip-us.apache.org/repos/asf/hive/blob/aed21d0b/ql/src/test/results/clientpositive/llap/vector_decimal_precision.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_decimal_precision.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_precision.q.out
index 90f80a3..5583874 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_precision.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_precision.q.out
@@ -545,14 +545,10 @@ NULL      NULL
 123456789.0123456789   15241578753238836.75019051998750190521
 1234567890.1234560000  NULL
 1234567890.1234567890  NULL
-PREHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT avg(dec), sum(dec) 
FROM DECIMAL_PRECISION
+PREHOOK: query: EXPLAIN SELECT avg(dec), sum(dec) FROM DECIMAL_PRECISION
 PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN VECTORIZATION EXPRESSION SELECT avg(dec), sum(dec) 
FROM DECIMAL_PRECISION
+POSTHOOK: query: EXPLAIN SELECT avg(dec), sum(dec) FROM DECIMAL_PRECISION
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -570,26 +566,12 @@ STAGE PLANS:
                 TableScan
                   alias: decimal_precision
                   Statistics: Num rows: 75 Data size: 3472 Basic stats: 
COMPLETE Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0]
                   Select Operator
                     expressions: dec (type: decimal(20,10))
                     outputColumnNames: dec
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [0]
                     Statistics: Num rows: 75 Data size: 3472 Basic stats: 
COMPLETE Column stats: NONE
                     Group By Operator
                       aggregations: avg(dec), sum(dec)
-                      Group By Vectorization:
-                          aggregators: VectorUDAFAvgDecimal(col 0) -> 
struct<count:bigint,sum:decimal(30,10)>, VectorUDAFSumDecimal(col 0) -> 
decimal(38,18)
-                          className: VectorGroupByOperator
-                          vectorOutput: false
-                          native: false
-                          projectedOutputColumns: [0, 1]
-                          vectorOutputConditionsNotMet: Vector output of 
VectorUDAFAvgDecimal(col 0) -> struct<count:bigint,sum:decimal(30,10)> output 
type STRUCT requires PRIMITIVE IS false
                       mode: hash
                       outputColumnNames: _col0, _col1
                       Statistics: Num rows: 1 Data size: 400 Basic stats: 
COMPLETE Column stats: NONE
@@ -599,21 +581,8 @@ STAGE PLANS:
                         value expressions: _col0 (type: 
struct<count:bigint,sum:decimal(30,10),input:decimal(20,10)>), _col1 (type: 
decimal(30,10))
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: false
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                notVectorizedReason: Aggregation Function UDF avg parameter 
expression for GROUPBY operator: Data type 
struct<count:bigint,sum:decimal(30,10),input:decimal(20,10)> of 
Column[VALUE._col0] not supported
-                vectorized: false
             Reduce Operator Tree:
               Group By Operator
                 aggregations: avg(VALUE._col0), sum(VALUE._col1)

http://git-wip-us.apache.org/repos/asf/hive/blob/aed21d0b/ql/src/test/results/clientpositive/llap/vector_decimal_round.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_decimal_round.q.out 
b/ql/src/test/results/clientpositive/llap/vector_decimal_round.q.out
index b99935d..4063199 100644
--- a/ql/src/test/results/clientpositive/llap/vector_decimal_round.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_decimal_round.q.out
@@ -30,16 +30,12 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@decimal_tbl_txt
 #### A masked pattern was here ####
 101
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select dec, round(dec, -1) from decimal_tbl_txt order by dec
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select dec, round(dec, -1) from decimal_tbl_txt order by dec
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -68,33 +64,15 @@ STAGE PLANS:
                       value expressions: _col1 (type: decimal(11,0))
             Execution mode: llap
             LLAP IO: no inputs
-            Map Vectorization:
-                enabled: false
-                enabledConditionsNotMet: 
hive.vectorized.use.vector.serde.deserialize IS false
-                inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: decimal(10,0)), 
VALUE._col0 (type: decimal(11,0))
                 outputColumnNames: _col0, _col1
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [0, 1]
                 Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
                   Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -116,16 +94,12 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@decimal_tbl_txt
 #### A masked pattern was here ####
 101    100
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select dec, round(dec, -1) from decimal_tbl_txt order by round(dec, -1)
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select dec, round(dec, -1) from decimal_tbl_txt order by round(dec, -1)
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -154,33 +128,15 @@ STAGE PLANS:
                       value expressions: _col0 (type: decimal(10,0))
             Execution mode: llap
             LLAP IO: no inputs
-            Map Vectorization:
-                enabled: false
-                enabledConditionsNotMet: 
hive.vectorized.use.vector.serde.deserialize IS false
-                inputFileFormats: org.apache.hadoop.mapred.TextInputFormat
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: VALUE._col0 (type: decimal(10,0)), 
KEY.reducesinkkey0 (type: decimal(11,0))
                 outputColumnNames: _col0, _col1
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [1, 0]
                 Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
                   Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -230,16 +186,12 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@decimal_tbl_rc
 #### A masked pattern was here ####
 101
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select dec, round(dec, -1) from decimal_tbl_rc order by dec
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select dec, round(dec, -1) from decimal_tbl_rc order by dec
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -268,33 +220,15 @@ STAGE PLANS:
                       value expressions: _col1 (type: decimal(11,0))
             Execution mode: llap
             LLAP IO: no inputs
-            Map Vectorization:
-                enabled: false
-                enabledConditionsNotMet: 
hive.vectorized.use.row.serde.deserialize IS false
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.RCFileInputFormat
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: decimal(10,0)), 
VALUE._col0 (type: decimal(11,0))
                 outputColumnNames: _col0, _col1
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [0, 1]
                 Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
                   Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -316,16 +250,12 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@decimal_tbl_rc
 #### A masked pattern was here ####
 101    100
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select dec, round(dec, -1) from decimal_tbl_rc order by round(dec, -1)
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select dec, round(dec, -1) from decimal_tbl_rc order by round(dec, -1)
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -354,33 +284,15 @@ STAGE PLANS:
                       value expressions: _col0 (type: decimal(10,0))
             Execution mode: llap
             LLAP IO: no inputs
-            Map Vectorization:
-                enabled: false
-                enabledConditionsNotMet: 
hive.vectorized.use.row.serde.deserialize IS false
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.RCFileInputFormat
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: VALUE._col0 (type: decimal(10,0)), 
KEY.reducesinkkey0 (type: decimal(11,0))
                 outputColumnNames: _col0, _col1
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [1, 0]
                 Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
                   Statistics: Num rows: 1 Data size: 3 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -430,16 +342,12 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@decimal_tbl_orc
 #### A masked pattern was here ####
 101
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select dec, round(dec, -1) from decimal_tbl_orc order by dec
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select dec, round(dec, -1) from decimal_tbl_orc order by dec
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -457,61 +365,26 @@ STAGE PLANS:
                 TableScan
                   alias: decimal_tbl_orc
                   Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE 
Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0]
                   Select Operator
                     expressions: dec (type: decimal(10,0)), round(dec, -1) 
(type: decimal(11,0))
                     outputColumnNames: _col0, _col1
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [0, 1]
-                        selectExpressions: 
FuncRoundWithNumDigitsDecimalToDecimal(col 0, decimalPlaces -1) -> 
1:decimal(11,0)
                     Statistics: Num rows: 1 Data size: 112 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       key expressions: _col0 (type: decimal(10,0))
                       sort order: +
-                      Reduce Sink Vectorization:
-                          className: VectorReduceSinkOperator
-                          native: false
-                          nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for 
keys IS true, LazyBinarySerDe for values IS true
-                          nativeConditionsNotMet: Uniform Hash IS false
                       Statistics: Num rows: 1 Data size: 112 Basic stats: 
COMPLETE Column stats: NONE
                       value expressions: _col1 (type: decimal(11,0))
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: decimal(10,0)), 
VALUE._col0 (type: decimal(11,0))
                 outputColumnNames: _col0, _col1
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [0, 1]
                 Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
                   Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -533,16 +406,12 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@decimal_tbl_orc
 #### A masked pattern was here ####
 101    100
-PREHOOK: query: explain vectorization expression
+PREHOOK: query: explain
 select dec, round(dec, -1) from decimal_tbl_orc order by round(dec, -1)
 PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
+POSTHOOK: query: explain
 select dec, round(dec, -1) from decimal_tbl_orc order by round(dec, -1)
 POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 depends on stages: Stage-1
@@ -560,60 +429,26 @@ STAGE PLANS:
                 TableScan
                   alias: decimal_tbl_orc
                   Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE 
Column stats: NONE
-                  TableScan Vectorization:
-                      native: true
-                      projectedOutputColumns: [0]
                   Select Operator
                     expressions: dec (type: decimal(10,0))
                     outputColumnNames: _col0
-                    Select Vectorization:
-                        className: VectorSelectOperator
-                        native: true
-                        projectedOutputColumns: [0]
                     Statistics: Num rows: 1 Data size: 112 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       key expressions: round(_col0, -1) (type: decimal(11,0))
                       sort order: +
-                      Reduce Sink Vectorization:
-                          className: VectorReduceSinkOperator
-                          native: false
-                          nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, Not ACID UPDATE or DELETE IS true, No buckets IS 
true, No TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for 
keys IS true, LazyBinarySerDe for values IS true
-                          nativeConditionsNotMet: Uniform Hash IS false
                       Statistics: Num rows: 1 Data size: 112 Basic stats: 
COMPLETE Column stats: NONE
                       value expressions: _col0 (type: decimal(10,0))
             Execution mode: vectorized, llap
             LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                groupByVectorOutput: true
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
         Reducer 2 
             Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                groupByVectorOutput: true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
             Reduce Operator Tree:
               Select Operator
                 expressions: VALUE._col0 (type: decimal(10,0)), 
KEY.reducesinkkey0 (type: decimal(11,0))
                 outputColumnNames: _col0, _col1
-                Select Vectorization:
-                    className: VectorSelectOperator
-                    native: true
-                    projectedOutputColumns: [1, 0]
                 Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
                   Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat

Reply via email to