http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/ctas.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/ctas.q.out 
b/ql/src/test/results/clientpositive/tez/ctas.q.out
deleted file mode 100644
index b26f7a2..0000000
--- a/ql/src/test/results/clientpositive/tez/ctas.q.out
+++ /dev/null
@@ -1,905 +0,0 @@
-PREHOOK: query: -- EXCLUDE_HADOOP_MAJOR_VERSIONS( 0.20S)
--- SORT_QUERY_RESULTS
-
-create table nzhang_Tmp(a int, b string)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@nzhang_Tmp
-POSTHOOK: query: -- EXCLUDE_HADOOP_MAJOR_VERSIONS( 0.20S)
--- SORT_QUERY_RESULTS
-
-create table nzhang_Tmp(a int, b string)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@nzhang_Tmp
-PREHOOK: query: select * from nzhang_Tmp
-PREHOOK: type: QUERY
-PREHOOK: Input: default@nzhang_tmp
-#### A masked pattern was here ####
-POSTHOOK: query: select * from nzhang_Tmp
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@nzhang_tmp
-#### A masked pattern was here ####
-PREHOOK: query: explain create table nzhang_CTAS1 as select key k, value from 
src sort by k, value limit 10
-PREHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: query: explain create table nzhang_CTAS1 as select key k, value from 
src sort by k, value limit 10
-POSTHOOK: type: CREATETABLE_AS_SELECT
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-2 depends on stages: Stage-1
-  Stage-4 depends on stages: Stage-2, Stage-0
-  Stage-3 depends on stages: Stage-4
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-        Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: src
-                  Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  Select Operator
-                    expressions: key (type: string), value (type: string)
-                    outputColumnNames: _col0, _col1
-                    Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    Reduce Output Operator
-                      key expressions: _col0 (type: string), _col1 (type: 
string)
-                      sort order: ++
-                      Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                      TopN Hash Memory Usage: 0.1
-        Reducer 2 
-            Reduce Operator Tree:
-              Select Operator
-                expressions: KEY.reducesinkkey0 (type: string), 
KEY.reducesinkkey1 (type: string)
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                Limit
-                  Number of rows: 10
-                  Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  Reduce Output Operator
-                    key expressions: _col0 (type: string), _col1 (type: string)
-                    sort order: ++
-                    Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    TopN Hash Memory Usage: 0.1
-        Reducer 3 
-            Reduce Operator Tree:
-              Select Operator
-                expressions: KEY.reducesinkkey0 (type: string), 
KEY.reducesinkkey1 (type: string)
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE 
Column stats: COMPLETE
-                Limit
-                  Number of rows: 10
-                  Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  File Output Operator
-                    compressed: false
-                    Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                        name: default.nzhang_CTAS1
-
-  Stage: Stage-2
-    Dependency Collection
-
-  Stage: Stage-4
-      Create Table Operator:
-        Create Table
-          columns: k string, value string
-          input format: org.apache.hadoop.mapred.TextInputFormat
-          output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-          serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: default.nzhang_CTAS1
-
-  Stage: Stage-3
-    Stats-Aggr Operator
-
-  Stage: Stage-0
-    Move Operator
-      files:
-          hdfs directory: true
-#### A masked pattern was here ####
-
-PREHOOK: query: create table nzhang_CTAS1 as select key k, value from src sort 
by k, value limit 10
-PREHOOK: type: CREATETABLE_AS_SELECT
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@nzhang_CTAS1
-POSTHOOK: query: create table nzhang_CTAS1 as select key k, value from src 
sort by k, value limit 10
-POSTHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@nzhang_CTAS1
-POSTHOOK: Lineage: nzhang_ctas1.k SIMPLE [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
-POSTHOOK: Lineage: nzhang_ctas1.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
-PREHOOK: query: select * from nzhang_CTAS1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@nzhang_ctas1
-#### A masked pattern was here ####
-POSTHOOK: query: select * from nzhang_CTAS1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@nzhang_ctas1
-#### A masked pattern was here ####
-0      val_0
-0      val_0
-0      val_0
-10     val_10
-100    val_100
-100    val_100
-103    val_103
-103    val_103
-104    val_104
-104    val_104
-PREHOOK: query: describe formatted nzhang_CTAS1
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@nzhang_ctas1
-POSTHOOK: query: describe formatted nzhang_CTAS1
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@nzhang_ctas1
-# col_name             data_type               comment             
-                
-k                      string                                      
-value                  string                                      
-                
-# Detailed Table Information            
-Database:              default                  
-#### A masked pattern was here ####
-Retention:             0                        
-#### A masked pattern was here ####
-Table Type:            MANAGED_TABLE            
-Table Parameters:               
-       COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
-       numFiles                1                   
-       numRows                 10                  
-       rawDataSize             96                  
-       totalSize               106                 
-#### A masked pattern was here ####
-                
-# Storage Information           
-SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
-InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
-OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
-Compressed:            No                       
-Num Buckets:           -1                       
-Bucket Columns:        []                       
-Sort Columns:          []                       
-Storage Desc Params:            
-       serialization.format    1                   
-PREHOOK: query: explain create table nzhang_ctas2 as select * from src sort by 
key, value limit 10
-PREHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: query: explain create table nzhang_ctas2 as select * from src sort 
by key, value limit 10
-POSTHOOK: type: CREATETABLE_AS_SELECT
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-2 depends on stages: Stage-1
-  Stage-4 depends on stages: Stage-2, Stage-0
-  Stage-3 depends on stages: Stage-4
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-        Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: src
-                  Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  Select Operator
-                    expressions: key (type: string), value (type: string)
-                    outputColumnNames: _col0, _col1
-                    Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    Reduce Output Operator
-                      key expressions: _col0 (type: string), _col1 (type: 
string)
-                      sort order: ++
-                      Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                      TopN Hash Memory Usage: 0.1
-        Reducer 2 
-            Reduce Operator Tree:
-              Select Operator
-                expressions: KEY.reducesinkkey0 (type: string), 
KEY.reducesinkkey1 (type: string)
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                Limit
-                  Number of rows: 10
-                  Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  Reduce Output Operator
-                    key expressions: _col0 (type: string), _col1 (type: string)
-                    sort order: ++
-                    Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    TopN Hash Memory Usage: 0.1
-        Reducer 3 
-            Reduce Operator Tree:
-              Select Operator
-                expressions: KEY.reducesinkkey0 (type: string), 
KEY.reducesinkkey1 (type: string)
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE 
Column stats: COMPLETE
-                Limit
-                  Number of rows: 10
-                  Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  File Output Operator
-                    compressed: false
-                    Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                        name: default.nzhang_ctas2
-
-  Stage: Stage-2
-    Dependency Collection
-
-  Stage: Stage-4
-      Create Table Operator:
-        Create Table
-          columns: key string, value string
-          input format: org.apache.hadoop.mapred.TextInputFormat
-          output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-          serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: default.nzhang_ctas2
-
-  Stage: Stage-3
-    Stats-Aggr Operator
-
-  Stage: Stage-0
-    Move Operator
-      files:
-          hdfs directory: true
-#### A masked pattern was here ####
-
-PREHOOK: query: create table nzhang_ctas2 as select * from src sort by key, 
value limit 10
-PREHOOK: type: CREATETABLE_AS_SELECT
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@nzhang_ctas2
-POSTHOOK: query: create table nzhang_ctas2 as select * from src sort by key, 
value limit 10
-POSTHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@nzhang_ctas2
-POSTHOOK: Lineage: nzhang_ctas2.key SIMPLE [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
-POSTHOOK: Lineage: nzhang_ctas2.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
-PREHOOK: query: select * from nzhang_ctas2
-PREHOOK: type: QUERY
-PREHOOK: Input: default@nzhang_ctas2
-#### A masked pattern was here ####
-POSTHOOK: query: select * from nzhang_ctas2
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@nzhang_ctas2
-#### A masked pattern was here ####
-0      val_0
-0      val_0
-0      val_0
-10     val_10
-100    val_100
-100    val_100
-103    val_103
-103    val_103
-104    val_104
-104    val_104
-PREHOOK: query: describe formatted nzhang_CTAS2
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@nzhang_ctas2
-POSTHOOK: query: describe formatted nzhang_CTAS2
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@nzhang_ctas2
-# col_name             data_type               comment             
-                
-key                    string                                      
-value                  string                                      
-                
-# Detailed Table Information            
-Database:              default                  
-#### A masked pattern was here ####
-Retention:             0                        
-#### A masked pattern was here ####
-Table Type:            MANAGED_TABLE            
-Table Parameters:               
-       COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
-       numFiles                1                   
-       numRows                 10                  
-       rawDataSize             96                  
-       totalSize               106                 
-#### A masked pattern was here ####
-                
-# Storage Information           
-SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
-InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
-OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
-Compressed:            No                       
-Num Buckets:           -1                       
-Bucket Columns:        []                       
-Sort Columns:          []                       
-Storage Desc Params:            
-       serialization.format    1                   
-PREHOOK: query: explain create table nzhang_ctas3 row format serde 
"org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe" stored as RCFile as 
select key/2 half_key, concat(value, "_con") conb  from src sort by half_key, 
conb limit 10
-PREHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: query: explain create table nzhang_ctas3 row format serde 
"org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe" stored as RCFile as 
select key/2 half_key, concat(value, "_con") conb  from src sort by half_key, 
conb limit 10
-POSTHOOK: type: CREATETABLE_AS_SELECT
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-2 depends on stages: Stage-1
-  Stage-4 depends on stages: Stage-2, Stage-0
-  Stage-3 depends on stages: Stage-4
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-        Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: src
-                  Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  Select Operator
-                    expressions: (key / 2) (type: double), concat(value, 
'_con') (type: string)
-                    outputColumnNames: _col0, _col1
-                    Statistics: Num rows: 500 Data size: 96000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    Reduce Output Operator
-                      key expressions: _col0 (type: double), _col1 (type: 
string)
-                      sort order: ++
-                      Statistics: Num rows: 500 Data size: 96000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                      TopN Hash Memory Usage: 0.1
-        Reducer 2 
-            Reduce Operator Tree:
-              Select Operator
-                expressions: KEY.reducesinkkey0 (type: double), 
KEY.reducesinkkey1 (type: string)
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 500 Data size: 96000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                Limit
-                  Number of rows: 10
-                  Statistics: Num rows: 10 Data size: 1920 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  Reduce Output Operator
-                    key expressions: _col0 (type: double), _col1 (type: string)
-                    sort order: ++
-                    Statistics: Num rows: 10 Data size: 1920 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    TopN Hash Memory Usage: 0.1
-        Reducer 3 
-            Reduce Operator Tree:
-              Select Operator
-                expressions: KEY.reducesinkkey0 (type: double), 
KEY.reducesinkkey1 (type: string)
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 10 Data size: 1920 Basic stats: COMPLETE 
Column stats: COMPLETE
-                Limit
-                  Number of rows: 10
-                  Statistics: Num rows: 10 Data size: 1920 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  File Output Operator
-                    compressed: false
-                    Statistics: Num rows: 10 Data size: 1920 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    table:
-                        input format: 
org.apache.hadoop.hive.ql.io.RCFileInputFormat
-                        output format: 
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-                        serde: 
org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
-                        name: default.nzhang_ctas3
-
-  Stage: Stage-2
-    Dependency Collection
-
-  Stage: Stage-4
-      Create Table Operator:
-        Create Table
-          columns: half_key double, conb string
-          input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
-          output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-          serde name: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
-          name: default.nzhang_ctas3
-
-  Stage: Stage-3
-    Stats-Aggr Operator
-
-  Stage: Stage-0
-    Move Operator
-      files:
-          hdfs directory: true
-#### A masked pattern was here ####
-
-PREHOOK: query: create table nzhang_ctas3 row format serde 
"org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe" stored as RCFile as 
select key/2 half_key, concat(value, "_con") conb  from src sort by half_key, 
conb limit 10
-PREHOOK: type: CREATETABLE_AS_SELECT
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@nzhang_ctas3
-POSTHOOK: query: create table nzhang_ctas3 row format serde 
"org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe" stored as RCFile as 
select key/2 half_key, concat(value, "_con") conb  from src sort by half_key, 
conb limit 10
-POSTHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@nzhang_ctas3
-POSTHOOK: Lineage: nzhang_ctas3.conb EXPRESSION 
[(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: nzhang_ctas3.half_key EXPRESSION 
[(src)src.FieldSchema(name:key, type:string, comment:default), ]
-PREHOOK: query: select * from nzhang_ctas3
-PREHOOK: type: QUERY
-PREHOOK: Input: default@nzhang_ctas3
-#### A masked pattern was here ####
-POSTHOOK: query: select * from nzhang_ctas3
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@nzhang_ctas3
-#### A masked pattern was here ####
-0.0    val_0_con
-0.0    val_0_con
-0.0    val_0_con
-1.0    val_2_con
-2.0    val_4_con
-2.5    val_5_con
-2.5    val_5_con
-2.5    val_5_con
-4.0    val_8_con
-4.5    val_9_con
-PREHOOK: query: describe formatted nzhang_CTAS3
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@nzhang_ctas3
-POSTHOOK: query: describe formatted nzhang_CTAS3
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@nzhang_ctas3
-# col_name             data_type               comment             
-                
-half_key               double                                      
-conb                   string                                      
-                
-# Detailed Table Information            
-Database:              default                  
-#### A masked pattern was here ####
-Retention:             0                        
-#### A masked pattern was here ####
-Table Type:            MANAGED_TABLE            
-Table Parameters:               
-       COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
-       numFiles                1                   
-       numRows                 10                  
-       rawDataSize             120                 
-       totalSize               199                 
-#### A masked pattern was here ####
-                
-# Storage Information           
-SerDe Library:         org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe    
 
-InputFormat:           org.apache.hadoop.hive.ql.io.RCFileInputFormat   
-OutputFormat:          org.apache.hadoop.hive.ql.io.RCFileOutputFormat  
-Compressed:            No                       
-Num Buckets:           -1                       
-Bucket Columns:        []                       
-Sort Columns:          []                       
-Storage Desc Params:            
-       serialization.format    1                   
-PREHOOK: query: explain create table if not exists nzhang_ctas3 as select key, 
value from src sort by key, value limit 2
-PREHOOK: type: CREATETABLE
-POSTHOOK: query: explain create table if not exists nzhang_ctas3 as select 
key, value from src sort by key, value limit 2
-POSTHOOK: type: CREATETABLE
-STAGE DEPENDENCIES:
-
-STAGE PLANS:
-PREHOOK: query: create table if not exists nzhang_ctas3 as select key, value 
from src sort by key, value limit 2
-PREHOOK: type: CREATETABLE
-POSTHOOK: query: create table if not exists nzhang_ctas3 as select key, value 
from src sort by key, value limit 2
-POSTHOOK: type: CREATETABLE
-PREHOOK: query: select * from nzhang_ctas3
-PREHOOK: type: QUERY
-PREHOOK: Input: default@nzhang_ctas3
-#### A masked pattern was here ####
-POSTHOOK: query: select * from nzhang_ctas3
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@nzhang_ctas3
-#### A masked pattern was here ####
-0.0    val_0_con
-0.0    val_0_con
-0.0    val_0_con
-1.0    val_2_con
-2.0    val_4_con
-2.5    val_5_con
-2.5    val_5_con
-2.5    val_5_con
-4.0    val_8_con
-4.5    val_9_con
-PREHOOK: query: describe formatted nzhang_CTAS3
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@nzhang_ctas3
-POSTHOOK: query: describe formatted nzhang_CTAS3
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@nzhang_ctas3
-# col_name             data_type               comment             
-                
-half_key               double                                      
-conb                   string                                      
-                
-# Detailed Table Information            
-Database:              default                  
-#### A masked pattern was here ####
-Retention:             0                        
-#### A masked pattern was here ####
-Table Type:            MANAGED_TABLE            
-Table Parameters:               
-       COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
-       numFiles                1                   
-       numRows                 10                  
-       rawDataSize             120                 
-       totalSize               199                 
-#### A masked pattern was here ####
-                
-# Storage Information           
-SerDe Library:         org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe    
 
-InputFormat:           org.apache.hadoop.hive.ql.io.RCFileInputFormat   
-OutputFormat:          org.apache.hadoop.hive.ql.io.RCFileOutputFormat  
-Compressed:            No                       
-Num Buckets:           -1                       
-Bucket Columns:        []                       
-Sort Columns:          []                       
-Storage Desc Params:            
-       serialization.format    1                   
-PREHOOK: query: explain create table nzhang_ctas4 row format delimited fields 
terminated by ',' stored as textfile as select key, value from src sort by key, 
value limit 10
-PREHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: query: explain create table nzhang_ctas4 row format delimited fields 
terminated by ',' stored as textfile as select key, value from src sort by key, 
value limit 10
-POSTHOOK: type: CREATETABLE_AS_SELECT
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-2 depends on stages: Stage-1
-  Stage-4 depends on stages: Stage-2, Stage-0
-  Stage-3 depends on stages: Stage-4
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-        Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: src
-                  Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  Select Operator
-                    expressions: key (type: string), value (type: string)
-                    outputColumnNames: _col0, _col1
-                    Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    Reduce Output Operator
-                      key expressions: _col0 (type: string), _col1 (type: 
string)
-                      sort order: ++
-                      Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                      TopN Hash Memory Usage: 0.1
-        Reducer 2 
-            Reduce Operator Tree:
-              Select Operator
-                expressions: KEY.reducesinkkey0 (type: string), 
KEY.reducesinkkey1 (type: string)
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                Limit
-                  Number of rows: 10
-                  Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  Reduce Output Operator
-                    key expressions: _col0 (type: string), _col1 (type: string)
-                    sort order: ++
-                    Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    TopN Hash Memory Usage: 0.1
-        Reducer 3 
-            Reduce Operator Tree:
-              Select Operator
-                expressions: KEY.reducesinkkey0 (type: string), 
KEY.reducesinkkey1 (type: string)
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE 
Column stats: COMPLETE
-                Limit
-                  Number of rows: 10
-                  Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  File Output Operator
-                    compressed: false
-                    Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                        name: default.nzhang_ctas4
-
-  Stage: Stage-2
-    Dependency Collection
-
-  Stage: Stage-4
-      Create Table Operator:
-        Create Table
-          columns: key string, value string
-          field delimiter: ,
-          input format: org.apache.hadoop.mapred.TextInputFormat
-          output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-          serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: default.nzhang_ctas4
-
-  Stage: Stage-3
-    Stats-Aggr Operator
-
-  Stage: Stage-0
-    Move Operator
-      files:
-          hdfs directory: true
-#### A masked pattern was here ####
-
-PREHOOK: query: create table nzhang_ctas4 row format delimited fields 
terminated by ',' stored as textfile as select key, value from src sort by key, 
value limit 10
-PREHOOK: type: CREATETABLE_AS_SELECT
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@nzhang_ctas4
-POSTHOOK: query: create table nzhang_ctas4 row format delimited fields 
terminated by ',' stored as textfile as select key, value from src sort by key, 
value limit 10
-POSTHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@nzhang_ctas4
-POSTHOOK: Lineage: nzhang_ctas4.key SIMPLE [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
-POSTHOOK: Lineage: nzhang_ctas4.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
-PREHOOK: query: select * from nzhang_ctas4
-PREHOOK: type: QUERY
-PREHOOK: Input: default@nzhang_ctas4
-#### A masked pattern was here ####
-POSTHOOK: query: select * from nzhang_ctas4
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@nzhang_ctas4
-#### A masked pattern was here ####
-0      val_0
-0      val_0
-0      val_0
-10     val_10
-100    val_100
-100    val_100
-103    val_103
-103    val_103
-104    val_104
-104    val_104
-PREHOOK: query: describe formatted nzhang_CTAS4
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@nzhang_ctas4
-POSTHOOK: query: describe formatted nzhang_CTAS4
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@nzhang_ctas4
-# col_name             data_type               comment             
-                
-key                    string                                      
-value                  string                                      
-                
-# Detailed Table Information            
-Database:              default                  
-#### A masked pattern was here ####
-Retention:             0                        
-#### A masked pattern was here ####
-Table Type:            MANAGED_TABLE            
-Table Parameters:               
-       COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
-       numFiles                1                   
-       numRows                 10                  
-       rawDataSize             96                  
-       totalSize               106                 
-#### A masked pattern was here ####
-                
-# Storage Information           
-SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
-InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
-OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
-Compressed:            No                       
-Num Buckets:           -1                       
-Bucket Columns:        []                       
-Sort Columns:          []                       
-Storage Desc Params:            
-       field.delim             ,                   
-       serialization.format    ,                   
-PREHOOK: query: explain extended create table nzhang_ctas5 row format 
delimited fields terminated by ',' lines terminated by '\012' stored as 
textfile as select key, value from src sort by key, value limit 10
-PREHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: query: explain extended create table nzhang_ctas5 row format 
delimited fields terminated by ',' lines terminated by '\012' stored as 
textfile as select key, value from src sort by key, value limit 10
-POSTHOOK: type: CREATETABLE_AS_SELECT
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-2 depends on stages: Stage-1
-  Stage-4 depends on stages: Stage-2, Stage-0
-  Stage-3 depends on stages: Stage-4
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-        Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: src
-                  Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  GatherStats: false
-                  Select Operator
-                    expressions: key (type: string), value (type: string)
-                    outputColumnNames: _col0, _col1
-                    Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    Reduce Output Operator
-                      key expressions: _col0 (type: string), _col1 (type: 
string)
-                      null sort order: aa
-                      sort order: ++
-                      Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                      tag: -1
-                      TopN: 10
-                      TopN Hash Memory Usage: 0.1
-                      auto parallelism: true
-            Path -> Alias:
-#### A masked pattern was here ####
-            Path -> Partition:
-#### A masked pattern was here ####
-                Partition
-                  base file name: src
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                  properties:
-                    COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
-                    bucket_count -1
-                    columns key,value
-                    columns.comments 'default','default'
-                    columns.types string:string
-#### A masked pattern was here ####
-                    name default.src
-                    numFiles 1
-                    numRows 500
-                    rawDataSize 5312
-                    serialization.ddl struct src { string key, string value}
-                    serialization.format 1
-                    serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    totalSize 5812
-#### A masked pattern was here ####
-                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                
-                    input format: org.apache.hadoop.mapred.TextInputFormat
-                    output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                    properties:
-                      COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
-                      bucket_count -1
-                      columns key,value
-                      columns.comments 'default','default'
-                      columns.types string:string
-#### A masked pattern was here ####
-                      name default.src
-                      numFiles 1
-                      numRows 500
-                      rawDataSize 5312
-                      serialization.ddl struct src { string key, string value}
-                      serialization.format 1
-                      serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                      totalSize 5812
-#### A masked pattern was here ####
-                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                    name: default.src
-                  name: default.src
-            Truncated Path -> Alias:
-              /src [src]
-        Reducer 2 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Select Operator
-                expressions: KEY.reducesinkkey0 (type: string), 
KEY.reducesinkkey1 (type: string)
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
-                Limit
-                  Number of rows: 10
-                  Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  Reduce Output Operator
-                    key expressions: _col0 (type: string), _col1 (type: string)
-                    null sort order: aa
-                    sort order: ++
-                    Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    tag: -1
-                    TopN: 10
-                    TopN Hash Memory Usage: 0.1
-                    auto parallelism: false
-        Reducer 3 
-            Needs Tagging: false
-            Reduce Operator Tree:
-              Select Operator
-                expressions: KEY.reducesinkkey0 (type: string), 
KEY.reducesinkkey1 (type: string)
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE 
Column stats: COMPLETE
-                Limit
-                  Number of rows: 10
-                  Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 1
-#### A masked pattern was here ####
-                    NumFilesPerFileSink: 1
-                    Statistics: Num rows: 10 Data size: 1780 Basic stats: 
COMPLETE Column stats: COMPLETE
-#### A masked pattern was here ####
-                    table:
-                        input format: org.apache.hadoop.mapred.TextInputFormat
-                        output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-                        properties:
-                          columns key,value
-                          columns.types string:string
-                          field.delim ,
-                          line.delim 
-
-                          name default.nzhang_ctas5
-                          serialization.format ,
-                          serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-                        name: default.nzhang_ctas5
-                    TotalFiles: 1
-                    GatherStats: true
-                    MultiFileSpray: false
-
-  Stage: Stage-2
-    Dependency Collection
-
-  Stage: Stage-4
-      Create Table Operator:
-        Create Table
-          columns: key string, value string
-          field delimiter: ,
-          input format: org.apache.hadoop.mapred.TextInputFormat
-          line delimiter: 
-
-          output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
-          serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-          name: default.nzhang_ctas5
-
-  Stage: Stage-3
-    Stats-Aggr Operator
-#### A masked pattern was here ####
-
-  Stage: Stage-0
-    Move Operator
-      files:
-          hdfs directory: true
-#### A masked pattern was here ####
-
-PREHOOK: query: create table nzhang_ctas5 row format delimited fields 
terminated by ',' lines terminated by '\012' stored as textfile as select key, 
value from src sort by key, value limit 10
-PREHOOK: type: CREATETABLE_AS_SELECT
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@nzhang_ctas5
-POSTHOOK: query: create table nzhang_ctas5 row format delimited fields 
terminated by ',' lines terminated by '\012' stored as textfile as select key, 
value from src sort by key, value limit 10
-POSTHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@nzhang_ctas5
-POSTHOOK: Lineage: nzhang_ctas5.key SIMPLE [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
-POSTHOOK: Lineage: nzhang_ctas5.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
-PREHOOK: query: create table nzhang_ctas6 (key string, `to` string)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@nzhang_ctas6
-POSTHOOK: query: create table nzhang_ctas6 (key string, `to` string)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@nzhang_ctas6
-PREHOOK: query: insert overwrite table nzhang_ctas6 select key, value from src 
tablesample (10 rows)
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Output: default@nzhang_ctas6
-POSTHOOK: query: insert overwrite table nzhang_ctas6 select key, value from 
src tablesample (10 rows)
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Output: default@nzhang_ctas6
-POSTHOOK: Lineage: nzhang_ctas6.key SIMPLE [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
-POSTHOOK: Lineage: nzhang_ctas6.to SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
-PREHOOK: query: create table nzhang_ctas7 as select key, `to` from nzhang_ctas6
-PREHOOK: type: CREATETABLE_AS_SELECT
-PREHOOK: Input: default@nzhang_ctas6
-PREHOOK: Output: database:default
-PREHOOK: Output: default@nzhang_ctas7
-POSTHOOK: query: create table nzhang_ctas7 as select key, `to` from 
nzhang_ctas6
-POSTHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: Input: default@nzhang_ctas6
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@nzhang_ctas7
-POSTHOOK: Lineage: nzhang_ctas7.key SIMPLE 
[(nzhang_ctas6)nzhang_ctas6.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: nzhang_ctas7.to SIMPLE 
[(nzhang_ctas6)nzhang_ctas6.FieldSchema(name:to, type:string, comment:null), ]

http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/cte_1.q.out 
b/ql/src/test/results/clientpositive/tez/cte_1.q.out
deleted file mode 100644
index 1b24fb0..0000000
--- a/ql/src/test/results/clientpositive/tez/cte_1.q.out
+++ /dev/null
@@ -1,111 +0,0 @@
-PREHOOK: query: explain
-with q1 as ( select key from src where key = '5')
-select *
-from q1
-PREHOOK: type: QUERY
-POSTHOOK: query: explain
-with q1 as ( select key from src where key = '5')
-select *
-from q1
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Stage-0
-  Fetch Operator
-    limit:-1
-    Select Operator [SEL_2]
-      Output:["_col0"]
-      Filter Operator [FIL_4]
-        predicate:(key = '5')
-        TableScan [TS_0]
-          Output:["key"]
-
-PREHOOK: query: with q1 as ( select key from src where key = '5')
-select *
-from q1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as ( select key from src where key = '5')
-select *
-from q1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-#### A masked pattern was here ####
-5
-5
-5
-PREHOOK: query: -- in subquery
-explain
-with q1 as ( select key from src where key = '5')
-select * from (select key from q1) a
-PREHOOK: type: QUERY
-POSTHOOK: query: -- in subquery
-explain
-with q1 as ( select key from src where key = '5')
-select * from (select key from q1) a
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Stage-0
-  Fetch Operator
-    limit:-1
-    Select Operator [SEL_2]
-      Output:["_col0"]
-      Filter Operator [FIL_4]
-        predicate:(key = '5')
-        TableScan [TS_0]
-          Output:["key"]
-
-PREHOOK: query: with q1 as ( select key from src where key = '5')
-select * from (select key from q1) a
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as ( select key from src where key = '5')
-select * from (select key from q1) a
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-#### A masked pattern was here ####
-5
-5
-5
-PREHOOK: query: -- chaining
-explain
-with q1 as ( select key from q2 where key = '5'),
-q2 as ( select key from src where key = '5')
-select * from (select key from q1) a
-PREHOOK: type: QUERY
-POSTHOOK: query: -- chaining
-explain
-with q1 as ( select key from q2 where key = '5'),
-q2 as ( select key from src where key = '5')
-select * from (select key from q1) a
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Stage-0
-  Fetch Operator
-    limit:-1
-    Select Operator [SEL_2]
-      Output:["_col0"]
-      Filter Operator [FIL_4]
-        predicate:(key = '5')
-        TableScan [TS_0]
-          Output:["key"]
-
-PREHOOK: query: with q1 as ( select key from q2 where key = '5'),
-q2 as ( select key from src where key = '5')
-select * from (select key from q1) a
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as ( select key from q2 where key = '5'),
-q2 as ( select key from src where key = '5')
-select * from (select key from q1) a
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-#### A masked pattern was here ####
-5
-5
-5

http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/cte_2.q.out 
b/ql/src/test/results/clientpositive/tez/cte_2.q.out
deleted file mode 100644
index 017bded..0000000
--- a/ql/src/test/results/clientpositive/tez/cte_2.q.out
+++ /dev/null
@@ -1,190 +0,0 @@
-PREHOOK: query: -- union test
-with q1 as (select * from src where key= '5'),
-q2 as (select * from src s2 where key = '4')
-select * from q1 union all select * from q2
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-#### A masked pattern was here ####
-POSTHOOK: query: -- union test
-with q1 as (select * from src where key= '5'),
-q2 as (select * from src s2 where key = '4')
-select * from q1 union all select * from q2
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-#### A masked pattern was here ####
-5      val_5
-5      val_5
-5      val_5
-4      val_4
-PREHOOK: query: -- insert test
-create table s1 like src
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@s1
-POSTHOOK: query: -- insert test
-create table s1 like src
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@s1
-PREHOOK: query: with q1 as ( select key, value from src where key = '5')
-from q1
-insert overwrite table s1
-select *
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Output: default@s1
-POSTHOOK: query: with q1 as ( select key, value from src where key = '5')
-from q1
-insert overwrite table s1
-select *
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Output: default@s1
-POSTHOOK: Lineage: s1.key SIMPLE []
-POSTHOOK: Lineage: s1.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
-PREHOOK: query: select * from s1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@s1
-#### A masked pattern was here ####
-POSTHOOK: query: select * from s1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@s1
-#### A masked pattern was here ####
-5      val_5
-5      val_5
-5      val_5
-PREHOOK: query: drop table s1
-PREHOOK: type: DROPTABLE
-PREHOOK: Input: default@s1
-PREHOOK: Output: default@s1
-POSTHOOK: query: drop table s1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Input: default@s1
-POSTHOOK: Output: default@s1
-PREHOOK: query: -- from style
-with q1 as (select * from src where key= '5')
-from q1
-select *
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-#### A masked pattern was here ####
-POSTHOOK: query: -- from style
-with q1 as (select * from src where key= '5')
-from q1
-select *
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-#### A masked pattern was here ####
-5      val_5
-5      val_5
-5      val_5
-PREHOOK: query: -- ctas
-create table s2 as
-with q1 as ( select key from src where key = '4')
-select * from q1
-PREHOOK: type: CREATETABLE_AS_SELECT
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@s2
-POSTHOOK: query: -- ctas
-create table s2 as
-with q1 as ( select key from src where key = '4')
-select * from q1
-POSTHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@s2
-POSTHOOK: Lineage: s2.key SIMPLE []
-PREHOOK: query: select * from s2
-PREHOOK: type: QUERY
-PREHOOK: Input: default@s2
-#### A masked pattern was here ####
-POSTHOOK: query: select * from s2
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@s2
-#### A masked pattern was here ####
-4
-PREHOOK: query: drop table s2
-PREHOOK: type: DROPTABLE
-PREHOOK: Input: default@s2
-PREHOOK: Output: default@s2
-POSTHOOK: query: drop table s2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Input: default@s2
-POSTHOOK: Output: default@s2
-PREHOOK: query: -- view test
-create view v1 as
-with q1 as ( select key from src where key = '5')
-select * from q1
-PREHOOK: type: CREATEVIEW
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@v1
-POSTHOOK: query: -- view test
-create view v1 as
-with q1 as ( select key from src where key = '5')
-select * from q1
-POSTHOOK: type: CREATEVIEW
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@v1
-PREHOOK: query: select * from v1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Input: default@v1
-#### A masked pattern was here ####
-POSTHOOK: query: select * from v1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Input: default@v1
-#### A masked pattern was here ####
-5
-5
-5
-PREHOOK: query: drop view v1
-PREHOOK: type: DROPVIEW
-PREHOOK: Input: default@v1
-PREHOOK: Output: default@v1
-POSTHOOK: query: drop view v1
-POSTHOOK: type: DROPVIEW
-POSTHOOK: Input: default@v1
-POSTHOOK: Output: default@v1
-PREHOOK: query: -- view test, name collision
-create view v1 as
-with q1 as ( select key from src where key = '5')
-select * from q1
-PREHOOK: type: CREATEVIEW
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@v1
-POSTHOOK: query: -- view test, name collision
-create view v1 as
-with q1 as ( select key from src where key = '5')
-select * from q1
-POSTHOOK: type: CREATEVIEW
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@v1
-PREHOOK: query: with q1 as ( select key from src where key = '4')
-select * from v1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Input: default@v1
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as ( select key from src where key = '4')
-select * from v1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Input: default@v1
-#### A masked pattern was here ####
-5
-5
-5
-PREHOOK: query: drop view v1
-PREHOOK: type: DROPVIEW
-PREHOOK: Input: default@v1
-PREHOOK: Output: default@v1
-POSTHOOK: query: drop view v1
-POSTHOOK: type: DROPVIEW
-POSTHOOK: Input: default@v1
-POSTHOOK: Output: default@v1

http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/cte_3.q.out 
b/ql/src/test/results/clientpositive/tez/cte_3.q.out
deleted file mode 100644
index cd5776e..0000000
--- a/ql/src/test/results/clientpositive/tez/cte_3.q.out
+++ /dev/null
@@ -1,187 +0,0 @@
-PREHOOK: query: explain
-with q1 as ( select key from src where key = '5')
-select *
-from q1
-PREHOOK: type: QUERY
-POSTHOOK: query: explain
-with q1 as ( select key from src where key = '5')
-select *
-from q1
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Stage-2
-  Dependency Collection{}
-    Stage-1
-      Map 1
-      File Output Operator [FS_3]
-        table:{"name:":"default.q1"}
-        Select Operator [SEL_2] (rows=2 width=85)
-          Output:["_col0"]
-          Filter Operator [FIL_4] (rows=2 width=87)
-            predicate:(key = '5')
-            TableScan [TS_0] (rows=500 width=87)
-              default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"]
-Stage-0
-  Move Operator
-     Please refer to the previous Stage-1
-Stage-3
-  Fetch Operator
-    limit:-1
-    Select Operator [SEL_6]
-      Output:["_col0"]
-      TableScan [TS_5]
-        Output:["key"]
-
-PREHOOK: query: with q1 as ( select key from src where key = '5')
-select *
-from q1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@q1
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@q1
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as ( select key from src where key = '5')
-select *
-from q1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@q1
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@q1
-#### A masked pattern was here ####
-5
-5
-5
-PREHOOK: query: -- in subquery
-explain
-with q1 as ( select key from src where key = '5')
-select * from (select key from q1) a
-PREHOOK: type: QUERY
-POSTHOOK: query: -- in subquery
-explain
-with q1 as ( select key from src where key = '5')
-select * from (select key from q1) a
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Stage-2
-  Dependency Collection{}
-    Stage-1
-      Map 1
-      File Output Operator [FS_3]
-        table:{"name:":"default.q1"}
-        Select Operator [SEL_2] (rows=2 width=85)
-          Output:["_col0"]
-          Filter Operator [FIL_4] (rows=2 width=87)
-            predicate:(key = '5')
-            TableScan [TS_0] (rows=500 width=87)
-              default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"]
-Stage-0
-  Move Operator
-     Please refer to the previous Stage-1
-Stage-3
-  Fetch Operator
-    limit:-1
-    Select Operator [SEL_6]
-      Output:["_col0"]
-      TableScan [TS_5]
-        Output:["key"]
-
-PREHOOK: query: with q1 as ( select key from src where key = '5')
-select * from (select key from q1) a
-PREHOOK: type: QUERY
-PREHOOK: Input: default@q1
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@q1
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as ( select key from src where key = '5')
-select * from (select key from q1) a
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@q1
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@q1
-#### A masked pattern was here ####
-5
-5
-5
-PREHOOK: query: -- chaining
-explain
-with q1 as ( select key from q2 where key = '5'),
-q2 as ( select key from src where key = '5')
-select * from (select key from q1) a
-PREHOOK: type: QUERY
-POSTHOOK: query: -- chaining
-explain
-with q1 as ( select key from q2 where key = '5'),
-q2 as ( select key from src where key = '5')
-select * from (select key from q1) a
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Stage-5
-  Dependency Collection{}
-    Stage-4
-      Map 2
-      File Output Operator [FS_8]
-        table:{"name:":"default.q1"}
-        Select Operator [SEL_7] (rows=1 width=0)
-          Output:["_col0"]
-          Filter Operator [FIL_9] (rows=1 width=0)
-            predicate:(key = '5')
-            TableScan [TS_5] (rows=1 width=0)
-              default@q2,q2,Tbl:PARTIAL,Col:NONE,Output:["key"]
-        Stage-2
-          Dependency Collection{}
-            Stage-1
-              Map 1
-              File Output Operator [FS_3]
-                table:{"name:":"default.q2"}
-                Select Operator [SEL_2] (rows=2 width=85)
-                  Output:["_col0"]
-                  Filter Operator [FIL_4] (rows=2 width=87)
-                    predicate:(key = '5')
-                    TableScan [TS_0] (rows=500 width=87)
-                      default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"]
-        Stage-0
-          Move Operator
-             Please refer to the previous Stage-1
-Stage-3
-  Move Operator
-     Please refer to the previous Stage-4
-Stage-6
-  Fetch Operator
-    limit:-1
-    Select Operator [SEL_11]
-      Output:["_col0"]
-      TableScan [TS_10]
-        Output:["key"]
-
-PREHOOK: query: with q1 as ( select key from q2 where key = '5'),
-q2 as ( select key from src where key = '5')
-select * from (select key from q1) a
-PREHOOK: type: QUERY
-PREHOOK: Input: default@q1
-PREHOOK: Input: default@q2
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@q1
-PREHOOK: Output: default@q2
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as ( select key from q2 where key = '5'),
-q2 as ( select key from src where key = '5')
-select * from (select key from q1) a
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@q1
-POSTHOOK: Input: default@q2
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@q1
-POSTHOOK: Output: default@q2
-#### A masked pattern was here ####
-5
-5
-5

http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_4.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/cte_4.q.out 
b/ql/src/test/results/clientpositive/tez/cte_4.q.out
deleted file mode 100644
index 6385abe..0000000
--- a/ql/src/test/results/clientpositive/tez/cte_4.q.out
+++ /dev/null
@@ -1,220 +0,0 @@
-PREHOOK: query: -- union test
-with q1 as (select * from src where key= '5'),
-q2 as (select * from src s2 where key = '4')
-select * from q1 union all select * from q2
-PREHOOK: type: QUERY
-PREHOOK: Input: default@q1
-PREHOOK: Input: default@q2
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@q1
-PREHOOK: Output: default@q2
-#### A masked pattern was here ####
-POSTHOOK: query: -- union test
-with q1 as (select * from src where key= '5'),
-q2 as (select * from src s2 where key = '4')
-select * from q1 union all select * from q2
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@q1
-POSTHOOK: Input: default@q2
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@q1
-POSTHOOK: Output: default@q2
-#### A masked pattern was here ####
-5      val_5
-5      val_5
-5      val_5
-4      val_4
-PREHOOK: query: -- insert test
-create table s1 like src
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@s1
-POSTHOOK: query: -- insert test
-create table s1 like src
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@s1
-PREHOOK: query: with q1 as ( select key, value from src where key = '5')
-from q1
-insert overwrite table s1
-select *
-PREHOOK: type: QUERY
-PREHOOK: Input: default@q1
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@q1
-PREHOOK: Output: default@s1
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as ( select key, value from src where key = '5')
-from q1
-insert overwrite table s1
-select *
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@q1
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@q1
-POSTHOOK: Output: default@s1
-#### A masked pattern was here ####
-POSTHOOK: Lineage: s1.key SIMPLE [(q1)q1.FieldSchema(name:key, type:string, 
comment:null), ]
-POSTHOOK: Lineage: s1.value SIMPLE [(q1)q1.FieldSchema(name:value, 
type:string, comment:null), ]
-PREHOOK: query: select * from s1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@s1
-#### A masked pattern was here ####
-POSTHOOK: query: select * from s1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@s1
-#### A masked pattern was here ####
-5      val_5
-5      val_5
-5      val_5
-PREHOOK: query: drop table s1
-PREHOOK: type: DROPTABLE
-PREHOOK: Input: default@s1
-PREHOOK: Output: default@s1
-POSTHOOK: query: drop table s1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Input: default@s1
-POSTHOOK: Output: default@s1
-PREHOOK: query: -- from style
-with q1 as (select * from src where key= '5')
-from q1
-select *
-PREHOOK: type: QUERY
-PREHOOK: Input: default@q1
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@q1
-#### A masked pattern was here ####
-POSTHOOK: query: -- from style
-with q1 as (select * from src where key= '5')
-from q1
-select *
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@q1
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@q1
-#### A masked pattern was here ####
-5      val_5
-5      val_5
-5      val_5
-PREHOOK: query: -- ctas
-create table s2 as
-with q1 as ( select key from src where key = '4')
-select * from q1
-PREHOOK: type: CREATETABLE_AS_SELECT
-PREHOOK: Input: default@q1
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@q1
-PREHOOK: Output: default@s2
-#### A masked pattern was here ####
-POSTHOOK: query: -- ctas
-create table s2 as
-with q1 as ( select key from src where key = '4')
-select * from q1
-POSTHOOK: type: CREATETABLE_AS_SELECT
-POSTHOOK: Input: default@q1
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@q1
-POSTHOOK: Output: default@s2
-#### A masked pattern was here ####
-POSTHOOK: Lineage: s2.key SIMPLE [(q1)q1.FieldSchema(name:key, type:string, 
comment:null), ]
-PREHOOK: query: select * from s2
-PREHOOK: type: QUERY
-PREHOOK: Input: default@s2
-#### A masked pattern was here ####
-POSTHOOK: query: select * from s2
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@s2
-#### A masked pattern was here ####
-4
-PREHOOK: query: drop table s2
-PREHOOK: type: DROPTABLE
-PREHOOK: Input: default@s2
-PREHOOK: Output: default@s2
-POSTHOOK: query: drop table s2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Input: default@s2
-POSTHOOK: Output: default@s2
-PREHOOK: query: -- view test
-create view v1 as
-with q1 as ( select key from src where key = '5')
-select * from q1
-PREHOOK: type: CREATEVIEW
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@v1
-POSTHOOK: query: -- view test
-create view v1 as
-with q1 as ( select key from src where key = '5')
-select * from q1
-POSTHOOK: type: CREATEVIEW
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@v1
-PREHOOK: query: select * from v1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Input: default@v1
-#### A masked pattern was here ####
-POSTHOOK: query: select * from v1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Input: default@v1
-#### A masked pattern was here ####
-5
-5
-5
-PREHOOK: query: drop view v1
-PREHOOK: type: DROPVIEW
-PREHOOK: Input: default@v1
-PREHOOK: Output: default@v1
-POSTHOOK: query: drop view v1
-POSTHOOK: type: DROPVIEW
-POSTHOOK: Input: default@v1
-POSTHOOK: Output: default@v1
-PREHOOK: query: -- view test, name collision
-create view v1 as
-with q1 as ( select key from src where key = '5')
-select * from q1
-PREHOOK: type: CREATEVIEW
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@v1
-POSTHOOK: query: -- view test, name collision
-create view v1 as
-with q1 as ( select key from src where key = '5')
-select * from q1
-POSTHOOK: type: CREATEVIEW
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@v1
-PREHOOK: query: with q1 as ( select key from src where key = '4')
-select * from v1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Input: default@v1
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as ( select key from src where key = '4')
-select * from v1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Input: default@v1
-#### A masked pattern was here ####
-5
-5
-5
-PREHOOK: query: drop view v1
-PREHOOK: type: DROPVIEW
-PREHOOK: Input: default@v1
-PREHOOK: Output: default@v1
-POSTHOOK: query: drop view v1
-POSTHOOK: type: DROPVIEW
-POSTHOOK: Input: default@v1
-POSTHOOK: Output: default@v1

http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_5.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/cte_5.q.out 
b/ql/src/test/results/clientpositive/tez/cte_5.q.out
deleted file mode 100644
index d8edfaf..0000000
--- a/ql/src/test/results/clientpositive/tez/cte_5.q.out
+++ /dev/null
@@ -1,128 +0,0 @@
-PREHOOK: query: create database mydb
-PREHOOK: type: CREATEDATABASE
-PREHOOK: Output: database:mydb
-POSTHOOK: query: create database mydb
-POSTHOOK: type: CREATEDATABASE
-POSTHOOK: Output: database:mydb
-PREHOOK: query: use mydb
-PREHOOK: type: SWITCHDATABASE
-PREHOOK: Input: database:mydb
-POSTHOOK: query: use mydb
-POSTHOOK: type: SWITCHDATABASE
-POSTHOOK: Input: database:mydb
-PREHOOK: query: create table q1 (colnum int, colstring string)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:mydb
-PREHOOK: Output: mydb@q1
-POSTHOOK: query: create table q1 (colnum int, colstring string)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:mydb
-POSTHOOK: Output: mydb@q1
-PREHOOK: query: insert into q1 values (5, 'A')
-PREHOOK: type: QUERY
-PREHOOK: Input: mydb@values__tmp__table__1
-PREHOOK: Output: mydb@q1
-POSTHOOK: query: insert into q1 values (5, 'A')
-POSTHOOK: type: QUERY
-POSTHOOK: Input: mydb@values__tmp__table__1
-POSTHOOK: Output: mydb@q1
-POSTHOOK: Lineage: q1.colnum EXPRESSION 
[(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, 
type:string, comment:), ]
-POSTHOOK: Lineage: q1.colstring SIMPLE 
[(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, 
type:string, comment:), ]
-PREHOOK: query: use default
-PREHOOK: type: SWITCHDATABASE
-PREHOOK: Input: database:default
-POSTHOOK: query: use default
-POSTHOOK: type: SWITCHDATABASE
-POSTHOOK: Input: database:default
-PREHOOK: query: show tables in mydb
-PREHOOK: type: SHOWTABLES
-PREHOOK: Input: database:mydb
-POSTHOOK: query: show tables in mydb
-POSTHOOK: type: SHOWTABLES
-POSTHOOK: Input: database:mydb
-q1
-values__tmp__table__1
-PREHOOK: query: show tables
-PREHOOK: type: SHOWTABLES
-PREHOOK: Input: database:default
-POSTHOOK: query: show tables
-POSTHOOK: type: SHOWTABLES
-POSTHOOK: Input: database:default
-alltypesorc
-cbo_t1
-cbo_t2
-cbo_t3
-lineitem
-part
-src
-src1
-src_cbo
-src_json
-src_sequencefile
-src_thrift
-srcbucket
-srcbucket2
-srcpart
-Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 
'Reducer 2' is a cross product
-PREHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.colnum
-from mydb.q1 as a join q1 as b
-on a.colnum=b.key
-PREHOOK: type: QUERY
-POSTHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.colnum
-from mydb.q1 as a join q1 as b
-on a.colnum=b.key
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Vertex dependency in root stage
-Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE)
-
-Stage-0
-  Fetch Operator
-    limit:-1
-    Stage-1
-      Reducer 2
-      File Output Operator [FS_10]
-        Select Operator [SEL_9] (rows=2 width=89)
-          Output:["_col0"]
-          Merge Join Operator [MERGEJOIN_13] (rows=2 width=89)
-            Conds:(Inner)
-          <-Map 1 [SIMPLE_EDGE]
-            SHUFFLE [RS_6]
-              Select Operator [SEL_2] (rows=1 width=3)
-                Filter Operator [FIL_11] (rows=1 width=3)
-                  predicate:(UDFToDouble(colnum) = 5.0)
-                  TableScan [TS_0] (rows=1 width=3)
-                    mydb@q1,a,Tbl:COMPLETE,Col:NONE,Output:["colnum"]
-          <-Map 3 [SIMPLE_EDGE]
-            SHUFFLE [RS_7]
-              Select Operator [SEL_5] (rows=2 width=85)
-                Filter Operator [FIL_12] (rows=2 width=87)
-                  predicate:(key = '5')
-                  TableScan [TS_3] (rows=500 width=87)
-                    default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"]
-
-Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 
'Reducer 2' is a cross product
-PREHOOK: query: with q1 as (select * from src where key= '5')
-select a.colnum
-from mydb.q1 as a join q1 as b
-on a.colnum=b.key
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Input: mydb@q1
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as (select * from src where key= '5')
-select a.colnum
-from mydb.q1 as a join q1 as b
-on a.colnum=b.key
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Input: mydb@q1
-#### A masked pattern was here ####
-5
-5
-5

http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_mat_1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/cte_mat_1.q.out 
b/ql/src/test/results/clientpositive/tez/cte_mat_1.q.out
deleted file mode 100644
index 0c81d95..0000000
--- a/ql/src/test/results/clientpositive/tez/cte_mat_1.q.out
+++ /dev/null
@@ -1,43 +0,0 @@
-Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 
'Reducer 2' is a cross product
-PREHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-PREHOOK: type: QUERY
-POSTHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Vertex dependency in root stage
-Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE)
-
-Stage-0
-  Fetch Operator
-    limit:-1
-    Stage-1
-      Reducer 2
-      File Output Operator [FS_10]
-        Select Operator [SEL_9] (rows=4 width=85)
-          Output:["_col0"]
-          Merge Join Operator [MERGEJOIN_13] (rows=4 width=8)
-            Conds:(Inner)
-          <-Map 1 [SIMPLE_EDGE]
-            SHUFFLE [RS_6]
-              Select Operator [SEL_2] (rows=2 width=85)
-                Filter Operator [FIL_11] (rows=2 width=87)
-                  predicate:(key = '5')
-                  TableScan [TS_0] (rows=500 width=87)
-                    default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"]
-          <-Map 3 [SIMPLE_EDGE]
-            SHUFFLE [RS_7]
-              Select Operator [SEL_5] (rows=2 width=85)
-                Filter Operator [FIL_12] (rows=2 width=87)
-                  predicate:(key = '5')
-                  TableScan [TS_3] (rows=500 width=87)
-                    default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"]
-

http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_mat_2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/cte_mat_2.q.out 
b/ql/src/test/results/clientpositive/tez/cte_mat_2.q.out
deleted file mode 100644
index 0c81d95..0000000
--- a/ql/src/test/results/clientpositive/tez/cte_mat_2.q.out
+++ /dev/null
@@ -1,43 +0,0 @@
-Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 
'Reducer 2' is a cross product
-PREHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-PREHOOK: type: QUERY
-POSTHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Vertex dependency in root stage
-Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE)
-
-Stage-0
-  Fetch Operator
-    limit:-1
-    Stage-1
-      Reducer 2
-      File Output Operator [FS_10]
-        Select Operator [SEL_9] (rows=4 width=85)
-          Output:["_col0"]
-          Merge Join Operator [MERGEJOIN_13] (rows=4 width=8)
-            Conds:(Inner)
-          <-Map 1 [SIMPLE_EDGE]
-            SHUFFLE [RS_6]
-              Select Operator [SEL_2] (rows=2 width=85)
-                Filter Operator [FIL_11] (rows=2 width=87)
-                  predicate:(key = '5')
-                  TableScan [TS_0] (rows=500 width=87)
-                    default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"]
-          <-Map 3 [SIMPLE_EDGE]
-            SHUFFLE [RS_7]
-              Select Operator [SEL_5] (rows=2 width=85)
-                Filter Operator [FIL_12] (rows=2 width=87)
-                  predicate:(key = '5')
-                  TableScan [TS_3] (rows=500 width=87)
-                    default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"]
-

http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_mat_3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/cte_mat_3.q.out 
b/ql/src/test/results/clientpositive/tez/cte_mat_3.q.out
deleted file mode 100644
index 8cd6bad..0000000
--- a/ql/src/test/results/clientpositive/tez/cte_mat_3.q.out
+++ /dev/null
@@ -1,59 +0,0 @@
-PREHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-PREHOOK: type: QUERY
-POSTHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Vertex dependency in root stage
-Reducer 3 <- Map 2 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE)
-
-Stage-3
-  Fetch Operator
-    limit:-1
-    Stage-4
-      Reducer 3
-      File Output Operator [FS_15]
-        Merge Join Operator [MERGEJOIN_20] (rows=1 width=0)
-          Conds:RS_11._col0=RS_12._col0(Inner),Output:["_col0"]
-        <-Map 2 [SIMPLE_EDGE]
-          SHUFFLE [RS_11]
-            PartitionCols:_col0
-            Select Operator [SEL_7] (rows=1 width=0)
-              Output:["_col0"]
-              Filter Operator [FIL_18] (rows=1 width=0)
-                predicate:key is not null
-                TableScan [TS_5] (rows=1 width=0)
-                  default@q1,a,Tbl:PARTIAL,Col:NONE,Output:["key"]
-        <-Map 4 [SIMPLE_EDGE]
-          SHUFFLE [RS_12]
-            PartitionCols:_col0
-            Select Operator [SEL_10] (rows=1 width=0)
-              Output:["_col0"]
-              Filter Operator [FIL_19] (rows=1 width=0)
-                predicate:key is not null
-                TableScan [TS_8] (rows=1 width=0)
-                  default@q1,b,Tbl:PARTIAL,Col:NONE,Output:["key"]
-        Stage-2
-          Dependency Collection{}
-            Stage-1
-              Map 1
-              File Output Operator [FS_3]
-                table:{"name:":"default.q1"}
-                Select Operator [SEL_2] (rows=2 width=176)
-                  Output:["_col0","_col1"]
-                  Filter Operator [FIL_4] (rows=2 width=178)
-                    predicate:(key = '5')
-                    TableScan [TS_0] (rows=500 width=178)
-                      
default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key","value"]
-        Stage-0
-          Move Operator
-             Please refer to the previous Stage-1
-

http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_mat_4.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/cte_mat_4.q.out 
b/ql/src/test/results/clientpositive/tez/cte_mat_4.q.out
deleted file mode 100644
index 8098170..0000000
--- a/ql/src/test/results/clientpositive/tez/cte_mat_4.q.out
+++ /dev/null
@@ -1,301 +0,0 @@
-PREHOOK: query: create temporary table q1 (a int, b string)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@q1
-POSTHOOK: query: create temporary table q1 (a int, b string)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@q1
-PREHOOK: query: insert into q1 values (1, 'A')
-PREHOOK: type: QUERY
-PREHOOK: Input: default@values__tmp__table__1
-PREHOOK: Output: default@q1
-POSTHOOK: query: insert into q1 values (1, 'A')
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@values__tmp__table__1
-POSTHOOK: Output: default@q1
-POSTHOOK: Lineage: q1.a EXPRESSION 
[(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, 
type:string, comment:), ]
-POSTHOOK: Lineage: q1.b SIMPLE 
[(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, 
type:string, comment:), ]
-PREHOOK: query: show tables
-PREHOOK: type: SHOWTABLES
-PREHOOK: Input: database:default
-POSTHOOK: query: show tables
-POSTHOOK: type: SHOWTABLES
-POSTHOOK: Input: database:default
-alltypesorc
-cbo_t1
-cbo_t2
-cbo_t3
-lineitem
-part
-q1
-src
-src1
-src_cbo
-src_json
-src_sequencefile
-src_thrift
-srcbucket
-srcbucket2
-srcpart
-values__tmp__table__1
-PREHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-PREHOOK: type: QUERY
-POSTHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Vertex dependency in root stage
-Reducer 3 <- Map 2 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE)
-
-Stage-3
-  Fetch Operator
-    limit:-1
-    Stage-4
-      Reducer 3
-      File Output Operator [FS_15]
-        Merge Join Operator [MERGEJOIN_20] (rows=1 width=0)
-          Conds:RS_11._col0=RS_12._col0(Inner),Output:["_col0"]
-        <-Map 2 [SIMPLE_EDGE]
-          SHUFFLE [RS_11]
-            PartitionCols:_col0
-            Select Operator [SEL_7] (rows=1 width=0)
-              Output:["_col0"]
-              Filter Operator [FIL_18] (rows=1 width=0)
-                predicate:key is not null
-                TableScan [TS_5] (rows=1 width=0)
-                  default@q1,a,Tbl:PARTIAL,Col:NONE,Output:["key"]
-        <-Map 4 [SIMPLE_EDGE]
-          SHUFFLE [RS_12]
-            PartitionCols:_col0
-            Select Operator [SEL_10] (rows=1 width=0)
-              Output:["_col0"]
-              Filter Operator [FIL_19] (rows=1 width=0)
-                predicate:key is not null
-                TableScan [TS_8] (rows=1 width=0)
-                  default@q1,b,Tbl:PARTIAL,Col:NONE,Output:["key"]
-        Stage-2
-          Dependency Collection{}
-            Stage-1
-              Map 1
-              File Output Operator [FS_3]
-                table:{"name:":"default.q1"}
-                Select Operator [SEL_2] (rows=2 width=176)
-                  Output:["_col0","_col1"]
-                  Filter Operator [FIL_4] (rows=2 width=178)
-                    predicate:(key = '5')
-                    TableScan [TS_0] (rows=500 width=178)
-                      
default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key","value"]
-        Stage-0
-          Move Operator
-             Please refer to the previous Stage-1
-
-PREHOOK: query: with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-PREHOOK: type: QUERY
-PREHOOK: Input: default@q1
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@q1
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@q1
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@q1
-#### A masked pattern was here ####
-5
-5
-5
-5
-5
-5
-5
-5
-5
-PREHOOK: query: show tables
-PREHOOK: type: SHOWTABLES
-PREHOOK: Input: database:default
-POSTHOOK: query: show tables
-POSTHOOK: type: SHOWTABLES
-POSTHOOK: Input: database:default
-alltypesorc
-cbo_t1
-cbo_t2
-cbo_t3
-lineitem
-part
-q1
-src
-src1
-src_cbo
-src_json
-src_sequencefile
-src_thrift
-srcbucket
-srcbucket2
-srcpart
-values__tmp__table__1
-PREHOOK: query: select * from q1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@q1
-#### A masked pattern was here ####
-POSTHOOK: query: select * from q1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@q1
-#### A masked pattern was here ####
-1      A
-PREHOOK: query: drop table q1
-PREHOOK: type: DROPTABLE
-PREHOOK: Input: default@q1
-PREHOOK: Output: default@q1
-POSTHOOK: query: drop table q1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Input: default@q1
-POSTHOOK: Output: default@q1
-PREHOOK: query: show tables
-PREHOOK: type: SHOWTABLES
-PREHOOK: Input: database:default
-POSTHOOK: query: show tables
-POSTHOOK: type: SHOWTABLES
-POSTHOOK: Input: database:default
-alltypesorc
-cbo_t1
-cbo_t2
-cbo_t3
-lineitem
-part
-src
-src1
-src_cbo
-src_json
-src_sequencefile
-src_thrift
-srcbucket
-srcbucket2
-srcpart
-values__tmp__table__1
-PREHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-PREHOOK: type: QUERY
-POSTHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Vertex dependency in root stage
-Reducer 3 <- Map 2 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE)
-
-Stage-3
-  Fetch Operator
-    limit:-1
-    Stage-4
-      Reducer 3
-      File Output Operator [FS_15]
-        Merge Join Operator [MERGEJOIN_20] (rows=1 width=0)
-          Conds:RS_11._col0=RS_12._col0(Inner),Output:["_col0"]
-        <-Map 2 [SIMPLE_EDGE]
-          SHUFFLE [RS_11]
-            PartitionCols:_col0
-            Select Operator [SEL_7] (rows=1 width=0)
-              Output:["_col0"]
-              Filter Operator [FIL_18] (rows=1 width=0)
-                predicate:key is not null
-                TableScan [TS_5] (rows=1 width=0)
-                  default@q1,a,Tbl:PARTIAL,Col:NONE,Output:["key"]
-        <-Map 4 [SIMPLE_EDGE]
-          SHUFFLE [RS_12]
-            PartitionCols:_col0
-            Select Operator [SEL_10] (rows=1 width=0)
-              Output:["_col0"]
-              Filter Operator [FIL_19] (rows=1 width=0)
-                predicate:key is not null
-                TableScan [TS_8] (rows=1 width=0)
-                  default@q1,b,Tbl:PARTIAL,Col:NONE,Output:["key"]
-        Stage-2
-          Dependency Collection{}
-            Stage-1
-              Map 1
-              File Output Operator [FS_3]
-                table:{"name:":"default.q1"}
-                Select Operator [SEL_2] (rows=2 width=176)
-                  Output:["_col0","_col1"]
-                  Filter Operator [FIL_4] (rows=2 width=178)
-                    predicate:(key = '5')
-                    TableScan [TS_0] (rows=500 width=178)
-                      
default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key","value"]
-        Stage-0
-          Move Operator
-             Please refer to the previous Stage-1
-
-PREHOOK: query: with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-PREHOOK: type: QUERY
-PREHOOK: Input: default@q1
-PREHOOK: Input: default@src
-PREHOOK: Output: database:default
-PREHOOK: Output: default@q1
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as (select * from src where key= '5')
-select a.key
-from q1 a join q1 b
-on a.key=b.key
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@q1
-POSTHOOK: Input: default@src
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@q1
-#### A masked pattern was here ####
-5
-5
-5
-5
-5
-5
-5
-5
-5
-PREHOOK: query: show tables
-PREHOOK: type: SHOWTABLES
-PREHOOK: Input: database:default
-POSTHOOK: query: show tables
-POSTHOOK: type: SHOWTABLES
-POSTHOOK: Input: database:default
-alltypesorc
-cbo_t1
-cbo_t2
-cbo_t3
-lineitem
-part
-src
-src1
-src_cbo
-src_json
-src_sequencefile
-src_thrift
-srcbucket
-srcbucket2
-srcpart
-values__tmp__table__1

http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_mat_5.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/cte_mat_5.q.out 
b/ql/src/test/results/clientpositive/tez/cte_mat_5.q.out
deleted file mode 100644
index 9bf7647..0000000
--- a/ql/src/test/results/clientpositive/tez/cte_mat_5.q.out
+++ /dev/null
@@ -1,149 +0,0 @@
-PREHOOK: query: create database mydb
-PREHOOK: type: CREATEDATABASE
-PREHOOK: Output: database:mydb
-POSTHOOK: query: create database mydb
-POSTHOOK: type: CREATEDATABASE
-POSTHOOK: Output: database:mydb
-PREHOOK: query: use mydb
-PREHOOK: type: SWITCHDATABASE
-PREHOOK: Input: database:mydb
-POSTHOOK: query: use mydb
-POSTHOOK: type: SWITCHDATABASE
-POSTHOOK: Input: database:mydb
-PREHOOK: query: create table q1 (colnum int, colstring string)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:mydb
-PREHOOK: Output: mydb@q1
-POSTHOOK: query: create table q1 (colnum int, colstring string)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:mydb
-POSTHOOK: Output: mydb@q1
-PREHOOK: query: insert into q1 values (5, 'A')
-PREHOOK: type: QUERY
-PREHOOK: Input: mydb@values__tmp__table__1
-PREHOOK: Output: mydb@q1
-POSTHOOK: query: insert into q1 values (5, 'A')
-POSTHOOK: type: QUERY
-POSTHOOK: Input: mydb@values__tmp__table__1
-POSTHOOK: Output: mydb@q1
-POSTHOOK: Lineage: q1.colnum EXPRESSION 
[(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, 
type:string, comment:), ]
-POSTHOOK: Lineage: q1.colstring SIMPLE 
[(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, 
type:string, comment:), ]
-PREHOOK: query: use default
-PREHOOK: type: SWITCHDATABASE
-PREHOOK: Input: database:default
-POSTHOOK: query: use default
-POSTHOOK: type: SWITCHDATABASE
-POSTHOOK: Input: database:default
-PREHOOK: query: show tables in mydb
-PREHOOK: type: SHOWTABLES
-PREHOOK: Input: database:mydb
-POSTHOOK: query: show tables in mydb
-POSTHOOK: type: SHOWTABLES
-POSTHOOK: Input: database:mydb
-q1
-values__tmp__table__1
-PREHOOK: query: show tables
-PREHOOK: type: SHOWTABLES
-PREHOOK: Input: database:default
-POSTHOOK: query: show tables
-POSTHOOK: type: SHOWTABLES
-POSTHOOK: Input: database:default
-alltypesorc
-cbo_t1
-cbo_t2
-cbo_t3
-lineitem
-part
-src
-src1
-src_cbo
-src_json
-src_sequencefile
-src_thrift
-srcbucket
-srcbucket2
-srcpart
-PREHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.colnum
-from mydb.q1 as a join q1 as b
-on a.colnum=b.key
-PREHOOK: type: QUERY
-POSTHOOK: query: explain
-with q1 as (select * from src where key= '5')
-select a.colnum
-from mydb.q1 as a join q1 as b
-on a.colnum=b.key
-POSTHOOK: type: QUERY
-Plan optimized by CBO.
-
-Vertex dependency in root stage
-Reducer 3 <- Map 2 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE)
-
-Stage-3
-  Fetch Operator
-    limit:-1
-    Stage-4
-      Reducer 3
-      File Output Operator [FS_15]
-        Merge Join Operator [MERGEJOIN_20] (rows=1 width=3)
-          
Conds:RS_11.UDFToDouble(_col0)=RS_12.UDFToDouble(_col0)(Inner),Output:["_col0"]
-        <-Map 2 [SIMPLE_EDGE]
-          SHUFFLE [RS_11]
-            PartitionCols:UDFToDouble(_col0)
-            Select Operator [SEL_7] (rows=1 width=3)
-              Output:["_col0"]
-              Filter Operator [FIL_18] (rows=1 width=3)
-                predicate:colnum is not null
-                TableScan [TS_5] (rows=1 width=3)
-                  mydb@q1,a,Tbl:COMPLETE,Col:NONE,Output:["colnum"]
-        <-Map 4 [SIMPLE_EDGE]
-          SHUFFLE [RS_12]
-            PartitionCols:UDFToDouble(_col0)
-            Select Operator [SEL_10] (rows=1 width=0)
-              Output:["_col0"]
-              Filter Operator [FIL_19] (rows=1 width=0)
-                predicate:key is not null
-                TableScan [TS_8] (rows=1 width=0)
-                  default@q1,b,Tbl:PARTIAL,Col:NONE,Output:["key"]
-        Stage-2
-          Dependency Collection{}
-            Stage-1
-              Map 1
-              File Output Operator [FS_3]
-                table:{"name:":"default.q1"}
-                Select Operator [SEL_2] (rows=2 width=176)
-                  Output:["_col0","_col1"]
-                  Filter Operator [FIL_4] (rows=2 width=178)
-                    predicate:(key = '5')
-                    TableScan [TS_0] (rows=500 width=178)
-                      
default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key","value"]
-        Stage-0
-          Move Operator
-             Please refer to the previous Stage-1
-
-PREHOOK: query: with q1 as (select * from src where key= '5')
-select a.colnum
-from mydb.q1 as a join q1 as b
-on a.colnum=b.key
-PREHOOK: type: QUERY
-PREHOOK: Input: default@q1
-PREHOOK: Input: default@src
-PREHOOK: Input: mydb@q1
-PREHOOK: Output: database:default
-PREHOOK: Output: default@q1
-#### A masked pattern was here ####
-POSTHOOK: query: with q1 as (select * from src where key= '5')
-select a.colnum
-from mydb.q1 as a join q1 as b
-on a.colnum=b.key
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@q1
-POSTHOOK: Input: default@src
-POSTHOOK: Input: mydb@q1
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@q1
-#### A masked pattern was here ####
-5
-5
-5

Reply via email to