http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/ctas.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/ctas.q.out b/ql/src/test/results/clientpositive/tez/ctas.q.out deleted file mode 100644 index b26f7a2..0000000 --- a/ql/src/test/results/clientpositive/tez/ctas.q.out +++ /dev/null @@ -1,905 +0,0 @@ -PREHOOK: query: -- EXCLUDE_HADOOP_MAJOR_VERSIONS( 0.20S) --- SORT_QUERY_RESULTS - -create table nzhang_Tmp(a int, b string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@nzhang_Tmp -POSTHOOK: query: -- EXCLUDE_HADOOP_MAJOR_VERSIONS( 0.20S) --- SORT_QUERY_RESULTS - -create table nzhang_Tmp(a int, b string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@nzhang_Tmp -PREHOOK: query: select * from nzhang_Tmp -PREHOOK: type: QUERY -PREHOOK: Input: default@nzhang_tmp -#### A masked pattern was here #### -POSTHOOK: query: select * from nzhang_Tmp -POSTHOOK: type: QUERY -POSTHOOK: Input: default@nzhang_tmp -#### A masked pattern was here #### -PREHOOK: query: explain create table nzhang_CTAS1 as select key k, value from src sort by k, value limit 10 -PREHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: query: explain create table nzhang_CTAS1 as select key k, value from src sort by k, value limit 10 -POSTHOOK: type: CREATETABLE_AS_SELECT -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-4 depends on stages: Stage-2, Stage-0 - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Tez -#### A masked pattern was here #### - Edges: - Reducer 2 <- Map 1 (SIMPLE_EDGE) - Reducer 3 <- Reducer 2 (SIMPLE_EDGE) -#### A masked pattern was here #### - Vertices: - Map 1 - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - sort order: ++ - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Reducer 2 - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - sort order: ++ - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Reducer 3 - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_CTAS1 - - Stage: Stage-2 - Dependency Collection - - Stage: Stage-4 - Create Table Operator: - Create Table - columns: k string, value string - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat - serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_CTAS1 - - Stage: Stage-3 - Stats-Aggr Operator - - Stage: Stage-0 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: create table nzhang_CTAS1 as select key k, value from src sort by k, value limit 10 -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@nzhang_CTAS1 -POSTHOOK: query: create table nzhang_CTAS1 as select key k, value from src sort by k, value limit 10 -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@nzhang_CTAS1 -POSTHOOK: Lineage: nzhang_ctas1.k SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: nzhang_ctas1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from nzhang_CTAS1 -PREHOOK: type: QUERY -PREHOOK: Input: default@nzhang_ctas1 -#### A masked pattern was here #### -POSTHOOK: query: select * from nzhang_CTAS1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@nzhang_ctas1 -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -10 val_10 -100 val_100 -100 val_100 -103 val_103 -103 val_103 -104 val_104 -104 val_104 -PREHOOK: query: describe formatted nzhang_CTAS1 -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@nzhang_ctas1 -POSTHOOK: query: describe formatted nzhang_CTAS1 -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@nzhang_ctas1 -# col_name data_type comment - -k string -value string - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 1 - numRows 10 - rawDataSize 96 - totalSize 106 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: explain create table nzhang_ctas2 as select * from src sort by key, value limit 10 -PREHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: query: explain create table nzhang_ctas2 as select * from src sort by key, value limit 10 -POSTHOOK: type: CREATETABLE_AS_SELECT -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-4 depends on stages: Stage-2, Stage-0 - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Tez -#### A masked pattern was here #### - Edges: - Reducer 2 <- Map 1 (SIMPLE_EDGE) - Reducer 3 <- Reducer 2 (SIMPLE_EDGE) -#### A masked pattern was here #### - Vertices: - Map 1 - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - sort order: ++ - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Reducer 2 - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - sort order: ++ - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Reducer 3 - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_ctas2 - - Stage: Stage-2 - Dependency Collection - - Stage: Stage-4 - Create Table Operator: - Create Table - columns: key string, value string - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat - serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_ctas2 - - Stage: Stage-3 - Stats-Aggr Operator - - Stage: Stage-0 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: create table nzhang_ctas2 as select * from src sort by key, value limit 10 -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@nzhang_ctas2 -POSTHOOK: query: create table nzhang_ctas2 as select * from src sort by key, value limit 10 -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@nzhang_ctas2 -POSTHOOK: Lineage: nzhang_ctas2.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: nzhang_ctas2.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from nzhang_ctas2 -PREHOOK: type: QUERY -PREHOOK: Input: default@nzhang_ctas2 -#### A masked pattern was here #### -POSTHOOK: query: select * from nzhang_ctas2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@nzhang_ctas2 -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -10 val_10 -100 val_100 -100 val_100 -103 val_103 -103 val_103 -104 val_104 -104 val_104 -PREHOOK: query: describe formatted nzhang_CTAS2 -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@nzhang_ctas2 -POSTHOOK: query: describe formatted nzhang_CTAS2 -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@nzhang_ctas2 -# col_name data_type comment - -key string -value string - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 1 - numRows 10 - rawDataSize 96 - totalSize 106 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: explain create table nzhang_ctas3 row format serde "org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe" stored as RCFile as select key/2 half_key, concat(value, "_con") conb from src sort by half_key, conb limit 10 -PREHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: query: explain create table nzhang_ctas3 row format serde "org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe" stored as RCFile as select key/2 half_key, concat(value, "_con") conb from src sort by half_key, conb limit 10 -POSTHOOK: type: CREATETABLE_AS_SELECT -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-4 depends on stages: Stage-2, Stage-0 - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Tez -#### A masked pattern was here #### - Edges: - Reducer 2 <- Map 1 (SIMPLE_EDGE) - Reducer 3 <- Reducer 2 (SIMPLE_EDGE) -#### A masked pattern was here #### - Vertices: - Map 1 - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: (key / 2) (type: double), concat(value, '_con') (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 96000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: double), _col1 (type: string) - sort order: ++ - Statistics: Num rows: 500 Data size: 96000 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Reducer 2 - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: double), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 96000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1920 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: double), _col1 (type: string) - sort order: ++ - Statistics: Num rows: 10 Data size: 1920 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Reducer 3 - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: double), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 1920 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1920 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1920 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.nzhang_ctas3 - - Stage: Stage-2 - Dependency Collection - - Stage: Stage-4 - Create Table Operator: - Create Table - columns: half_key double, conb string - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde name: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.nzhang_ctas3 - - Stage: Stage-3 - Stats-Aggr Operator - - Stage: Stage-0 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: create table nzhang_ctas3 row format serde "org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe" stored as RCFile as select key/2 half_key, concat(value, "_con") conb from src sort by half_key, conb limit 10 -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@nzhang_ctas3 -POSTHOOK: query: create table nzhang_ctas3 row format serde "org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe" stored as RCFile as select key/2 half_key, concat(value, "_con") conb from src sort by half_key, conb limit 10 -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@nzhang_ctas3 -POSTHOOK: Lineage: nzhang_ctas3.conb EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: nzhang_ctas3.half_key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: select * from nzhang_ctas3 -PREHOOK: type: QUERY -PREHOOK: Input: default@nzhang_ctas3 -#### A masked pattern was here #### -POSTHOOK: query: select * from nzhang_ctas3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@nzhang_ctas3 -#### A masked pattern was here #### -0.0 val_0_con -0.0 val_0_con -0.0 val_0_con -1.0 val_2_con -2.0 val_4_con -2.5 val_5_con -2.5 val_5_con -2.5 val_5_con -4.0 val_8_con -4.5 val_9_con -PREHOOK: query: describe formatted nzhang_CTAS3 -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@nzhang_ctas3 -POSTHOOK: query: describe formatted nzhang_CTAS3 -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@nzhang_ctas3 -# col_name data_type comment - -half_key double -conb string - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 1 - numRows 10 - rawDataSize 120 - totalSize 199 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: explain create table if not exists nzhang_ctas3 as select key, value from src sort by key, value limit 2 -PREHOOK: type: CREATETABLE -POSTHOOK: query: explain create table if not exists nzhang_ctas3 as select key, value from src sort by key, value limit 2 -POSTHOOK: type: CREATETABLE -STAGE DEPENDENCIES: - -STAGE PLANS: -PREHOOK: query: create table if not exists nzhang_ctas3 as select key, value from src sort by key, value limit 2 -PREHOOK: type: CREATETABLE -POSTHOOK: query: create table if not exists nzhang_ctas3 as select key, value from src sort by key, value limit 2 -POSTHOOK: type: CREATETABLE -PREHOOK: query: select * from nzhang_ctas3 -PREHOOK: type: QUERY -PREHOOK: Input: default@nzhang_ctas3 -#### A masked pattern was here #### -POSTHOOK: query: select * from nzhang_ctas3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@nzhang_ctas3 -#### A masked pattern was here #### -0.0 val_0_con -0.0 val_0_con -0.0 val_0_con -1.0 val_2_con -2.0 val_4_con -2.5 val_5_con -2.5 val_5_con -2.5 val_5_con -4.0 val_8_con -4.5 val_9_con -PREHOOK: query: describe formatted nzhang_CTAS3 -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@nzhang_ctas3 -POSTHOOK: query: describe formatted nzhang_CTAS3 -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@nzhang_ctas3 -# col_name data_type comment - -half_key double -conb string - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 1 - numRows 10 - rawDataSize 120 - totalSize 199 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: explain create table nzhang_ctas4 row format delimited fields terminated by ',' stored as textfile as select key, value from src sort by key, value limit 10 -PREHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: query: explain create table nzhang_ctas4 row format delimited fields terminated by ',' stored as textfile as select key, value from src sort by key, value limit 10 -POSTHOOK: type: CREATETABLE_AS_SELECT -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-4 depends on stages: Stage-2, Stage-0 - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Tez -#### A masked pattern was here #### - Edges: - Reducer 2 <- Map 1 (SIMPLE_EDGE) - Reducer 3 <- Reducer 2 (SIMPLE_EDGE) -#### A masked pattern was here #### - Vertices: - Map 1 - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - sort order: ++ - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Reducer 2 - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - sort order: ++ - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Reducer 3 - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_ctas4 - - Stage: Stage-2 - Dependency Collection - - Stage: Stage-4 - Create Table Operator: - Create Table - columns: key string, value string - field delimiter: , - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat - serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_ctas4 - - Stage: Stage-3 - Stats-Aggr Operator - - Stage: Stage-0 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: create table nzhang_ctas4 row format delimited fields terminated by ',' stored as textfile as select key, value from src sort by key, value limit 10 -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@nzhang_ctas4 -POSTHOOK: query: create table nzhang_ctas4 row format delimited fields terminated by ',' stored as textfile as select key, value from src sort by key, value limit 10 -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@nzhang_ctas4 -POSTHOOK: Lineage: nzhang_ctas4.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: nzhang_ctas4.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from nzhang_ctas4 -PREHOOK: type: QUERY -PREHOOK: Input: default@nzhang_ctas4 -#### A masked pattern was here #### -POSTHOOK: query: select * from nzhang_ctas4 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@nzhang_ctas4 -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -10 val_10 -100 val_100 -100 val_100 -103 val_103 -103 val_103 -104 val_104 -104 val_104 -PREHOOK: query: describe formatted nzhang_CTAS4 -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@nzhang_ctas4 -POSTHOOK: query: describe formatted nzhang_CTAS4 -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@nzhang_ctas4 -# col_name data_type comment - -key string -value string - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 1 - numRows 10 - rawDataSize 96 - totalSize 106 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - field.delim , - serialization.format , -PREHOOK: query: explain extended create table nzhang_ctas5 row format delimited fields terminated by ',' lines terminated by '\012' stored as textfile as select key, value from src sort by key, value limit 10 -PREHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: query: explain extended create table nzhang_ctas5 row format delimited fields terminated by ',' lines terminated by '\012' stored as textfile as select key, value from src sort by key, value limit 10 -POSTHOOK: type: CREATETABLE_AS_SELECT -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-4 depends on stages: Stage-2, Stage-0 - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Tez -#### A masked pattern was here #### - Edges: - Reducer 2 <- Map 1 (SIMPLE_EDGE) - Reducer 3 <- Reducer 2 (SIMPLE_EDGE) -#### A masked pattern was here #### - Vertices: - Map 1 - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: aa - sort order: ++ - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - TopN: 10 - TopN Hash Memory Usage: 0.1 - auto parallelism: true - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src - Truncated Path -> Alias: - /src [src] - Reducer 2 - Needs Tagging: false - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: aa - sort order: ++ - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - TopN: 10 - TopN Hash Memory Usage: 0.1 - auto parallelism: false - Reducer 3 - Needs Tagging: false - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - columns key,value - columns.types string:string - field.delim , - line.delim - - name default.nzhang_ctas5 - serialization.format , - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_ctas5 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - - Stage: Stage-2 - Dependency Collection - - Stage: Stage-4 - Create Table Operator: - Create Table - columns: key string, value string - field delimiter: , - input format: org.apache.hadoop.mapred.TextInputFormat - line delimiter: - - output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat - serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_ctas5 - - Stage: Stage-3 - Stats-Aggr Operator -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: create table nzhang_ctas5 row format delimited fields terminated by ',' lines terminated by '\012' stored as textfile as select key, value from src sort by key, value limit 10 -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@nzhang_ctas5 -POSTHOOK: query: create table nzhang_ctas5 row format delimited fields terminated by ',' lines terminated by '\012' stored as textfile as select key, value from src sort by key, value limit 10 -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@nzhang_ctas5 -POSTHOOK: Lineage: nzhang_ctas5.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: nzhang_ctas5.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: create table nzhang_ctas6 (key string, `to` string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@nzhang_ctas6 -POSTHOOK: query: create table nzhang_ctas6 (key string, `to` string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@nzhang_ctas6 -PREHOOK: query: insert overwrite table nzhang_ctas6 select key, value from src tablesample (10 rows) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@nzhang_ctas6 -POSTHOOK: query: insert overwrite table nzhang_ctas6 select key, value from src tablesample (10 rows) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@nzhang_ctas6 -POSTHOOK: Lineage: nzhang_ctas6.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: nzhang_ctas6.to SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: create table nzhang_ctas7 as select key, `to` from nzhang_ctas6 -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@nzhang_ctas6 -PREHOOK: Output: database:default -PREHOOK: Output: default@nzhang_ctas7 -POSTHOOK: query: create table nzhang_ctas7 as select key, `to` from nzhang_ctas6 -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@nzhang_ctas6 -POSTHOOK: Output: database:default -POSTHOOK: Output: default@nzhang_ctas7 -POSTHOOK: Lineage: nzhang_ctas7.key SIMPLE [(nzhang_ctas6)nzhang_ctas6.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: nzhang_ctas7.to SIMPLE [(nzhang_ctas6)nzhang_ctas6.FieldSchema(name:to, type:string, comment:null), ]
http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_1.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/cte_1.q.out b/ql/src/test/results/clientpositive/tez/cte_1.q.out deleted file mode 100644 index 1b24fb0..0000000 --- a/ql/src/test/results/clientpositive/tez/cte_1.q.out +++ /dev/null @@ -1,111 +0,0 @@ -PREHOOK: query: explain -with q1 as ( select key from src where key = '5') -select * -from q1 -PREHOOK: type: QUERY -POSTHOOK: query: explain -with q1 as ( select key from src where key = '5') -select * -from q1 -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Stage-0 - Fetch Operator - limit:-1 - Select Operator [SEL_2] - Output:["_col0"] - Filter Operator [FIL_4] - predicate:(key = '5') - TableScan [TS_0] - Output:["key"] - -PREHOOK: query: with q1 as ( select key from src where key = '5') -select * -from q1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: with q1 as ( select key from src where key = '5') -select * -from q1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -5 -5 -5 -PREHOOK: query: -- in subquery -explain -with q1 as ( select key from src where key = '5') -select * from (select key from q1) a -PREHOOK: type: QUERY -POSTHOOK: query: -- in subquery -explain -with q1 as ( select key from src where key = '5') -select * from (select key from q1) a -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Stage-0 - Fetch Operator - limit:-1 - Select Operator [SEL_2] - Output:["_col0"] - Filter Operator [FIL_4] - predicate:(key = '5') - TableScan [TS_0] - Output:["key"] - -PREHOOK: query: with q1 as ( select key from src where key = '5') -select * from (select key from q1) a -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: with q1 as ( select key from src where key = '5') -select * from (select key from q1) a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -5 -5 -5 -PREHOOK: query: -- chaining -explain -with q1 as ( select key from q2 where key = '5'), -q2 as ( select key from src where key = '5') -select * from (select key from q1) a -PREHOOK: type: QUERY -POSTHOOK: query: -- chaining -explain -with q1 as ( select key from q2 where key = '5'), -q2 as ( select key from src where key = '5') -select * from (select key from q1) a -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Stage-0 - Fetch Operator - limit:-1 - Select Operator [SEL_2] - Output:["_col0"] - Filter Operator [FIL_4] - predicate:(key = '5') - TableScan [TS_0] - Output:["key"] - -PREHOOK: query: with q1 as ( select key from q2 where key = '5'), -q2 as ( select key from src where key = '5') -select * from (select key from q1) a -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: with q1 as ( select key from q2 where key = '5'), -q2 as ( select key from src where key = '5') -select * from (select key from q1) a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -5 -5 -5 http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_2.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/cte_2.q.out b/ql/src/test/results/clientpositive/tez/cte_2.q.out deleted file mode 100644 index 017bded..0000000 --- a/ql/src/test/results/clientpositive/tez/cte_2.q.out +++ /dev/null @@ -1,190 +0,0 @@ -PREHOOK: query: -- union test -with q1 as (select * from src where key= '5'), -q2 as (select * from src s2 where key = '4') -select * from q1 union all select * from q2 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: -- union test -with q1 as (select * from src where key= '5'), -q2 as (select * from src s2 where key = '4') -select * from q1 union all select * from q2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -5 val_5 -5 val_5 -5 val_5 -4 val_4 -PREHOOK: query: -- insert test -create table s1 like src -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@s1 -POSTHOOK: query: -- insert test -create table s1 like src -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@s1 -PREHOOK: query: with q1 as ( select key, value from src where key = '5') -from q1 -insert overwrite table s1 -select * -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@s1 -POSTHOOK: query: with q1 as ( select key, value from src where key = '5') -from q1 -insert overwrite table s1 -select * -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@s1 -POSTHOOK: Lineage: s1.key SIMPLE [] -POSTHOOK: Lineage: s1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from s1 -PREHOOK: type: QUERY -PREHOOK: Input: default@s1 -#### A masked pattern was here #### -POSTHOOK: query: select * from s1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@s1 -#### A masked pattern was here #### -5 val_5 -5 val_5 -5 val_5 -PREHOOK: query: drop table s1 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@s1 -PREHOOK: Output: default@s1 -POSTHOOK: query: drop table s1 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@s1 -POSTHOOK: Output: default@s1 -PREHOOK: query: -- from style -with q1 as (select * from src where key= '5') -from q1 -select * -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: -- from style -with q1 as (select * from src where key= '5') -from q1 -select * -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -5 val_5 -5 val_5 -5 val_5 -PREHOOK: query: -- ctas -create table s2 as -with q1 as ( select key from src where key = '4') -select * from q1 -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@s2 -POSTHOOK: query: -- ctas -create table s2 as -with q1 as ( select key from src where key = '4') -select * from q1 -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@s2 -POSTHOOK: Lineage: s2.key SIMPLE [] -PREHOOK: query: select * from s2 -PREHOOK: type: QUERY -PREHOOK: Input: default@s2 -#### A masked pattern was here #### -POSTHOOK: query: select * from s2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@s2 -#### A masked pattern was here #### -4 -PREHOOK: query: drop table s2 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@s2 -PREHOOK: Output: default@s2 -POSTHOOK: query: drop table s2 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@s2 -POSTHOOK: Output: default@s2 -PREHOOK: query: -- view test -create view v1 as -with q1 as ( select key from src where key = '5') -select * from q1 -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@v1 -POSTHOOK: query: -- view test -create view v1 as -with q1 as ( select key from src where key = '5') -select * from q1 -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@v1 -PREHOOK: query: select * from v1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@v1 -#### A masked pattern was here #### -POSTHOOK: query: select * from v1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@v1 -#### A masked pattern was here #### -5 -5 -5 -PREHOOK: query: drop view v1 -PREHOOK: type: DROPVIEW -PREHOOK: Input: default@v1 -PREHOOK: Output: default@v1 -POSTHOOK: query: drop view v1 -POSTHOOK: type: DROPVIEW -POSTHOOK: Input: default@v1 -POSTHOOK: Output: default@v1 -PREHOOK: query: -- view test, name collision -create view v1 as -with q1 as ( select key from src where key = '5') -select * from q1 -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@v1 -POSTHOOK: query: -- view test, name collision -create view v1 as -with q1 as ( select key from src where key = '5') -select * from q1 -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@v1 -PREHOOK: query: with q1 as ( select key from src where key = '4') -select * from v1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@v1 -#### A masked pattern was here #### -POSTHOOK: query: with q1 as ( select key from src where key = '4') -select * from v1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@v1 -#### A masked pattern was here #### -5 -5 -5 -PREHOOK: query: drop view v1 -PREHOOK: type: DROPVIEW -PREHOOK: Input: default@v1 -PREHOOK: Output: default@v1 -POSTHOOK: query: drop view v1 -POSTHOOK: type: DROPVIEW -POSTHOOK: Input: default@v1 -POSTHOOK: Output: default@v1 http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_3.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/cte_3.q.out b/ql/src/test/results/clientpositive/tez/cte_3.q.out deleted file mode 100644 index cd5776e..0000000 --- a/ql/src/test/results/clientpositive/tez/cte_3.q.out +++ /dev/null @@ -1,187 +0,0 @@ -PREHOOK: query: explain -with q1 as ( select key from src where key = '5') -select * -from q1 -PREHOOK: type: QUERY -POSTHOOK: query: explain -with q1 as ( select key from src where key = '5') -select * -from q1 -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Stage-2 - Dependency Collection{} - Stage-1 - Map 1 - File Output Operator [FS_3] - table:{"name:":"default.q1"} - Select Operator [SEL_2] (rows=2 width=85) - Output:["_col0"] - Filter Operator [FIL_4] (rows=2 width=87) - predicate:(key = '5') - TableScan [TS_0] (rows=500 width=87) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"] -Stage-0 - Move Operator - Please refer to the previous Stage-1 -Stage-3 - Fetch Operator - limit:-1 - Select Operator [SEL_6] - Output:["_col0"] - TableScan [TS_5] - Output:["key"] - -PREHOOK: query: with q1 as ( select key from src where key = '5') -select * -from q1 -PREHOOK: type: QUERY -PREHOOK: Input: default@q1 -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@q1 -#### A masked pattern was here #### -POSTHOOK: query: with q1 as ( select key from src where key = '5') -select * -from q1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@q1 -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@q1 -#### A masked pattern was here #### -5 -5 -5 -PREHOOK: query: -- in subquery -explain -with q1 as ( select key from src where key = '5') -select * from (select key from q1) a -PREHOOK: type: QUERY -POSTHOOK: query: -- in subquery -explain -with q1 as ( select key from src where key = '5') -select * from (select key from q1) a -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Stage-2 - Dependency Collection{} - Stage-1 - Map 1 - File Output Operator [FS_3] - table:{"name:":"default.q1"} - Select Operator [SEL_2] (rows=2 width=85) - Output:["_col0"] - Filter Operator [FIL_4] (rows=2 width=87) - predicate:(key = '5') - TableScan [TS_0] (rows=500 width=87) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"] -Stage-0 - Move Operator - Please refer to the previous Stage-1 -Stage-3 - Fetch Operator - limit:-1 - Select Operator [SEL_6] - Output:["_col0"] - TableScan [TS_5] - Output:["key"] - -PREHOOK: query: with q1 as ( select key from src where key = '5') -select * from (select key from q1) a -PREHOOK: type: QUERY -PREHOOK: Input: default@q1 -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@q1 -#### A masked pattern was here #### -POSTHOOK: query: with q1 as ( select key from src where key = '5') -select * from (select key from q1) a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@q1 -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@q1 -#### A masked pattern was here #### -5 -5 -5 -PREHOOK: query: -- chaining -explain -with q1 as ( select key from q2 where key = '5'), -q2 as ( select key from src where key = '5') -select * from (select key from q1) a -PREHOOK: type: QUERY -POSTHOOK: query: -- chaining -explain -with q1 as ( select key from q2 where key = '5'), -q2 as ( select key from src where key = '5') -select * from (select key from q1) a -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Stage-5 - Dependency Collection{} - Stage-4 - Map 2 - File Output Operator [FS_8] - table:{"name:":"default.q1"} - Select Operator [SEL_7] (rows=1 width=0) - Output:["_col0"] - Filter Operator [FIL_9] (rows=1 width=0) - predicate:(key = '5') - TableScan [TS_5] (rows=1 width=0) - default@q2,q2,Tbl:PARTIAL,Col:NONE,Output:["key"] - Stage-2 - Dependency Collection{} - Stage-1 - Map 1 - File Output Operator [FS_3] - table:{"name:":"default.q2"} - Select Operator [SEL_2] (rows=2 width=85) - Output:["_col0"] - Filter Operator [FIL_4] (rows=2 width=87) - predicate:(key = '5') - TableScan [TS_0] (rows=500 width=87) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"] - Stage-0 - Move Operator - Please refer to the previous Stage-1 -Stage-3 - Move Operator - Please refer to the previous Stage-4 -Stage-6 - Fetch Operator - limit:-1 - Select Operator [SEL_11] - Output:["_col0"] - TableScan [TS_10] - Output:["key"] - -PREHOOK: query: with q1 as ( select key from q2 where key = '5'), -q2 as ( select key from src where key = '5') -select * from (select key from q1) a -PREHOOK: type: QUERY -PREHOOK: Input: default@q1 -PREHOOK: Input: default@q2 -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@q1 -PREHOOK: Output: default@q2 -#### A masked pattern was here #### -POSTHOOK: query: with q1 as ( select key from q2 where key = '5'), -q2 as ( select key from src where key = '5') -select * from (select key from q1) a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@q1 -POSTHOOK: Input: default@q2 -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@q1 -POSTHOOK: Output: default@q2 -#### A masked pattern was here #### -5 -5 -5 http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_4.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/cte_4.q.out b/ql/src/test/results/clientpositive/tez/cte_4.q.out deleted file mode 100644 index 6385abe..0000000 --- a/ql/src/test/results/clientpositive/tez/cte_4.q.out +++ /dev/null @@ -1,220 +0,0 @@ -PREHOOK: query: -- union test -with q1 as (select * from src where key= '5'), -q2 as (select * from src s2 where key = '4') -select * from q1 union all select * from q2 -PREHOOK: type: QUERY -PREHOOK: Input: default@q1 -PREHOOK: Input: default@q2 -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@q1 -PREHOOK: Output: default@q2 -#### A masked pattern was here #### -POSTHOOK: query: -- union test -with q1 as (select * from src where key= '5'), -q2 as (select * from src s2 where key = '4') -select * from q1 union all select * from q2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@q1 -POSTHOOK: Input: default@q2 -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@q1 -POSTHOOK: Output: default@q2 -#### A masked pattern was here #### -5 val_5 -5 val_5 -5 val_5 -4 val_4 -PREHOOK: query: -- insert test -create table s1 like src -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@s1 -POSTHOOK: query: -- insert test -create table s1 like src -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@s1 -PREHOOK: query: with q1 as ( select key, value from src where key = '5') -from q1 -insert overwrite table s1 -select * -PREHOOK: type: QUERY -PREHOOK: Input: default@q1 -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@q1 -PREHOOK: Output: default@s1 -#### A masked pattern was here #### -POSTHOOK: query: with q1 as ( select key, value from src where key = '5') -from q1 -insert overwrite table s1 -select * -POSTHOOK: type: QUERY -POSTHOOK: Input: default@q1 -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@q1 -POSTHOOK: Output: default@s1 -#### A masked pattern was here #### -POSTHOOK: Lineage: s1.key SIMPLE [(q1)q1.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: s1.value SIMPLE [(q1)q1.FieldSchema(name:value, type:string, comment:null), ] -PREHOOK: query: select * from s1 -PREHOOK: type: QUERY -PREHOOK: Input: default@s1 -#### A masked pattern was here #### -POSTHOOK: query: select * from s1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@s1 -#### A masked pattern was here #### -5 val_5 -5 val_5 -5 val_5 -PREHOOK: query: drop table s1 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@s1 -PREHOOK: Output: default@s1 -POSTHOOK: query: drop table s1 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@s1 -POSTHOOK: Output: default@s1 -PREHOOK: query: -- from style -with q1 as (select * from src where key= '5') -from q1 -select * -PREHOOK: type: QUERY -PREHOOK: Input: default@q1 -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@q1 -#### A masked pattern was here #### -POSTHOOK: query: -- from style -with q1 as (select * from src where key= '5') -from q1 -select * -POSTHOOK: type: QUERY -POSTHOOK: Input: default@q1 -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@q1 -#### A masked pattern was here #### -5 val_5 -5 val_5 -5 val_5 -PREHOOK: query: -- ctas -create table s2 as -with q1 as ( select key from src where key = '4') -select * from q1 -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@q1 -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@q1 -PREHOOK: Output: default@s2 -#### A masked pattern was here #### -POSTHOOK: query: -- ctas -create table s2 as -with q1 as ( select key from src where key = '4') -select * from q1 -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@q1 -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@q1 -POSTHOOK: Output: default@s2 -#### A masked pattern was here #### -POSTHOOK: Lineage: s2.key SIMPLE [(q1)q1.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: select * from s2 -PREHOOK: type: QUERY -PREHOOK: Input: default@s2 -#### A masked pattern was here #### -POSTHOOK: query: select * from s2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@s2 -#### A masked pattern was here #### -4 -PREHOOK: query: drop table s2 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@s2 -PREHOOK: Output: default@s2 -POSTHOOK: query: drop table s2 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@s2 -POSTHOOK: Output: default@s2 -PREHOOK: query: -- view test -create view v1 as -with q1 as ( select key from src where key = '5') -select * from q1 -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@v1 -POSTHOOK: query: -- view test -create view v1 as -with q1 as ( select key from src where key = '5') -select * from q1 -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@v1 -PREHOOK: query: select * from v1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@v1 -#### A masked pattern was here #### -POSTHOOK: query: select * from v1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@v1 -#### A masked pattern was here #### -5 -5 -5 -PREHOOK: query: drop view v1 -PREHOOK: type: DROPVIEW -PREHOOK: Input: default@v1 -PREHOOK: Output: default@v1 -POSTHOOK: query: drop view v1 -POSTHOOK: type: DROPVIEW -POSTHOOK: Input: default@v1 -POSTHOOK: Output: default@v1 -PREHOOK: query: -- view test, name collision -create view v1 as -with q1 as ( select key from src where key = '5') -select * from q1 -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@v1 -POSTHOOK: query: -- view test, name collision -create view v1 as -with q1 as ( select key from src where key = '5') -select * from q1 -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@v1 -PREHOOK: query: with q1 as ( select key from src where key = '4') -select * from v1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@v1 -#### A masked pattern was here #### -POSTHOOK: query: with q1 as ( select key from src where key = '4') -select * from v1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@v1 -#### A masked pattern was here #### -5 -5 -5 -PREHOOK: query: drop view v1 -PREHOOK: type: DROPVIEW -PREHOOK: Input: default@v1 -PREHOOK: Output: default@v1 -POSTHOOK: query: drop view v1 -POSTHOOK: type: DROPVIEW -POSTHOOK: Input: default@v1 -POSTHOOK: Output: default@v1 http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_5.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/cte_5.q.out b/ql/src/test/results/clientpositive/tez/cte_5.q.out deleted file mode 100644 index d8edfaf..0000000 --- a/ql/src/test/results/clientpositive/tez/cte_5.q.out +++ /dev/null @@ -1,128 +0,0 @@ -PREHOOK: query: create database mydb -PREHOOK: type: CREATEDATABASE -PREHOOK: Output: database:mydb -POSTHOOK: query: create database mydb -POSTHOOK: type: CREATEDATABASE -POSTHOOK: Output: database:mydb -PREHOOK: query: use mydb -PREHOOK: type: SWITCHDATABASE -PREHOOK: Input: database:mydb -POSTHOOK: query: use mydb -POSTHOOK: type: SWITCHDATABASE -POSTHOOK: Input: database:mydb -PREHOOK: query: create table q1 (colnum int, colstring string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:mydb -PREHOOK: Output: mydb@q1 -POSTHOOK: query: create table q1 (colnum int, colstring string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:mydb -POSTHOOK: Output: mydb@q1 -PREHOOK: query: insert into q1 values (5, 'A') -PREHOOK: type: QUERY -PREHOOK: Input: mydb@values__tmp__table__1 -PREHOOK: Output: mydb@q1 -POSTHOOK: query: insert into q1 values (5, 'A') -POSTHOOK: type: QUERY -POSTHOOK: Input: mydb@values__tmp__table__1 -POSTHOOK: Output: mydb@q1 -POSTHOOK: Lineage: q1.colnum EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ] -POSTHOOK: Lineage: q1.colstring SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, type:string, comment:), ] -PREHOOK: query: use default -PREHOOK: type: SWITCHDATABASE -PREHOOK: Input: database:default -POSTHOOK: query: use default -POSTHOOK: type: SWITCHDATABASE -POSTHOOK: Input: database:default -PREHOOK: query: show tables in mydb -PREHOOK: type: SHOWTABLES -PREHOOK: Input: database:mydb -POSTHOOK: query: show tables in mydb -POSTHOOK: type: SHOWTABLES -POSTHOOK: Input: database:mydb -q1 -values__tmp__table__1 -PREHOOK: query: show tables -PREHOOK: type: SHOWTABLES -PREHOOK: Input: database:default -POSTHOOK: query: show tables -POSTHOOK: type: SHOWTABLES -POSTHOOK: Input: database:default -alltypesorc -cbo_t1 -cbo_t2 -cbo_t3 -lineitem -part -src -src1 -src_cbo -src_json -src_sequencefile -src_thrift -srcbucket -srcbucket2 -srcpart -Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product -PREHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.colnum -from mydb.q1 as a join q1 as b -on a.colnum=b.key -PREHOOK: type: QUERY -POSTHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.colnum -from mydb.q1 as a join q1 as b -on a.colnum=b.key -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Vertex dependency in root stage -Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) - -Stage-0 - Fetch Operator - limit:-1 - Stage-1 - Reducer 2 - File Output Operator [FS_10] - Select Operator [SEL_9] (rows=2 width=89) - Output:["_col0"] - Merge Join Operator [MERGEJOIN_13] (rows=2 width=89) - Conds:(Inner) - <-Map 1 [SIMPLE_EDGE] - SHUFFLE [RS_6] - Select Operator [SEL_2] (rows=1 width=3) - Filter Operator [FIL_11] (rows=1 width=3) - predicate:(UDFToDouble(colnum) = 5.0) - TableScan [TS_0] (rows=1 width=3) - mydb@q1,a,Tbl:COMPLETE,Col:NONE,Output:["colnum"] - <-Map 3 [SIMPLE_EDGE] - SHUFFLE [RS_7] - Select Operator [SEL_5] (rows=2 width=85) - Filter Operator [FIL_12] (rows=2 width=87) - predicate:(key = '5') - TableScan [TS_3] (rows=500 width=87) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"] - -Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product -PREHOOK: query: with q1 as (select * from src where key= '5') -select a.colnum -from mydb.q1 as a join q1 as b -on a.colnum=b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: mydb@q1 -#### A masked pattern was here #### -POSTHOOK: query: with q1 as (select * from src where key= '5') -select a.colnum -from mydb.q1 as a join q1 as b -on a.colnum=b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: mydb@q1 -#### A masked pattern was here #### -5 -5 -5 http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_mat_1.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/cte_mat_1.q.out b/ql/src/test/results/clientpositive/tez/cte_mat_1.q.out deleted file mode 100644 index 0c81d95..0000000 --- a/ql/src/test/results/clientpositive/tez/cte_mat_1.q.out +++ /dev/null @@ -1,43 +0,0 @@ -Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product -PREHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -PREHOOK: type: QUERY -POSTHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Vertex dependency in root stage -Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) - -Stage-0 - Fetch Operator - limit:-1 - Stage-1 - Reducer 2 - File Output Operator [FS_10] - Select Operator [SEL_9] (rows=4 width=85) - Output:["_col0"] - Merge Join Operator [MERGEJOIN_13] (rows=4 width=8) - Conds:(Inner) - <-Map 1 [SIMPLE_EDGE] - SHUFFLE [RS_6] - Select Operator [SEL_2] (rows=2 width=85) - Filter Operator [FIL_11] (rows=2 width=87) - predicate:(key = '5') - TableScan [TS_0] (rows=500 width=87) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"] - <-Map 3 [SIMPLE_EDGE] - SHUFFLE [RS_7] - Select Operator [SEL_5] (rows=2 width=85) - Filter Operator [FIL_12] (rows=2 width=87) - predicate:(key = '5') - TableScan [TS_3] (rows=500 width=87) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"] - http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_mat_2.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/cte_mat_2.q.out b/ql/src/test/results/clientpositive/tez/cte_mat_2.q.out deleted file mode 100644 index 0c81d95..0000000 --- a/ql/src/test/results/clientpositive/tez/cte_mat_2.q.out +++ /dev/null @@ -1,43 +0,0 @@ -Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product -PREHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -PREHOOK: type: QUERY -POSTHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Vertex dependency in root stage -Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) - -Stage-0 - Fetch Operator - limit:-1 - Stage-1 - Reducer 2 - File Output Operator [FS_10] - Select Operator [SEL_9] (rows=4 width=85) - Output:["_col0"] - Merge Join Operator [MERGEJOIN_13] (rows=4 width=8) - Conds:(Inner) - <-Map 1 [SIMPLE_EDGE] - SHUFFLE [RS_6] - Select Operator [SEL_2] (rows=2 width=85) - Filter Operator [FIL_11] (rows=2 width=87) - predicate:(key = '5') - TableScan [TS_0] (rows=500 width=87) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"] - <-Map 3 [SIMPLE_EDGE] - SHUFFLE [RS_7] - Select Operator [SEL_5] (rows=2 width=85) - Filter Operator [FIL_12] (rows=2 width=87) - predicate:(key = '5') - TableScan [TS_3] (rows=500 width=87) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key"] - http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_mat_3.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/cte_mat_3.q.out b/ql/src/test/results/clientpositive/tez/cte_mat_3.q.out deleted file mode 100644 index 8cd6bad..0000000 --- a/ql/src/test/results/clientpositive/tez/cte_mat_3.q.out +++ /dev/null @@ -1,59 +0,0 @@ -PREHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -PREHOOK: type: QUERY -POSTHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Vertex dependency in root stage -Reducer 3 <- Map 2 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) - -Stage-3 - Fetch Operator - limit:-1 - Stage-4 - Reducer 3 - File Output Operator [FS_15] - Merge Join Operator [MERGEJOIN_20] (rows=1 width=0) - Conds:RS_11._col0=RS_12._col0(Inner),Output:["_col0"] - <-Map 2 [SIMPLE_EDGE] - SHUFFLE [RS_11] - PartitionCols:_col0 - Select Operator [SEL_7] (rows=1 width=0) - Output:["_col0"] - Filter Operator [FIL_18] (rows=1 width=0) - predicate:key is not null - TableScan [TS_5] (rows=1 width=0) - default@q1,a,Tbl:PARTIAL,Col:NONE,Output:["key"] - <-Map 4 [SIMPLE_EDGE] - SHUFFLE [RS_12] - PartitionCols:_col0 - Select Operator [SEL_10] (rows=1 width=0) - Output:["_col0"] - Filter Operator [FIL_19] (rows=1 width=0) - predicate:key is not null - TableScan [TS_8] (rows=1 width=0) - default@q1,b,Tbl:PARTIAL,Col:NONE,Output:["key"] - Stage-2 - Dependency Collection{} - Stage-1 - Map 1 - File Output Operator [FS_3] - table:{"name:":"default.q1"} - Select Operator [SEL_2] (rows=2 width=176) - Output:["_col0","_col1"] - Filter Operator [FIL_4] (rows=2 width=178) - predicate:(key = '5') - TableScan [TS_0] (rows=500 width=178) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key","value"] - Stage-0 - Move Operator - Please refer to the previous Stage-1 - http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_mat_4.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/cte_mat_4.q.out b/ql/src/test/results/clientpositive/tez/cte_mat_4.q.out deleted file mode 100644 index 8098170..0000000 --- a/ql/src/test/results/clientpositive/tez/cte_mat_4.q.out +++ /dev/null @@ -1,301 +0,0 @@ -PREHOOK: query: create temporary table q1 (a int, b string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@q1 -POSTHOOK: query: create temporary table q1 (a int, b string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@q1 -PREHOOK: query: insert into q1 values (1, 'A') -PREHOOK: type: QUERY -PREHOOK: Input: default@values__tmp__table__1 -PREHOOK: Output: default@q1 -POSTHOOK: query: insert into q1 values (1, 'A') -POSTHOOK: type: QUERY -POSTHOOK: Input: default@values__tmp__table__1 -POSTHOOK: Output: default@q1 -POSTHOOK: Lineage: q1.a EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ] -POSTHOOK: Lineage: q1.b SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, type:string, comment:), ] -PREHOOK: query: show tables -PREHOOK: type: SHOWTABLES -PREHOOK: Input: database:default -POSTHOOK: query: show tables -POSTHOOK: type: SHOWTABLES -POSTHOOK: Input: database:default -alltypesorc -cbo_t1 -cbo_t2 -cbo_t3 -lineitem -part -q1 -src -src1 -src_cbo -src_json -src_sequencefile -src_thrift -srcbucket -srcbucket2 -srcpart -values__tmp__table__1 -PREHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -PREHOOK: type: QUERY -POSTHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Vertex dependency in root stage -Reducer 3 <- Map 2 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) - -Stage-3 - Fetch Operator - limit:-1 - Stage-4 - Reducer 3 - File Output Operator [FS_15] - Merge Join Operator [MERGEJOIN_20] (rows=1 width=0) - Conds:RS_11._col0=RS_12._col0(Inner),Output:["_col0"] - <-Map 2 [SIMPLE_EDGE] - SHUFFLE [RS_11] - PartitionCols:_col0 - Select Operator [SEL_7] (rows=1 width=0) - Output:["_col0"] - Filter Operator [FIL_18] (rows=1 width=0) - predicate:key is not null - TableScan [TS_5] (rows=1 width=0) - default@q1,a,Tbl:PARTIAL,Col:NONE,Output:["key"] - <-Map 4 [SIMPLE_EDGE] - SHUFFLE [RS_12] - PartitionCols:_col0 - Select Operator [SEL_10] (rows=1 width=0) - Output:["_col0"] - Filter Operator [FIL_19] (rows=1 width=0) - predicate:key is not null - TableScan [TS_8] (rows=1 width=0) - default@q1,b,Tbl:PARTIAL,Col:NONE,Output:["key"] - Stage-2 - Dependency Collection{} - Stage-1 - Map 1 - File Output Operator [FS_3] - table:{"name:":"default.q1"} - Select Operator [SEL_2] (rows=2 width=176) - Output:["_col0","_col1"] - Filter Operator [FIL_4] (rows=2 width=178) - predicate:(key = '5') - TableScan [TS_0] (rows=500 width=178) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key","value"] - Stage-0 - Move Operator - Please refer to the previous Stage-1 - -PREHOOK: query: with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@q1 -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@q1 -#### A masked pattern was here #### -POSTHOOK: query: with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@q1 -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@q1 -#### A masked pattern was here #### -5 -5 -5 -5 -5 -5 -5 -5 -5 -PREHOOK: query: show tables -PREHOOK: type: SHOWTABLES -PREHOOK: Input: database:default -POSTHOOK: query: show tables -POSTHOOK: type: SHOWTABLES -POSTHOOK: Input: database:default -alltypesorc -cbo_t1 -cbo_t2 -cbo_t3 -lineitem -part -q1 -src -src1 -src_cbo -src_json -src_sequencefile -src_thrift -srcbucket -srcbucket2 -srcpart -values__tmp__table__1 -PREHOOK: query: select * from q1 -PREHOOK: type: QUERY -PREHOOK: Input: default@q1 -#### A masked pattern was here #### -POSTHOOK: query: select * from q1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@q1 -#### A masked pattern was here #### -1 A -PREHOOK: query: drop table q1 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@q1 -PREHOOK: Output: default@q1 -POSTHOOK: query: drop table q1 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@q1 -POSTHOOK: Output: default@q1 -PREHOOK: query: show tables -PREHOOK: type: SHOWTABLES -PREHOOK: Input: database:default -POSTHOOK: query: show tables -POSTHOOK: type: SHOWTABLES -POSTHOOK: Input: database:default -alltypesorc -cbo_t1 -cbo_t2 -cbo_t3 -lineitem -part -src -src1 -src_cbo -src_json -src_sequencefile -src_thrift -srcbucket -srcbucket2 -srcpart -values__tmp__table__1 -PREHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -PREHOOK: type: QUERY -POSTHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Vertex dependency in root stage -Reducer 3 <- Map 2 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) - -Stage-3 - Fetch Operator - limit:-1 - Stage-4 - Reducer 3 - File Output Operator [FS_15] - Merge Join Operator [MERGEJOIN_20] (rows=1 width=0) - Conds:RS_11._col0=RS_12._col0(Inner),Output:["_col0"] - <-Map 2 [SIMPLE_EDGE] - SHUFFLE [RS_11] - PartitionCols:_col0 - Select Operator [SEL_7] (rows=1 width=0) - Output:["_col0"] - Filter Operator [FIL_18] (rows=1 width=0) - predicate:key is not null - TableScan [TS_5] (rows=1 width=0) - default@q1,a,Tbl:PARTIAL,Col:NONE,Output:["key"] - <-Map 4 [SIMPLE_EDGE] - SHUFFLE [RS_12] - PartitionCols:_col0 - Select Operator [SEL_10] (rows=1 width=0) - Output:["_col0"] - Filter Operator [FIL_19] (rows=1 width=0) - predicate:key is not null - TableScan [TS_8] (rows=1 width=0) - default@q1,b,Tbl:PARTIAL,Col:NONE,Output:["key"] - Stage-2 - Dependency Collection{} - Stage-1 - Map 1 - File Output Operator [FS_3] - table:{"name:":"default.q1"} - Select Operator [SEL_2] (rows=2 width=176) - Output:["_col0","_col1"] - Filter Operator [FIL_4] (rows=2 width=178) - predicate:(key = '5') - TableScan [TS_0] (rows=500 width=178) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key","value"] - Stage-0 - Move Operator - Please refer to the previous Stage-1 - -PREHOOK: query: with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@q1 -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@q1 -#### A masked pattern was here #### -POSTHOOK: query: with q1 as (select * from src where key= '5') -select a.key -from q1 a join q1 b -on a.key=b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@q1 -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@q1 -#### A masked pattern was here #### -5 -5 -5 -5 -5 -5 -5 -5 -5 -PREHOOK: query: show tables -PREHOOK: type: SHOWTABLES -PREHOOK: Input: database:default -POSTHOOK: query: show tables -POSTHOOK: type: SHOWTABLES -POSTHOOK: Input: database:default -alltypesorc -cbo_t1 -cbo_t2 -cbo_t3 -lineitem -part -src -src1 -src_cbo -src_json -src_sequencefile -src_thrift -srcbucket -srcbucket2 -srcpart -values__tmp__table__1 http://git-wip-us.apache.org/repos/asf/hive/blob/975a49b6/ql/src/test/results/clientpositive/tez/cte_mat_5.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/cte_mat_5.q.out b/ql/src/test/results/clientpositive/tez/cte_mat_5.q.out deleted file mode 100644 index 9bf7647..0000000 --- a/ql/src/test/results/clientpositive/tez/cte_mat_5.q.out +++ /dev/null @@ -1,149 +0,0 @@ -PREHOOK: query: create database mydb -PREHOOK: type: CREATEDATABASE -PREHOOK: Output: database:mydb -POSTHOOK: query: create database mydb -POSTHOOK: type: CREATEDATABASE -POSTHOOK: Output: database:mydb -PREHOOK: query: use mydb -PREHOOK: type: SWITCHDATABASE -PREHOOK: Input: database:mydb -POSTHOOK: query: use mydb -POSTHOOK: type: SWITCHDATABASE -POSTHOOK: Input: database:mydb -PREHOOK: query: create table q1 (colnum int, colstring string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:mydb -PREHOOK: Output: mydb@q1 -POSTHOOK: query: create table q1 (colnum int, colstring string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:mydb -POSTHOOK: Output: mydb@q1 -PREHOOK: query: insert into q1 values (5, 'A') -PREHOOK: type: QUERY -PREHOOK: Input: mydb@values__tmp__table__1 -PREHOOK: Output: mydb@q1 -POSTHOOK: query: insert into q1 values (5, 'A') -POSTHOOK: type: QUERY -POSTHOOK: Input: mydb@values__tmp__table__1 -POSTHOOK: Output: mydb@q1 -POSTHOOK: Lineage: q1.colnum EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ] -POSTHOOK: Lineage: q1.colstring SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, type:string, comment:), ] -PREHOOK: query: use default -PREHOOK: type: SWITCHDATABASE -PREHOOK: Input: database:default -POSTHOOK: query: use default -POSTHOOK: type: SWITCHDATABASE -POSTHOOK: Input: database:default -PREHOOK: query: show tables in mydb -PREHOOK: type: SHOWTABLES -PREHOOK: Input: database:mydb -POSTHOOK: query: show tables in mydb -POSTHOOK: type: SHOWTABLES -POSTHOOK: Input: database:mydb -q1 -values__tmp__table__1 -PREHOOK: query: show tables -PREHOOK: type: SHOWTABLES -PREHOOK: Input: database:default -POSTHOOK: query: show tables -POSTHOOK: type: SHOWTABLES -POSTHOOK: Input: database:default -alltypesorc -cbo_t1 -cbo_t2 -cbo_t3 -lineitem -part -src -src1 -src_cbo -src_json -src_sequencefile -src_thrift -srcbucket -srcbucket2 -srcpart -PREHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.colnum -from mydb.q1 as a join q1 as b -on a.colnum=b.key -PREHOOK: type: QUERY -POSTHOOK: query: explain -with q1 as (select * from src where key= '5') -select a.colnum -from mydb.q1 as a join q1 as b -on a.colnum=b.key -POSTHOOK: type: QUERY -Plan optimized by CBO. - -Vertex dependency in root stage -Reducer 3 <- Map 2 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) - -Stage-3 - Fetch Operator - limit:-1 - Stage-4 - Reducer 3 - File Output Operator [FS_15] - Merge Join Operator [MERGEJOIN_20] (rows=1 width=3) - Conds:RS_11.UDFToDouble(_col0)=RS_12.UDFToDouble(_col0)(Inner),Output:["_col0"] - <-Map 2 [SIMPLE_EDGE] - SHUFFLE [RS_11] - PartitionCols:UDFToDouble(_col0) - Select Operator [SEL_7] (rows=1 width=3) - Output:["_col0"] - Filter Operator [FIL_18] (rows=1 width=3) - predicate:colnum is not null - TableScan [TS_5] (rows=1 width=3) - mydb@q1,a,Tbl:COMPLETE,Col:NONE,Output:["colnum"] - <-Map 4 [SIMPLE_EDGE] - SHUFFLE [RS_12] - PartitionCols:UDFToDouble(_col0) - Select Operator [SEL_10] (rows=1 width=0) - Output:["_col0"] - Filter Operator [FIL_19] (rows=1 width=0) - predicate:key is not null - TableScan [TS_8] (rows=1 width=0) - default@q1,b,Tbl:PARTIAL,Col:NONE,Output:["key"] - Stage-2 - Dependency Collection{} - Stage-1 - Map 1 - File Output Operator [FS_3] - table:{"name:":"default.q1"} - Select Operator [SEL_2] (rows=2 width=176) - Output:["_col0","_col1"] - Filter Operator [FIL_4] (rows=2 width=178) - predicate:(key = '5') - TableScan [TS_0] (rows=500 width=178) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key","value"] - Stage-0 - Move Operator - Please refer to the previous Stage-1 - -PREHOOK: query: with q1 as (select * from src where key= '5') -select a.colnum -from mydb.q1 as a join q1 as b -on a.colnum=b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@q1 -PREHOOK: Input: default@src -PREHOOK: Input: mydb@q1 -PREHOOK: Output: database:default -PREHOOK: Output: default@q1 -#### A masked pattern was here #### -POSTHOOK: query: with q1 as (select * from src where key= '5') -select a.colnum -from mydb.q1 as a join q1 as b -on a.colnum=b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@q1 -POSTHOOK: Input: default@src -POSTHOOK: Input: mydb@q1 -POSTHOOK: Output: database:default -POSTHOOK: Output: default@q1 -#### A masked pattern was here #### -5 -5 -5
