HIVE-14877: Move slow CliDriver tests to MiniLlap (Prasanth Jayachandran 
reviewed by Siddharth Seth)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/1f258e96
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/1f258e96
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/1f258e96

Branch: refs/heads/master
Commit: 1f258e96aaaba693066fd96e8a2fa97502fd54b6
Parents: 42d997a
Author: Prasanth Jayachandran <prasan...@apache.org>
Authored: Wed Oct 12 00:27:01 2016 -0700
Committer: Prasanth Jayachandran <prasan...@apache.org>
Committed: Wed Oct 12 00:27:01 2016 -0700

----------------------------------------------------------------------
 .../test/resources/testconfiguration.properties |    80 +-
 .../dynpart_sort_optimization_acid.q            |     6 +
 .../clientpositive/insert_into_with_schema.q    |    10 +-
 .../clientpositive/llap/autoColumnStats_1.q.out |  1361 ++
 .../clientpositive/llap/autoColumnStats_2.q.out |  1480 ++
 .../clientpositive/llap/bucket_groupby.q.out    |  1666 ++
 .../llap/bucketsortoptimize_insert_2.q.out      |  1256 ++
 .../clientpositive/llap/cbo_rp_gby.q.out        |   124 +
 .../clientpositive/llap/cbo_rp_join.q.out       | 15028 +++++++++++++
 .../clientpositive/llap/cbo_rp_lineage2.q.out   |   677 +
 .../clientpositive/llap/cbo_rp_semijoin.q.out   |   440 +
 .../llap/cbo_rp_subq_not_in.q.out               |   365 +
 .../llap/cbo_rp_unionDistinct_2.q.out           |   551 +
 .../llap/cbo_rp_windowing_2.q.out               |  2338 ++
 .../llap/correlationoptimizer2.q.out            |  2130 ++
 .../llap/correlationoptimizer4.q.out            |  1922 ++
 .../llap/correlationoptimizer6.q.out            |  3994 ++++
 .../llap/dynpart_sort_optimization_acid.q.out   |  1709 ++
 .../results/clientpositive/llap/escape1.q.out   |   Bin 0 -> 54831 bytes
 .../results/clientpositive/llap/escape2.q.out   |   Bin 0 -> 78516 bytes
 .../clientpositive/llap/global_limit.q.out      |  1677 ++
 .../clientpositive/llap/groupby_sort_1_23.q.out |  4310 ++++
 .../llap/groupby_sort_skew_1_23.q.out           |  4482 ++++
 .../llap/insert_into_with_schema.q.out          |   364 +
 .../results/clientpositive/llap/join43.q.out    |   648 +
 .../clientpositive/llap/join_filters.q.out      |  1484 ++
 .../clientpositive/llap/join_nulls.q.out        |   652 +
 .../llap/limit_join_transpose.q.out             |  1825 ++
 .../results/clientpositive/llap/lineage2.q.out  |   707 +
 .../results/clientpositive/llap/lineage3.q.out  |   388 +
 .../clientpositive/llap/llap_partitioned.q.out  |  2019 ++
 .../clientpositive/llap/load_dyn_part5.q.out    |  1042 +
 .../clientpositive/llap/multiMapJoin1.q.out     |  2145 ++
 .../clientpositive/llap/multiMapJoin2.q.out     |  3187 +++
 ...i_insert_move_tasks_share_dependencies.q.out |  3765 ++++
 .../clientpositive/llap/orc_ppd_date.q.out      |   301 +
 .../clientpositive/llap/orc_ppd_decimal.q.out   |   490 +
 .../clientpositive/llap/orc_ppd_timestamp.q.out |   292 +
 .../llap/parquet_ppd_decimal.q.out              |   858 +
 .../llap/partition_multilevels.q.out            |  1644 ++
 .../clientpositive/llap/rcfile_createas1.q.out  |   180 +
 .../clientpositive/llap/rcfile_merge2.q.out     |   230 +
 .../clientpositive/llap/rcfile_merge3.q.out     |   169 +
 .../clientpositive/llap/rcfile_merge4.q.out     |   169 +
 .../results/clientpositive/llap/sample10.q.out  |   437 +
 .../results/clientpositive/llap/semijoin.q.out  |  2751 +++
 .../llap/special_character_in_tabnames_1.q.out  | 19609 +++++++++++++++++
 .../clientpositive/llap/subquery_notin.q.out    |  1539 ++
 .../llap/table_access_keys_stats.q.out          |   606 +
 .../llap/udaf_collect_set_2.q.out               |   742 +
 .../clientpositive/llap/union_remove_26.q.out   |   849 +
 .../clientpositive/llap/union_top_level.q.out   |  1193 +
 .../llap/vector_number_compare_projection.q.out |   736 +
 .../clientpositive/llap/vector_udf1.q.out       |  1767 ++
 .../results/clientpositive/llap/windowing.q.out |  2338 ++
 .../llap/windowing_windowspec2.q.out            |  3084 +++
 56 files changed, 103810 insertions(+), 6 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hive/blob/1f258e96/itests/src/test/resources/testconfiguration.properties
----------------------------------------------------------------------
diff --git a/itests/src/test/resources/testconfiguration.properties 
b/itests/src/test/resources/testconfiguration.properties
index fbba0cd..97e310d 100644
--- a/itests/src/test/resources/testconfiguration.properties
+++ b/itests/src/test/resources/testconfiguration.properties
@@ -427,13 +427,30 @@ minillap.query.files=acid_bucket_pruning.q,\
   temp_table_external.q,\
   table_nonprintable.q,\
   llap_nullscan.q,\
+  unionDistinct_1.q,\
+  rcfile_merge2.q,\
+  rcfile_merge3.q,\
+  rcfile_merge4.q,\
+  rcfile_createas1.q,\
+  orc_ppd_schema_evol_3a.q,\
+  global_limit.q,\
   dynamic_partition_pruning_2.q,\
   tez_union_dynamic_partition.q,\
   load_fs2.q
 
 minillaplocal.query.files=acid_vectorization_missing_cols.q,\
+  alter_merge_stats_orc.q,\
+  auto_join30.q,\
+  auto_join_filters.q,\
+  auto_join_nulls.q,\
   auto_sortmerge_join_16.q,\
+  auto_sortmerge_join_6.q,\
+  auto_sortmerge_join_8.q,\
+  auto_sortmerge_join_9.q,\
+  autoColumnStats_1.q,\
+  autoColumnStats_2.q,\
   bucket4.q,\
+  bucket_groupby.q,\
   bucket_many.q,\
   bucket_map_join_tez1.q,\
   bucket_map_join_tez2.q,\
@@ -441,9 +458,30 @@ 
minillaplocal.query.files=acid_vectorization_missing_cols.q,\
   bucketmapjoin6.q,\
   bucketmapjoin7.q,\
   bucketpruning1.q,\
+  bucketsortoptimize_insert_2.q,\
+  cbo_gby.q,\
+  cbo_join.q,\
+  cbo_limit.q,\
+  cbo_rp_gby.q,\
+  cbo_rp_join.q,\
+  cbo_rp_lineage2.q,\
+  cbo_rp_semijoin.q,\
+  cbo_rp_subq_not_in.q,\
+  cbo_rp_unionDistinct_2.q,\
+  cbo_rp_windowing_2.q,\
+  cbo_subq_not_in.q,\
   constprog_dpp.q,\
+  correlationoptimizer1.q,\
+  correlationoptimizer2.q,\
+  correlationoptimizer4.q,\
+  correlationoptimizer6.q,\
   disable_merge_for_bucketing.q,\
   dynamic_partition_pruning.q,\
+  dynpart_sort_opt_vectorization.q,\
+  dynpart_sort_optimization.q,\
+  dynpart_sort_optimization_acid.q,\
+  escape1.q,\
+  escape2.q,\
   exchgpartition2lel.q,\
   explainuser_1.q,\
   explainuser_4.q,\
@@ -453,18 +491,37 @@ 
minillaplocal.query.files=acid_vectorization_missing_cols.q,\
   infer_bucket_sort_bucketed_table.q,\
   input16_cc.q,\
   insert_dir_distcp.q,\
+  insert_into_with_schema.q,\
   join1.q,\
   join_acid_non_acid.q,\
+  join_filters.q,\
+  join_nulls.q,\
+  join_nullsafe.q,\
   leftsemijoin_mr.q,\
+  limit_join_transpose.q,\
+  lineage2.q,\
+  lineage3.q,\
   list_bucket_dml_10.q,\
+  llap_partitioned.q,\
+  load_dyn_part5.q,\
   lvj_mapjoin.q,\
   mapjoin_decimal.q,\
   mergejoin_3way.q,\
   mrr.q,\
+  multiMapJoin1.q,\
+  multiMapJoin2.q,\
   non_native_window_udf.q,\
+  orc_analyze.q,\
   orc_llap.q,\
   orc_llap_nonvector.q,\
+  orc_ppd_date.q,\
+  orc_ppd_decimal.q,\
+  orc_ppd_timestamp.q,\
+  partition_multilevels.q,\
+  ptf.q,\
+  ptf_streaming.q,\
   quotedid_smb.q,\
+  sample10.q,\
   schema_evol_orc_acid_part.q,\
   schema_evol_orc_acid_part_update.q,\
   schema_evol_orc_acid_table.q,\
@@ -494,7 +551,12 @@ 
minillaplocal.query.files=acid_vectorization_missing_cols.q,\
   schema_evol_text_vecrow_part_all_complex.q,\
   schema_evol_text_vecrow_part_all_primitive.q,\
   schema_evol_text_vecrow_table.q,\
+  selectDistinctStar.q,\
+  semijoin.q,\
   smb_cache.q,\
+  special_character_in_tabnames_1.q,\
+  subquery_notin.q,\
+  table_access_keys_stats.q,\
   tez_bmj_schema_evolution.q,\
   tez_dml.q,\
   tez_dynpart_hashjoin_1.q,\
@@ -520,8 +582,24 @@ 
minillaplocal.query.files=acid_vectorization_missing_cols.q,\
   tez_vector_dynpart_hashjoin_1.q,\
   tez_vector_dynpart_hashjoin_2.q,\
   uber_reduce.q,\
+  udaf_collect_set_2.q,\
+  union_remove_26.q,\
+  union_top_level.q,\
+  vector_auto_smb_mapjoin_14.q,\
+  vector_decimal_2.q,\
+  vector_decimal_udf.q,\
+  vector_join30.q,\
+  vector_join_filters.q,\
+  vector_leftsemi_mapjoin.q,\
+  vector_number_compare_projection.q,\
+  vector_partitioned_date_time.q,\
+  vector_udf1.q,\
+  vectorization_short_regress.q,\
   vectorized_dynamic_partition_pruning.q,\
-  windowing_gby.q
+  vectorized_ptf.q,\
+  windowing.q,\
+  windowing_gby.q,\
+  windowing_windowspec2.q
 
 encrypted.query.files=encryption_join_unencrypted_tbl.q,\
   encryption_insert_partition_static.q,\

http://git-wip-us.apache.org/repos/asf/hive/blob/1f258e96/ql/src/test/queries/clientpositive/dynpart_sort_optimization_acid.q
----------------------------------------------------------------------
diff --git 
a/ql/src/test/queries/clientpositive/dynpart_sort_optimization_acid.q 
b/ql/src/test/queries/clientpositive/dynpart_sort_optimization_acid.q
index 88ff1e1..414d070 100644
--- a/ql/src/test/queries/clientpositive/dynpart_sort_optimization_acid.q
+++ b/ql/src/test/queries/clientpositive/dynpart_sort_optimization_acid.q
@@ -10,6 +10,8 @@ set hive.optimize.sort.dynamic.partition=false;
 drop table acid;
 CREATE TABLE acid(key string, value string) PARTITIONED BY(ds string) 
CLUSTERED BY(key) INTO 2 BUCKETS STORED AS ORC TBLPROPERTIES 
('transactional'='true');
 insert into table acid partition(ds)  select key,value,ds from srcpart;
+-- explicitly set statistics to avoid flakiness
+alter table acid partition(ds='2008-04-08') update statistics 
set('numRows'='1600', 'rawDataSize'='18000');
 select count(*) from acid where ds='2008-04-08';
 
 insert into table acid partition(ds='2008-04-08') values("foo", "bar");
@@ -32,6 +34,7 @@ set hive.optimize.sort.dynamic.partition=true;
 drop table acid;
 CREATE TABLE acid(key string, value string) PARTITIONED BY(ds string) 
CLUSTERED BY(key) INTO 2 BUCKETS STORED AS ORC TBLPROPERTIES 
('transactional'='true');
 insert into table acid partition(ds)  select key,value,ds from srcpart;
+alter table acid partition(ds='2008-04-08') update statistics 
set('numRows'='1600', 'rawDataSize'='18000');
 select count(*) from acid where ds='2008-04-08';
 
 insert into table acid partition(ds='2008-04-08') values("foo", "bar");
@@ -54,6 +57,7 @@ set hive.optimize.sort.dynamic.partition=false;
 drop table acid;
 CREATE TABLE acid(key string, value string) PARTITIONED BY(ds string, hr int) 
CLUSTERED BY(key) INTO 2 BUCKETS STORED AS ORC TBLPROPERTIES 
('transactional'='true');
 insert into table acid partition(ds,hr)  select * from srcpart;
+alter table acid partition(ds='2008-04-08') update statistics 
set('numRows'='1600', 'rawDataSize'='18000');
 select count(*) from acid where ds='2008-04-08' and hr=11;
 
 insert into table acid partition(ds='2008-04-08',hr=11) values("foo", "bar");
@@ -82,6 +86,7 @@ set hive.optimize.sort.dynamic.partition=true;
 drop table acid;
 CREATE TABLE acid(key string, value string) PARTITIONED BY(ds string, hr int) 
CLUSTERED BY(key) INTO 2 BUCKETS STORED AS ORC TBLPROPERTIES 
('transactional'='true');
 insert into table acid partition(ds,hr)  select * from srcpart;
+alter table acid partition(ds='2008-04-08') update statistics 
set('numRows'='1600', 'rawDataSize'='18000');
 select count(*) from acid where ds='2008-04-08' and hr=11;
 
 insert into table acid partition(ds='2008-04-08',hr=11) values("foo", "bar");
@@ -112,6 +117,7 @@ set hive.optimize.constant.propagation=false;
 drop table acid;
 CREATE TABLE acid(key string, value string) PARTITIONED BY(ds string, hr int) 
CLUSTERED BY(key) INTO 2 BUCKETS STORED AS ORC TBLPROPERTIES 
('transactional'='true');
 insert into table acid partition(ds,hr)  select * from srcpart;
+alter table acid partition(ds='2008-04-08') update statistics 
set('numRows'='1600', 'rawDataSize'='18000');
 select count(*) from acid where ds='2008-04-08' and hr=11;
 
 insert into table acid partition(ds='2008-04-08',hr=11) values("foo", "bar");

http://git-wip-us.apache.org/repos/asf/hive/blob/1f258e96/ql/src/test/queries/clientpositive/insert_into_with_schema.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/insert_into_with_schema.q 
b/ql/src/test/queries/clientpositive/insert_into_with_schema.q
index 62dfcf6..b2d9d06 100644
--- a/ql/src/test/queries/clientpositive/insert_into_with_schema.q
+++ b/ql/src/test/queries/clientpositive/insert_into_with_schema.q
@@ -7,6 +7,7 @@ use x314;
 create table source(s1 int, s2 int);
 create table target1(x int, y int, z int);
 create table target2(x int, y int, z int);
+create table target3(x int, y int, z int);
 
 insert into source(s2,s1) values(2,1);
 -- expect source to contain 1 row (1,2)
@@ -29,15 +30,14 @@ select * from target1 order by x,y,z;
 -- expect target2 to have 2 rows: (2,NULL,1), (NULL, 1,2)
 select * from target2 order by x,y,z;
 
-truncate table target1;
 create table source2(s1 int, s2 int);
-insert into target1 (x,z) select source.s1,source2.s2 from source left outer 
join source2 on source.s1=source2.s2;
---expect target1 to have 1 row (1,NULL,NULL)
-select * from target1;
+insert into target3 (x,z) select source.s1,source2.s2 from source left outer 
join source2 on source.s1=source2.s2;
+--expect target3 to have 1 row (1,NULL,NULL)
+select * from target3;
 
 
 -- partitioned tables
-CREATE TABLE pageviews (userid VARCHAR(64), link STRING, source STRING) 
PARTITIONED BY (datestamp STRING, i int) CLUSTERED BY (userid) INTO 256 BUCKETS 
STORED AS ORC;
+CREATE TABLE pageviews (userid VARCHAR(64), link STRING, source STRING) 
PARTITIONED BY (datestamp STRING, i int) CLUSTERED BY (userid) INTO 4 BUCKETS 
STORED AS ORC;
 INSERT INTO TABLE pageviews PARTITION (datestamp = '2014-09-23', i = 
1)(userid,link) VALUES ('jsmith', 'mail.com');
 -- expect 1 row: ('jsmith', 'mail.com', NULL) in partition '2014-09-23'/'1'
 select * from pageviews;

http://git-wip-us.apache.org/repos/asf/hive/blob/1f258e96/ql/src/test/results/clientpositive/llap/autoColumnStats_1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/autoColumnStats_1.q.out 
b/ql/src/test/results/clientpositive/llap/autoColumnStats_1.q.out
new file mode 100644
index 0000000..712beba
--- /dev/null
+++ b/ql/src/test/results/clientpositive/llap/autoColumnStats_1.q.out
@@ -0,0 +1,1361 @@
+PREHOOK: query: drop table src_multi1
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table src_multi1
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: create table src_multi1 like src
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@src_multi1
+POSTHOOK: query: create table src_multi1 like src
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@src_multi1
+PREHOOK: query: insert overwrite table src_multi1 select * from src
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@src_multi1
+POSTHOOK: query: insert overwrite table src_multi1 select * from src
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@src_multi1
+POSTHOOK: Lineage: src_multi1.key SIMPLE [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
+POSTHOOK: Lineage: src_multi1.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
+PREHOOK: query: explain extended select * from src_multi1
+PREHOOK: type: QUERY
+POSTHOOK: query: explain extended select * from src_multi1
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        TableScan
+          alias: src_multi1
+          GatherStats: false
+          Select Operator
+            expressions: key (type: string), value (type: string)
+            outputColumnNames: _col0, _col1
+            ListSink
+
+PREHOOK: query: describe formatted src_multi1
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@src_multi1
+POSTHOOK: query: describe formatted src_multi1
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@src_multi1
+# col_name             data_type               comment             
+                
+key                    string                  default             
+value                  string                  default             
+                
+# Detailed Table Information            
+Database:              default                  
+#### A masked pattern was here ####
+Retention:             0                        
+#### A masked pattern was here ####
+Table Type:            MANAGED_TABLE            
+Table Parameters:               
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                1                   
+       numRows                 500                 
+       rawDataSize             5312                
+       totalSize               5812                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: drop table a
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table a
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: drop table b
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table b
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: create table a like src
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@a
+POSTHOOK: query: create table a like src
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@a
+PREHOOK: query: create table b like src
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@b
+POSTHOOK: query: create table b like src
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@b
+PREHOOK: query: from src
+insert overwrite table a select *
+insert overwrite table b select *
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@a
+PREHOOK: Output: default@b
+POSTHOOK: query: from src
+insert overwrite table a select *
+insert overwrite table b select *
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@a
+POSTHOOK: Output: default@b
+POSTHOOK: Lineage: a.key SIMPLE [(src)src.FieldSchema(name:key, type:string, 
comment:default), ]
+POSTHOOK: Lineage: a.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
+POSTHOOK: Lineage: b.key SIMPLE [(src)src.FieldSchema(name:key, type:string, 
comment:default), ]
+POSTHOOK: Lineage: b.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
+PREHOOK: query: describe formatted a
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@a
+POSTHOOK: query: describe formatted a
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@a
+# col_name             data_type               comment             
+                
+key                    string                  default             
+value                  string                  default             
+                
+# Detailed Table Information            
+Database:              default                  
+#### A masked pattern was here ####
+Retention:             0                        
+#### A masked pattern was here ####
+Table Type:            MANAGED_TABLE            
+Table Parameters:               
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                1                   
+       numRows                 500                 
+       rawDataSize             5312                
+       totalSize               5812                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: describe formatted b
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@b
+POSTHOOK: query: describe formatted b
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@b
+# col_name             data_type               comment             
+                
+key                    string                  default             
+value                  string                  default             
+                
+# Detailed Table Information            
+Database:              default                  
+#### A masked pattern was here ####
+Retention:             0                        
+#### A masked pattern was here ####
+Table Type:            MANAGED_TABLE            
+Table Parameters:               
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                1                   
+       numRows                 500                 
+       rawDataSize             5312                
+       totalSize               5812                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: drop table a
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@a
+PREHOOK: Output: default@a
+POSTHOOK: query: drop table a
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@a
+POSTHOOK: Output: default@a
+PREHOOK: query: drop table b
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@b
+PREHOOK: Output: default@b
+POSTHOOK: query: drop table b
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@b
+POSTHOOK: Output: default@b
+PREHOOK: query: create table a like src
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@a
+POSTHOOK: query: create table a like src
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@a
+PREHOOK: query: create table b like src
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@b
+POSTHOOK: query: create table b like src
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@b
+PREHOOK: query: from src
+insert overwrite table a select *
+insert into table b select *
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@a
+PREHOOK: Output: default@b
+POSTHOOK: query: from src
+insert overwrite table a select *
+insert into table b select *
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@a
+POSTHOOK: Output: default@b
+POSTHOOK: Lineage: a.key SIMPLE [(src)src.FieldSchema(name:key, type:string, 
comment:default), ]
+POSTHOOK: Lineage: a.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
+POSTHOOK: Lineage: b.key SIMPLE [(src)src.FieldSchema(name:key, type:string, 
comment:default), ]
+POSTHOOK: Lineage: b.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
+PREHOOK: query: describe formatted a
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@a
+POSTHOOK: query: describe formatted a
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@a
+# col_name             data_type               comment             
+                
+key                    string                  default             
+value                  string                  default             
+                
+# Detailed Table Information            
+Database:              default                  
+#### A masked pattern was here ####
+Retention:             0                        
+#### A masked pattern was here ####
+Table Type:            MANAGED_TABLE            
+Table Parameters:               
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                1                   
+       numRows                 500                 
+       rawDataSize             5312                
+       totalSize               5812                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: describe formatted b
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@b
+POSTHOOK: query: describe formatted b
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@b
+# col_name             data_type               comment             
+                
+key                    string                  default             
+value                  string                  default             
+                
+# Detailed Table Information            
+Database:              default                  
+#### A masked pattern was here ####
+Retention:             0                        
+#### A masked pattern was here ####
+Table Type:            MANAGED_TABLE            
+Table Parameters:               
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                1                   
+       numRows                 500                 
+       rawDataSize             5312                
+       totalSize               5812                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: drop table src_multi2
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table src_multi2
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: create table src_multi2 like src
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@src_multi2
+POSTHOOK: query: create table src_multi2 like src
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@src_multi2
+PREHOOK: query: insert overwrite table src_multi2 select subq.key, src.value 
from (select * from src union select * from src1)subq join src on 
subq.key=src.key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Input: default@src1
+PREHOOK: Output: default@src_multi2
+POSTHOOK: query: insert overwrite table src_multi2 select subq.key, src.value 
from (select * from src union select * from src1)subq join src on 
subq.key=src.key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Input: default@src1
+POSTHOOK: Output: default@src_multi2
+POSTHOOK: Lineage: src_multi2.key EXPRESSION [(src)src.FieldSchema(name:key, 
type:string, comment:default), (src1)src1.FieldSchema(name:key, type:string, 
comment:default), ]
+POSTHOOK: Lineage: src_multi2.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
+PREHOOK: query: describe formatted src_multi2
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@src_multi2
+POSTHOOK: query: describe formatted src_multi2
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@src_multi2
+# col_name             data_type               comment             
+                
+key                    string                  default             
+value                  string                  default             
+                
+# Detailed Table Information            
+Database:              default                  
+#### A masked pattern was here ####
+Retention:             0                        
+#### A masked pattern was here ####
+Table Type:            MANAGED_TABLE            
+Table Parameters:               
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                1                   
+       numRows                 508                 
+       rawDataSize             5400                
+       totalSize               5908                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: drop table nzhang_part14
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table nzhang_part14
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: create table if not exists nzhang_part14 (key string)
+  partitioned by (value string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@nzhang_part14
+POSTHOOK: query: create table if not exists nzhang_part14 (key string)
+  partitioned by (value string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@nzhang_part14
+PREHOOK: query: insert overwrite table nzhang_part14 partition(value) 
+select key, value from (
+  select * from (select 'k1' as key, cast(null as string) as value from src 
limit 2)a 
+  union all
+  select * from (select 'k2' as key, '' as value from src limit 2)b
+  union all 
+  select * from (select 'k3' as key, ' ' as value from src limit 2)c
+) T
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@nzhang_part14
+POSTHOOK: query: insert overwrite table nzhang_part14 partition(value) 
+select key, value from (
+  select * from (select 'k1' as key, cast(null as string) as value from src 
limit 2)a 
+  union all
+  select * from (select 'k2' as key, '' as value from src limit 2)b
+  union all 
+  select * from (select 'k3' as key, ' ' as value from src limit 2)c
+) T
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@nzhang_part14@value= 
+POSTHOOK: Output: default@nzhang_part14@value=__HIVE_DEFAULT_PARTITION__
+POSTHOOK: Lineage: nzhang_part14 PARTITION(value= ).key EXPRESSION []
+POSTHOOK: Lineage: nzhang_part14 
PARTITION(value=__HIVE_DEFAULT_PARTITION__).key EXPRESSION []
+PREHOOK: query: explain select key from nzhang_part14
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select key from nzhang_part14
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        TableScan
+          alias: nzhang_part14
+          Select Operator
+            expressions: key (type: string)
+            outputColumnNames: _col0
+            ListSink
+
+PREHOOK: query: drop table src5
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table src5
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: create table src5 as select key, value from src limit 5
+PREHOOK: type: CREATETABLE_AS_SELECT
+PREHOOK: Input: default@src
+PREHOOK: Output: database:default
+PREHOOK: Output: default@src5
+POSTHOOK: query: create table src5 as select key, value from src limit 5
+POSTHOOK: type: CREATETABLE_AS_SELECT
+POSTHOOK: Input: default@src
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@src5
+POSTHOOK: Lineage: src5.key SIMPLE [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
+POSTHOOK: Lineage: src5.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
+PREHOOK: query: insert overwrite table nzhang_part14 partition(value)
+select key, value from src5
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src5
+PREHOOK: Output: default@nzhang_part14
+POSTHOOK: query: insert overwrite table nzhang_part14 partition(value)
+select key, value from src5
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src5
+POSTHOOK: Output: default@nzhang_part14@value=val_165
+POSTHOOK: Output: default@nzhang_part14@value=val_238
+POSTHOOK: Output: default@nzhang_part14@value=val_27
+POSTHOOK: Output: default@nzhang_part14@value=val_311
+POSTHOOK: Output: default@nzhang_part14@value=val_86
+POSTHOOK: Lineage: nzhang_part14 PARTITION(value=val_165).key SIMPLE 
[(src5)src5.FieldSchema(name:key, type:string, comment:null), ]
+POSTHOOK: Lineage: nzhang_part14 PARTITION(value=val_238).key SIMPLE 
[(src5)src5.FieldSchema(name:key, type:string, comment:null), ]
+POSTHOOK: Lineage: nzhang_part14 PARTITION(value=val_27).key SIMPLE 
[(src5)src5.FieldSchema(name:key, type:string, comment:null), ]
+POSTHOOK: Lineage: nzhang_part14 PARTITION(value=val_311).key SIMPLE 
[(src5)src5.FieldSchema(name:key, type:string, comment:null), ]
+POSTHOOK: Lineage: nzhang_part14 PARTITION(value=val_86).key SIMPLE 
[(src5)src5.FieldSchema(name:key, type:string, comment:null), ]
+PREHOOK: query: explain select key from nzhang_part14
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select key from nzhang_part14
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        TableScan
+          alias: nzhang_part14
+          Select Operator
+            expressions: key (type: string)
+            outputColumnNames: _col0
+            ListSink
+
+PREHOOK: query: create table alter5 ( col1 string ) partitioned by (dt string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@alter5
+POSTHOOK: query: create table alter5 ( col1 string ) partitioned by (dt string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@alter5
+PREHOOK: query: alter table alter5 add partition (dt='a') location 'parta'
+PREHOOK: type: ALTERTABLE_ADDPARTS
+#### A masked pattern was here ####
+PREHOOK: Output: default@alter5
+POSTHOOK: query: alter table alter5 add partition (dt='a') location 'parta'
+POSTHOOK: type: ALTERTABLE_ADDPARTS
+#### A masked pattern was here ####
+POSTHOOK: Output: default@alter5
+POSTHOOK: Output: default@alter5@dt=a
+PREHOOK: query: describe formatted alter5 partition (dt='a')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@alter5
+POSTHOOK: query: describe formatted alter5 partition (dt='a')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@alter5
+# col_name             data_type               comment             
+                
+col1                   string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+dt                     string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [a]                      
+Database:              default                  
+Table:                 alter5                   
+#### A masked pattern was here ####
+Partition Parameters:           
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: insert overwrite table alter5 partition (dt='a') select key 
from src
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@alter5@dt=a
+POSTHOOK: query: insert overwrite table alter5 partition (dt='a') select key 
from src
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@alter5@dt=a
+POSTHOOK: Lineage: alter5 PARTITION(dt=a).col1 SIMPLE 
[(src)src.FieldSchema(name:key, type:string, comment:default), ]
+PREHOOK: query: describe formatted alter5 partition (dt='a')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@alter5
+POSTHOOK: query: describe formatted alter5 partition (dt='a')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@alter5
+# col_name             data_type               comment             
+                
+col1                   string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+dt                     string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [a]                      
+Database:              default                  
+Table:                 alter5                   
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"col1\":\"true\"}}
+       numFiles                1                   
+       numRows                 500                 
+       rawDataSize             1406                
+       totalSize               1906                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: explain select * from alter5 where dt='a'
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select * from alter5 where dt='a'
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        TableScan
+          alias: alter5
+          Select Operator
+            expressions: col1 (type: string), 'a' (type: string)
+            outputColumnNames: _col0, _col1
+            ListSink
+
+PREHOOK: query: drop table src_stat_part
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table src_stat_part
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: create table src_stat_part(key string, value string) 
partitioned by (partitionId int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@src_stat_part
+POSTHOOK: query: create table src_stat_part(key string, value string) 
partitioned by (partitionId int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@src_stat_part
+PREHOOK: query: insert overwrite table src_stat_part partition (partitionId=1)
+select * from src1 limit 5
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src1
+PREHOOK: Output: default@src_stat_part@partitionid=1
+POSTHOOK: query: insert overwrite table src_stat_part partition (partitionId=1)
+select * from src1 limit 5
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src1
+POSTHOOK: Output: default@src_stat_part@partitionid=1
+POSTHOOK: Lineage: src_stat_part PARTITION(partitionid=1).key SIMPLE 
[(src1)src1.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: src_stat_part PARTITION(partitionid=1).value SIMPLE 
[(src1)src1.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: describe formatted src_stat_part PARTITION(partitionId=1)
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@src_stat_part
+POSTHOOK: query: describe formatted src_stat_part PARTITION(partitionId=1)
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@src_stat_part
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+partitionid            int                                         
+                
+# Detailed Partition Information                
+Partition Value:       [1]                      
+Database:              default                  
+Table:                 src_stat_part            
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                1                   
+       numRows                 5                   
+       rawDataSize             38                  
+       totalSize               43                  
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: insert overwrite table src_stat_part partition (partitionId=2)
+select * from src1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src1
+PREHOOK: Output: default@src_stat_part@partitionid=2
+POSTHOOK: query: insert overwrite table src_stat_part partition (partitionId=2)
+select * from src1
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src1
+POSTHOOK: Output: default@src_stat_part@partitionid=2
+POSTHOOK: Lineage: src_stat_part PARTITION(partitionid=2).key SIMPLE 
[(src1)src1.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: src_stat_part PARTITION(partitionid=2).value SIMPLE 
[(src1)src1.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: describe formatted src_stat_part PARTITION(partitionId=2)
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@src_stat_part
+POSTHOOK: query: describe formatted src_stat_part PARTITION(partitionId=2)
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@src_stat_part
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+partitionid            int                                         
+                
+# Detailed Partition Information                
+Partition Value:       [2]                      
+Database:              default                  
+Table:                 src_stat_part            
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                1                   
+       numRows                 25                  
+       rawDataSize             191                 
+       totalSize               216                 
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: drop table srcbucket_mapjoin
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table srcbucket_mapjoin
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE srcbucket_mapjoin(key int, value string) 
partitioned by (ds string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@srcbucket_mapjoin
+POSTHOOK: query: CREATE TABLE srcbucket_mapjoin(key int, value string) 
partitioned by (ds string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@srcbucket_mapjoin
+PREHOOK: query: drop table tab_part
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table tab_part
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE tab_part (key int, value string) PARTITIONED 
BY(ds STRING) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS 
TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@tab_part
+POSTHOOK: query: CREATE TABLE tab_part (key int, value string) PARTITIONED 
BY(ds STRING) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS 
TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@tab_part
+PREHOOK: query: drop table srcbucket_mapjoin_part
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table srcbucket_mapjoin_part
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE srcbucket_mapjoin_part (key int, value string) 
partitioned by (ds string) CLUSTERED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@srcbucket_mapjoin_part
+POSTHOOK: query: CREATE TABLE srcbucket_mapjoin_part (key int, value string) 
partitioned by (ds string) CLUSTERED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@srcbucket_mapjoin_part
+PREHOOK: query: load data local inpath '../../data/files/srcbucket20.txt' INTO 
TABLE srcbucket_mapjoin partition(ds='2008-04-08')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@srcbucket_mapjoin
+POSTHOOK: query: load data local inpath '../../data/files/srcbucket20.txt' 
INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@srcbucket_mapjoin
+POSTHOOK: Output: default@srcbucket_mapjoin@ds=2008-04-08
+PREHOOK: query: load data local inpath '../../data/files/srcbucket22.txt' INTO 
TABLE srcbucket_mapjoin partition(ds='2008-04-08')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@srcbucket_mapjoin@ds=2008-04-08
+POSTHOOK: query: load data local inpath '../../data/files/srcbucket22.txt' 
INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@srcbucket_mapjoin@ds=2008-04-08
+PREHOOK: query: load data local inpath '../../data/files/srcbucket20.txt' INTO 
TABLE srcbucket_mapjoin_part partition(ds='2008-04-08')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@srcbucket_mapjoin_part
+POSTHOOK: query: load data local inpath '../../data/files/srcbucket20.txt' 
INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@srcbucket_mapjoin_part
+POSTHOOK: Output: default@srcbucket_mapjoin_part@ds=2008-04-08
+PREHOOK: query: load data local inpath '../../data/files/srcbucket21.txt' INTO 
TABLE srcbucket_mapjoin_part partition(ds='2008-04-08')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@srcbucket_mapjoin_part@ds=2008-04-08
+POSTHOOK: query: load data local inpath '../../data/files/srcbucket21.txt' 
INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@srcbucket_mapjoin_part@ds=2008-04-08
+PREHOOK: query: load data local inpath '../../data/files/srcbucket22.txt' INTO 
TABLE srcbucket_mapjoin_part partition(ds='2008-04-08')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@srcbucket_mapjoin_part@ds=2008-04-08
+POSTHOOK: query: load data local inpath '../../data/files/srcbucket22.txt' 
INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@srcbucket_mapjoin_part@ds=2008-04-08
+PREHOOK: query: load data local inpath '../../data/files/srcbucket23.txt' INTO 
TABLE srcbucket_mapjoin_part partition(ds='2008-04-08')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@srcbucket_mapjoin_part@ds=2008-04-08
+POSTHOOK: query: load data local inpath '../../data/files/srcbucket23.txt' 
INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@srcbucket_mapjoin_part@ds=2008-04-08
+PREHOOK: query: insert overwrite table tab_part partition (ds='2008-04-08')
+select key,value from srcbucket_mapjoin_part
+PREHOOK: type: QUERY
+PREHOOK: Input: default@srcbucket_mapjoin_part
+PREHOOK: Input: default@srcbucket_mapjoin_part@ds=2008-04-08
+PREHOOK: Output: default@tab_part@ds=2008-04-08
+POSTHOOK: query: insert overwrite table tab_part partition (ds='2008-04-08')
+select key,value from srcbucket_mapjoin_part
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@srcbucket_mapjoin_part
+POSTHOOK: Input: default@srcbucket_mapjoin_part@ds=2008-04-08
+POSTHOOK: Output: default@tab_part@ds=2008-04-08
+POSTHOOK: Lineage: tab_part PARTITION(ds=2008-04-08).key SIMPLE 
[(srcbucket_mapjoin_part)srcbucket_mapjoin_part.FieldSchema(name:key, type:int, 
comment:null), ]
+POSTHOOK: Lineage: tab_part PARTITION(ds=2008-04-08).value SIMPLE 
[(srcbucket_mapjoin_part)srcbucket_mapjoin_part.FieldSchema(name:value, 
type:string, comment:null), ]
+PREHOOK: query: describe formatted tab_part partition (ds='2008-04-08')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@tab_part
+POSTHOOK: query: describe formatted tab_part partition (ds='2008-04-08')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@tab_part
+# col_name             data_type               comment             
+                
+key                    int                                         
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+ds                     string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [2008-04-08]             
+Database:              default                  
+Table:                 tab_part                 
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                4                   
+       numRows                 500                 
+       rawDataSize             5312                
+       totalSize               5812                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           4                        
+Bucket Columns:        [key]                    
+Sort Columns:          [Order(col:key, order:1)]        
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: CREATE TABLE tab(key int, value string) PARTITIONED BY(ds 
STRING) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@tab
+POSTHOOK: query: CREATE TABLE tab(key int, value string) PARTITIONED BY(ds 
STRING) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@tab
+PREHOOK: query: insert overwrite table tab partition (ds='2008-04-08')
+select key,value from srcbucket_mapjoin
+PREHOOK: type: QUERY
+PREHOOK: Input: default@srcbucket_mapjoin
+PREHOOK: Input: default@srcbucket_mapjoin@ds=2008-04-08
+PREHOOK: Output: default@tab@ds=2008-04-08
+POSTHOOK: query: insert overwrite table tab partition (ds='2008-04-08')
+select key,value from srcbucket_mapjoin
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@srcbucket_mapjoin
+POSTHOOK: Input: default@srcbucket_mapjoin@ds=2008-04-08
+POSTHOOK: Output: default@tab@ds=2008-04-08
+POSTHOOK: Lineage: tab PARTITION(ds=2008-04-08).key SIMPLE 
[(srcbucket_mapjoin)srcbucket_mapjoin.FieldSchema(name:key, type:int, 
comment:null), ]
+POSTHOOK: Lineage: tab PARTITION(ds=2008-04-08).value SIMPLE 
[(srcbucket_mapjoin)srcbucket_mapjoin.FieldSchema(name:value, type:string, 
comment:null), ]
+PREHOOK: query: describe formatted tab partition (ds='2008-04-08')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@tab
+POSTHOOK: query: describe formatted tab partition (ds='2008-04-08')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@tab
+# col_name             data_type               comment             
+                
+key                    int                                         
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+ds                     string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [2008-04-08]             
+Database:              default                  
+Table:                 tab                      
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                2                   
+       numRows                 242                 
+       rawDataSize             2566                
+       totalSize               2808                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           2                        
+Bucket Columns:        [key]                    
+Sort Columns:          [Order(col:key, order:1)]        
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: drop table nzhang_part14
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@nzhang_part14
+PREHOOK: Output: default@nzhang_part14
+POSTHOOK: query: drop table nzhang_part14
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@nzhang_part14
+POSTHOOK: Output: default@nzhang_part14
+PREHOOK: query: create table if not exists nzhang_part14 (key string, value 
string)
+  partitioned by (ds string, hr string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@nzhang_part14
+POSTHOOK: query: create table if not exists nzhang_part14 (key string, value 
string)
+  partitioned by (ds string, hr string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@nzhang_part14
+PREHOOK: query: describe formatted nzhang_part14
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@nzhang_part14
+POSTHOOK: query: describe formatted nzhang_part14
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@nzhang_part14
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+ds                     string                                      
+hr                     string                                      
+                
+# Detailed Table Information            
+Database:              default                  
+#### A masked pattern was here ####
+Retention:             0                        
+#### A masked pattern was here ####
+Table Type:            MANAGED_TABLE            
+Table Parameters:               
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: insert overwrite table nzhang_part14 partition(ds, hr) 
+select key, value, ds, hr from (
+  select * from (select 'k1' as key, cast(null as string) as value, '1' as ds, 
'2' as hr from src limit 2)a 
+  union all
+  select * from (select 'k2' as key, '' as value, '1' as ds, '3' as hr from 
src limit 2)b
+  union all 
+  select * from (select 'k3' as key, ' ' as value, '2' as ds, '1' as hr from 
src limit 2)c
+) T
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@nzhang_part14
+POSTHOOK: query: insert overwrite table nzhang_part14 partition(ds, hr) 
+select key, value, ds, hr from (
+  select * from (select 'k1' as key, cast(null as string) as value, '1' as ds, 
'2' as hr from src limit 2)a 
+  union all
+  select * from (select 'k2' as key, '' as value, '1' as ds, '3' as hr from 
src limit 2)b
+  union all 
+  select * from (select 'k3' as key, ' ' as value, '2' as ds, '1' as hr from 
src limit 2)c
+) T
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@nzhang_part14@ds=1/hr=2
+POSTHOOK: Output: default@nzhang_part14@ds=1/hr=3
+POSTHOOK: Output: default@nzhang_part14@ds=2/hr=1
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=1,hr=2).key EXPRESSION []
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=1,hr=2).value EXPRESSION []
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=1,hr=3).key EXPRESSION []
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=1,hr=3).value EXPRESSION []
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=2,hr=1).key EXPRESSION []
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=2,hr=1).value EXPRESSION []
+PREHOOK: query: desc formatted nzhang_part14 partition(ds='1', hr='3')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@nzhang_part14
+POSTHOOK: query: desc formatted nzhang_part14 partition(ds='1', hr='3')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@nzhang_part14
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+ds                     string                                      
+hr                     string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [1, 3]                   
+Database:              default                  
+Table:                 nzhang_part14            
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                1                   
+       numRows                 2                   
+       rawDataSize             6                   
+       totalSize               8                   
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: INSERT OVERWRITE TABLE nzhang_part14 PARTITION 
(ds='2010-03-03', hr)
+SELECT key, value, hr FROM srcpart WHERE ds is not null and hr>10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@srcpart
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+PREHOOK: Output: default@nzhang_part14@ds=2010-03-03
+POSTHOOK: query: INSERT OVERWRITE TABLE nzhang_part14 PARTITION 
(ds='2010-03-03', hr)
+SELECT key, value, hr FROM srcpart WHERE ds is not null and hr>10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@srcpart
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+POSTHOOK: Output: default@nzhang_part14@ds=2010-03-03/hr=11
+POSTHOOK: Output: default@nzhang_part14@ds=2010-03-03/hr=12
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=2010-03-03,hr=11).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=2010-03-03,hr=11).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=2010-03-03,hr=12).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=2010-03-03,hr=12).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: desc formatted nzhang_part14 PARTITION(ds='2010-03-03', 
hr='12')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@nzhang_part14
+POSTHOOK: query: desc formatted nzhang_part14 PARTITION(ds='2010-03-03', 
hr='12')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@nzhang_part14
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+ds                     string                                      
+hr                     string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [2010-03-03, 12]         
+Database:              default                  
+Table:                 nzhang_part14            
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                1                   
+       numRows                 1000                
+       rawDataSize             10624               
+       totalSize               11624               
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: drop table nzhang_part14
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@nzhang_part14
+PREHOOK: Output: default@nzhang_part14
+POSTHOOK: query: drop table nzhang_part14
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@nzhang_part14
+POSTHOOK: Output: default@nzhang_part14
+PREHOOK: query: create table if not exists nzhang_part14 (key string, value 
string)
+partitioned by (ds string, hr string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@nzhang_part14
+POSTHOOK: query: create table if not exists nzhang_part14 (key string, value 
string)
+partitioned by (ds string, hr string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@nzhang_part14
+PREHOOK: query: INSERT OVERWRITE TABLE nzhang_part14 PARTITION 
(ds='2010-03-03', hr)
+SELECT key, value, hr FROM srcpart WHERE ds is not null and hr>10
+PREHOOK: type: QUERY
+PREHOOK: Input: default@srcpart
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+PREHOOK: Output: default@nzhang_part14@ds=2010-03-03
+POSTHOOK: query: INSERT OVERWRITE TABLE nzhang_part14 PARTITION 
(ds='2010-03-03', hr)
+SELECT key, value, hr FROM srcpart WHERE ds is not null and hr>10
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@srcpart
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+POSTHOOK: Output: default@nzhang_part14@ds=2010-03-03/hr=11
+POSTHOOK: Output: default@nzhang_part14@ds=2010-03-03/hr=12
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=2010-03-03,hr=11).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=2010-03-03,hr=11).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=2010-03-03,hr=12).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: nzhang_part14 PARTITION(ds=2010-03-03,hr=12).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: desc formatted nzhang_part14 PARTITION(ds='2010-03-03', 
hr='12')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@nzhang_part14
+POSTHOOK: query: desc formatted nzhang_part14 PARTITION(ds='2010-03-03', 
hr='12')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@nzhang_part14
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+ds                     string                                      
+hr                     string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [2010-03-03, 12]         
+Database:              default                  
+Table:                 nzhang_part14            
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+       numFiles                1                   
+       numRows                 1000                
+       rawDataSize             10624               
+       totalSize               11624               
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: drop table a
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@a
+PREHOOK: Output: default@a
+POSTHOOK: query: drop table a
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@a
+POSTHOOK: Output: default@a
+PREHOOK: query: create table a (key string, value string)
+partitioned by (ds string, hr string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@a
+POSTHOOK: query: create table a (key string, value string)
+partitioned by (ds string, hr string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@a
+PREHOOK: query: drop table b
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@b
+PREHOOK: Output: default@b
+POSTHOOK: query: drop table b
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@b
+POSTHOOK: Output: default@b
+PREHOOK: query: create table b (key string, value string)
+partitioned by (ds string, hr string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@b
+POSTHOOK: query: create table b (key string, value string)
+partitioned by (ds string, hr string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@b
+PREHOOK: query: drop table c
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table c
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: create table c (key string, value string)
+partitioned by (ds string, hr string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@c
+POSTHOOK: query: create table c (key string, value string)
+partitioned by (ds string, hr string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@c
+PREHOOK: query: FROM srcpart 
+INSERT OVERWRITE TABLE a PARTITION (ds='2010-03-11', hr) SELECT key, value, hr 
WHERE ds is not null and hr>10
+INSERT OVERWRITE TABLE b PARTITION (ds='2010-04-11', hr) SELECT key, value, hr 
WHERE ds is not null and hr>11
+INSERT OVERWRITE TABLE c PARTITION (ds='2010-05-11', hr) SELECT key, value, hr 
WHERE hr>0
+PREHOOK: type: QUERY
+PREHOOK: Input: default@srcpart
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+PREHOOK: Output: default@a@ds=2010-03-11
+PREHOOK: Output: default@b@ds=2010-04-11
+PREHOOK: Output: default@c@ds=2010-05-11
+POSTHOOK: query: FROM srcpart 
+INSERT OVERWRITE TABLE a PARTITION (ds='2010-03-11', hr) SELECT key, value, hr 
WHERE ds is not null and hr>10
+INSERT OVERWRITE TABLE b PARTITION (ds='2010-04-11', hr) SELECT key, value, hr 
WHERE ds is not null and hr>11
+INSERT OVERWRITE TABLE c PARTITION (ds='2010-05-11', hr) SELECT key, value, hr 
WHERE hr>0
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@srcpart
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
+POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
+POSTHOOK: Output: default@a@ds=2010-03-11/hr=11
+POSTHOOK: Output: default@a@ds=2010-03-11/hr=12
+POSTHOOK: Output: default@b@ds=2010-04-11/hr=12
+POSTHOOK: Output: default@c@ds=2010-05-11/hr=11
+POSTHOOK: Output: default@c@ds=2010-05-11/hr=12
+POSTHOOK: Lineage: a PARTITION(ds=2010-03-11,hr=11).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: a PARTITION(ds=2010-03-11,hr=11).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: a PARTITION(ds=2010-03-11,hr=12).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: a PARTITION(ds=2010-03-11,hr=12).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: b PARTITION(ds=2010-04-11,hr=12).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: b PARTITION(ds=2010-04-11,hr=12).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: c PARTITION(ds=2010-05-11,hr=11).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: c PARTITION(ds=2010-05-11,hr=11).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
+POSTHOOK: Lineage: c PARTITION(ds=2010-05-11,hr=12).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: c PARTITION(ds=2010-05-11,hr=12).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: explain select key from a
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select key from a
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        TableScan
+          alias: a
+          Select Operator
+            expressions: key (type: string)
+            outputColumnNames: _col0
+            ListSink
+
+PREHOOK: query: explain select value from b
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select value from b
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        TableScan
+          alias: b
+          Select Operator
+            expressions: value (type: string)
+            outputColumnNames: _col0
+            ListSink
+
+PREHOOK: query: explain select key from b
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select key from b
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        TableScan
+          alias: b
+          Select Operator
+            expressions: key (type: string)
+            outputColumnNames: _col0
+            ListSink
+
+PREHOOK: query: explain select value from c
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select value from c
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        TableScan
+          alias: c
+          Select Operator
+            expressions: value (type: string)
+            outputColumnNames: _col0
+            ListSink
+
+PREHOOK: query: explain select key from c
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select key from c
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        TableScan
+          alias: c
+          Select Operator
+            expressions: key (type: string)
+            outputColumnNames: _col0
+            ListSink
+

Reply via email to