http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q b/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q index 692c414..ed75c57 100644 --- a/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q +++ b/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q @@ -2,7 +2,7 @@ set hive.stats.column.autogather=false; set hive.stats.dbclass=fs; set hive.compute.query.using.stats=true; set hive.explain.user=false; -create table over10k( +create table over10k_n23( t tinyint, si smallint, i int, @@ -17,9 +17,9 @@ create table over10k( row format delimited fields terminated by '|'; -load data local inpath '../../data/files/over10k' into table over10k; +load data local inpath '../../data/files/over10k' into table over10k_n23; -create table stats_tbl_part( +create table stats_tbl_part_n0( t tinyint, si smallint, i int, @@ -33,22 +33,22 @@ create table stats_tbl_part( bin binary) partitioned by (dt int); -from over10k -insert overwrite table stats_tbl_part partition (dt=2010) select t,si,i,b,f,d,bo,s,ts,`dec`,bin where t>0 and t<30 -insert overwrite table stats_tbl_part partition (dt=2014) select t,si,i,b,f,d,bo,s,ts,`dec`,bin where t > 30 and t<60; +from over10k_n23 +insert overwrite table stats_tbl_part_n0 partition (dt=2010) select t,si,i,b,f,d,bo,s,ts,`dec`,bin where t>0 and t<30 +insert overwrite table stats_tbl_part_n0 partition (dt=2014) select t,si,i,b,f,d,bo,s,ts,`dec`,bin where t > 30 and t<60; -analyze table stats_tbl_part partition(dt) compute statistics; -analyze table stats_tbl_part partition(dt=2010) compute statistics for columns t,si,i,b,f,d,bo,s,bin; -analyze table stats_tbl_part partition(dt=2014) compute statistics for columns t,si,i,b,f,d,bo,s,bin; +analyze table stats_tbl_part_n0 partition(dt) compute statistics; +analyze table stats_tbl_part_n0 partition(dt=2010) compute statistics for columns t,si,i,b,f,d,bo,s,bin; +analyze table stats_tbl_part_n0 partition(dt=2014) compute statistics for columns t,si,i,b,f,d,bo,s,bin; explain -select count(*), count(1), sum(1), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part where dt = 2010; -select count(*), count(1), sum(1), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part where dt = 2010; +select count(*), count(1), sum(1), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part_n0 where dt = 2010; +select count(*), count(1), sum(1), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part_n0 where dt = 2010; explain -select count(*), count(1), sum(1), sum(2), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part where dt > 2010; -select count(*), count(1), sum(1), sum(2), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part where dt > 2010; +select count(*), count(1), sum(1), sum(2), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part_n0 where dt > 2010; +select count(*), count(1), sum(1), sum(2), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part_n0 where dt > 2010; -select count(*) from stats_tbl_part; -select count(*)/2 from stats_tbl_part; -drop table stats_tbl_part; +select count(*) from stats_tbl_part_n0; +select count(*)/2 from stats_tbl_part_n0; +drop table stats_tbl_part_n0; set hive.compute.query.using.stats=false;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/metadataonly1.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/metadataonly1.q b/ql/src/test/queries/clientpositive/metadataonly1.q index e3bf819..e247c6a 100644 --- a/ql/src/test/queries/clientpositive/metadataonly1.q +++ b/ql/src/test/queries/clientpositive/metadataonly1.q @@ -1,48 +1,48 @@ --! qt:dataset:srcpart set hive.mapred.mode=nonstrict; set hive.optimize.metadataonly=true; -CREATE TABLE TEST1(A INT, B DOUBLE) partitioned by (ds string); -explain extended select max(ds) from TEST1; -select max(ds) from TEST1; +CREATE TABLE TEST1_n12(A INT, B DOUBLE) partitioned by (ds string); +explain extended select max(ds) from TEST1_n12; +select max(ds) from TEST1_n12; -alter table TEST1 add partition (ds='1'); -explain extended select max(ds) from TEST1; -select max(ds) from TEST1; +alter table TEST1_n12 add partition (ds='1'); +explain extended select max(ds) from TEST1_n12; +select max(ds) from TEST1_n12; -explain extended select count(distinct ds) from TEST1; -select count(distinct ds) from TEST1; +explain extended select count(distinct ds) from TEST1_n12; +select count(distinct ds) from TEST1_n12; -explain extended select count(ds) from TEST1; -select count(ds) from TEST1; +explain extended select count(ds) from TEST1_n12; +select count(ds) from TEST1_n12; -alter table TEST1 add partition (ds='2'); +alter table TEST1_n12 add partition (ds='2'); explain extended -select count(*) from TEST1 a2 join (select max(ds) m from TEST1) b on a2.ds=b.m; -select count(*) from TEST1 a2 join (select max(ds) m from TEST1) b on a2.ds=b.m; +select count(*) from TEST1_n12 a2 join (select max(ds) m from TEST1_n12) b on a2.ds=b.m; +select count(*) from TEST1_n12 a2 join (select max(ds) m from TEST1_n12) b on a2.ds=b.m; -CREATE TABLE TEST2(A INT, B DOUBLE) partitioned by (ds string, hr string); -alter table TEST2 add partition (ds='1', hr='1'); -alter table TEST2 add partition (ds='1', hr='2'); -alter table TEST2 add partition (ds='1', hr='3'); +CREATE TABLE TEST2_n8(A INT, B DOUBLE) partitioned by (ds string, hr string); +alter table TEST2_n8 add partition (ds='1', hr='1'); +alter table TEST2_n8 add partition (ds='1', hr='2'); +alter table TEST2_n8 add partition (ds='1', hr='3'); -explain extended select ds, count(distinct hr) from TEST2 group by ds; -select ds, count(distinct hr) from TEST2 group by ds; +explain extended select ds, count(distinct hr) from TEST2_n8 group by ds; +select ds, count(distinct hr) from TEST2_n8 group by ds; -explain extended select ds, count(hr) from TEST2 group by ds; -select ds, count(hr) from TEST2 group by ds; +explain extended select ds, count(hr) from TEST2_n8 group by ds; +select ds, count(hr) from TEST2_n8 group by ds; set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat; -explain extended select max(ds) from TEST1; -select max(ds) from TEST1; +explain extended select max(ds) from TEST1_n12; +select max(ds) from TEST1_n12; select distinct ds from srcpart; select min(ds),max(ds) from srcpart; -- HIVE-3594 URI encoding for temporary path -alter table TEST2 add partition (ds='01:10:10', hr='01'); -alter table TEST2 add partition (ds='01:10:20', hr='02'); +alter table TEST2_n8 add partition (ds='01:10:10', hr='01'); +alter table TEST2_n8 add partition (ds='01:10:20', hr='02'); -explain extended select ds, count(distinct hr) from TEST2 group by ds; -select ds, count(distinct hr) from TEST2 group by ds; +explain extended select ds, count(distinct hr) from TEST2_n8 group by ds; +select ds, count(distinct hr) from TEST2_n8 group by ds; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/misc_json.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/misc_json.q b/ql/src/test/queries/clientpositive/misc_json.q index 541e369..95a3e5b 100644 --- a/ql/src/test/queries/clientpositive/misc_json.q +++ b/ql/src/test/queries/clientpositive/misc_json.q @@ -1,10 +1,10 @@ set hive.ddl.output.format=json; -CREATE TABLE IF NOT EXISTS jsontable (key INT, value STRING) COMMENT 'json table' STORED AS TEXTFILE; +CREATE TABLE IF NOT EXISTS jsontable_n0 (key INT, value STRING) COMMENT 'json table' STORED AS TEXTFILE; -ALTER TABLE jsontable ADD COLUMNS (name STRING COMMENT 'a new column'); +ALTER TABLE jsontable_n0 ADD COLUMNS (name STRING COMMENT 'a new column'); -ALTER TABLE jsontable RENAME TO jsontable2; +ALTER TABLE jsontable_n0 RENAME TO jsontable2; SHOW TABLE EXTENDED LIKE jsontable2; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_all.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/mm_all.q b/ql/src/test/queries/clientpositive/mm_all.q index 7d0955b..61dd3e7 100644 --- a/ql/src/test/queries/clientpositive/mm_all.q +++ b/ql/src/test/queries/clientpositive/mm_all.q @@ -13,32 +13,32 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -- Force multiple writers when reading -drop table intermediate; -create table intermediate(key int) partitioned by (p int) stored as orc; -insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2; -insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2; -insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2; - - -drop table part_mm; -create table part_mm(key int) partitioned by (key_mm int) stored as orc tblproperties ("transactional"="true", "transactional_properties"="insert_only"); -explain insert into table part_mm partition(key_mm=455) select key from intermediate; -insert into table part_mm partition(key_mm=455) select key from intermediate; -insert into table part_mm partition(key_mm=456) select key from intermediate; -insert into table part_mm partition(key_mm=455) select key from intermediate; -select * from part_mm order by key, key_mm; - --- TODO: doesn't work truncate table part_mm partition(key_mm=455); -select * from part_mm order by key, key_mm; -truncate table part_mm; -select * from part_mm order by key, key_mm; -drop table part_mm; +drop table intermediate_n0; +create table intermediate_n0(key int) partitioned by (p int) stored as orc; +insert into table intermediate_n0 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2; +insert into table intermediate_n0 partition(p='456') select distinct key from src where key is not null order by key asc limit 2; +insert into table intermediate_n0 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2; + + +drop table part_mm_n0; +create table part_mm_n0(key int) partitioned by (key_mm int) stored as orc tblproperties ("transactional"="true", "transactional_properties"="insert_only"); +explain insert into table part_mm_n0 partition(key_mm=455) select key from intermediate_n0; +insert into table part_mm_n0 partition(key_mm=455) select key from intermediate_n0; +insert into table part_mm_n0 partition(key_mm=456) select key from intermediate_n0; +insert into table part_mm_n0 partition(key_mm=455) select key from intermediate_n0; +select * from part_mm_n0 order by key, key_mm; + +-- TODO: doesn't work truncate table part_mm_n0 partition(key_mm=455); +select * from part_mm_n0 order by key, key_mm; +truncate table part_mm_n0; +select * from part_mm_n0 order by key, key_mm; +drop table part_mm_n0; drop table simple_mm; create table simple_mm(key int) stored as orc tblproperties ("transactional"="true", "transactional_properties"="insert_only"); -insert into table simple_mm select key from intermediate; +insert into table simple_mm select key from intermediate_n0; select * from simple_mm order by key; -insert into table simple_mm select key from intermediate; +insert into table simple_mm select key from intermediate_n0; select * from simple_mm order by key; truncate table simple_mm; select * from simple_mm; @@ -57,7 +57,7 @@ set hive.merge.tezfiles=false; create table dp_mm (key int) partitioned by (key1 string, key2 int) stored as orc tblproperties ("transactional"="true", "transactional_properties"="insert_only"); -insert into table dp_mm partition (key1='123', key2) select key, key from intermediate; +insert into table dp_mm partition (key1='123', key2) select key, key from intermediate_n0; select * from dp_mm order by key; @@ -69,35 +69,35 @@ drop table dp_mm; create table union_mm(id int) tblproperties ("transactional"="true", "transactional_properties"="insert_only"); insert into table union_mm select temps.p from ( -select key as p from intermediate +select key as p from intermediate_n0 union all -select key + 1 as p from intermediate ) temps; +select key + 1 as p from intermediate_n0 ) temps; select * from union_mm order by id; insert into table union_mm select p from ( -select key + 1 as p from intermediate +select key + 1 as p from intermediate_n0 union all -select key from intermediate +select key from intermediate_n0 ) tab group by p union all -select key + 2 as p from intermediate; +select key + 2 as p from intermediate_n0; select * from union_mm order by id; insert into table union_mm SELECT p FROM ( - SELECT key + 1 as p FROM intermediate + SELECT key + 1 as p FROM intermediate_n0 UNION ALL SELECT key as p FROM ( SELECT distinct key FROM ( SELECT key FROM ( - SELECT key + 2 as key FROM intermediate + SELECT key + 2 as key FROM intermediate_n0 UNION ALL - SELECT key FROM intermediate + SELECT key FROM intermediate_n0 )t1 group by key)t2 )t3 @@ -112,9 +112,9 @@ drop table union_mm; create table partunion_mm(id int) partitioned by (key int) tblproperties ("transactional"="true", "transactional_properties"="insert_only"); insert into table partunion_mm partition(key) select temps.* from ( -select key as p, key from intermediate +select key as p, key from intermediate_n0 union all -select key + 1 as p, key + 1 from intermediate ) temps; +select key + 1 as p, key + 1 from intermediate_n0 ) temps; select * from partunion_mm order by id; drop table partunion_mm; @@ -127,7 +127,7 @@ create table skew_mm(k1 int, k2 int, k4 int) skewed by (k1, k4) on ((0,0),(1,1), stored as directories tblproperties ("transactional"="true", "transactional_properties"="insert_only"); insert into table skew_mm -select key, key, key from intermediate; +select key, key, key from intermediate_n0; select * from skew_mm order by k2, k1, k4; drop table skew_mm; @@ -137,9 +137,9 @@ create table skew_dp_union_mm(k1 int, k2 int, k4 int) partitioned by (k3 int) skewed by (k1, k4) on ((0,0),(1,1),(2,2),(3,3)) stored as directories tblproperties ("transactional"="true", "transactional_properties"="insert_only"); insert into table skew_dp_union_mm partition (k3) -select key as i, key as j, key as k, key as l from intermediate +select key as i, key as j, key as k, key as l from intermediate_n0 union all -select key +1 as i, key +2 as j, key +3 as k, key +4 as l from intermediate; +select key +1 as i, key +2 as j, key +3 as k, key +4 as l from intermediate_n0; select * from skew_dp_union_mm order by k2, k1, k4; @@ -155,11 +155,11 @@ set hive.merge.mapredfiles=true; create table merge0_mm (id int) stored as orc tblproperties("transactional"="true", "transactional_properties"="insert_only"); -insert into table merge0_mm select key from intermediate; +insert into table merge0_mm select key from intermediate_n0; select * from merge0_mm; set tez.grouping.split-count=1; -insert into table merge0_mm select key from intermediate; +insert into table merge0_mm select key from intermediate_n0; set tez.grouping.split-count=0; select * from merge0_mm; @@ -168,11 +168,11 @@ drop table merge0_mm; create table merge2_mm (id int) tblproperties("transactional"="true", "transactional_properties"="insert_only"); -insert into table merge2_mm select key from intermediate; +insert into table merge2_mm select key from intermediate_n0; select * from merge2_mm; set tez.grouping.split-count=1; -insert into table merge2_mm select key from intermediate; +insert into table merge2_mm select key from intermediate_n0; set tez.grouping.split-count=0; select * from merge2_mm; @@ -181,11 +181,11 @@ drop table merge2_mm; create table merge1_mm (id int) partitioned by (key int) stored as orc tblproperties("transactional"="true", "transactional_properties"="insert_only"); -insert into table merge1_mm partition (key) select key, key from intermediate; +insert into table merge1_mm partition (key) select key, key from intermediate_n0; select * from merge1_mm order by id, key; set tez.grouping.split-count=1; -insert into table merge1_mm partition (key) select key, key from intermediate; +insert into table merge1_mm partition (key) select key, key from intermediate_n0; set tez.grouping.split-count=0; select * from merge1_mm order by id, key; @@ -199,13 +199,13 @@ set hive.merge.mapredfiles=false; drop table ctas0_mm; -create table ctas0_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as select * from intermediate; +create table ctas0_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as select * from intermediate_n0; select * from ctas0_mm; drop table ctas0_mm; drop table ctas1_mm; create table ctas1_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as - select * from intermediate union all select * from intermediate; + select * from intermediate_n0 union all select * from intermediate_n0; select * from ctas1_mm; drop table ctas1_mm; @@ -215,7 +215,7 @@ drop table multi0_2_mm; create table multi0_1_mm (key int, key2 int) tblproperties("transactional"="true", "transactional_properties"="insert_only"); create table multi0_2_mm (key int, key2 int) tblproperties("transactional"="true", "transactional_properties"="insert_only"); -from intermediate +from intermediate_n0 insert overwrite table multi0_1_mm select key, p insert overwrite table multi0_2_mm select p, key; @@ -226,7 +226,7 @@ set hive.merge.mapredfiles=true; set hive.merge.sparkfiles=true; set hive.merge.tezfiles=true; -from intermediate +from intermediate_n0 insert into table multi0_1_mm select p, key insert overwrite table multi0_2_mm select key, p; @@ -243,25 +243,25 @@ drop table multi0_2_mm; drop table multi1_mm; create table multi1_mm (key int, key2 int) partitioned by (p int) tblproperties("transactional"="true", "transactional_properties"="insert_only"); -from intermediate +from intermediate_n0 insert into table multi1_mm partition(p=1) select p, key insert into table multi1_mm partition(p=2) select key, p; select * from multi1_mm order by key, key2, p; -from intermediate +from intermediate_n0 insert into table multi1_mm partition(p=2) select p, key insert overwrite table multi1_mm partition(p=1) select key, p; select * from multi1_mm order by key, key2, p; -from intermediate +from intermediate_n0 insert into table multi1_mm partition(p) select p, key, p insert into table multi1_mm partition(p=1) select key, p; select key, key2, p from multi1_mm order by key, key2, p; -from intermediate +from intermediate_n0 insert into table multi1_mm partition(p) select p, key, 1 insert into table multi1_mm partition(p=1) select key, p; @@ -277,11 +277,11 @@ set hive.stats.autogather=true; drop table stats_mm; create table stats_mm(key int) tblproperties("transactional"="true", "transactional_properties"="insert_only"); ---insert overwrite table stats_mm select key from intermediate; -insert into table stats_mm select key from intermediate; +--insert overwrite table stats_mm select key from intermediate_n0; +insert into table stats_mm select key from intermediate_n0; desc formatted stats_mm; -insert into table stats_mm select key from intermediate; +insert into table stats_mm select key from intermediate_n0; desc formatted stats_mm; drop table stats_mm; @@ -321,12 +321,12 @@ set hive.auto.convert.join=true; DROP TABLE IF EXISTS temp1; CREATE TEMPORARY TABLE temp1 (a int) TBLPROPERTIES ("transactional"="true", "transactional_properties"="insert_only"); -INSERT INTO temp1 SELECT key FROM intermediate; +INSERT INTO temp1 SELECT key FROM intermediate_n0; DESC EXTENDED temp1; SELECT * FROM temp1; -drop table intermediate; +drop table intermediate_n0; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_buckets.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/mm_buckets.q b/ql/src/test/queries/clientpositive/mm_buckets.q index 3e91cdd..12dc4fe 100644 --- a/ql/src/test/queries/clientpositive/mm_buckets.q +++ b/ql/src/test/queries/clientpositive/mm_buckets.q @@ -12,11 +12,11 @@ set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -- Bucketing tests are slow and some tablesample ones don't work w/o MM -- Force multiple writers when reading -drop table intermediate; -create table intermediate(key int) partitioned by (p int) stored as orc; -insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2; -insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2; -insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2; +drop table intermediate_n2; +create table intermediate_n2(key int) partitioned by (p int) stored as orc; +insert into table intermediate_n2 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2; +insert into table intermediate_n2 partition(p='456') select distinct key from src where key is not null order by key asc limit 2; +insert into table intermediate_n2 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2; @@ -24,11 +24,11 @@ drop table bucket0_mm; create table bucket0_mm(key int, id int) clustered by (key) into 2 buckets tblproperties("transactional"="true", "transactional_properties"="insert_only"); -insert into table bucket0_mm select key, key from intermediate; +insert into table bucket0_mm select key, key from intermediate_n2; select * from bucket0_mm order by key, id; select * from bucket0_mm tablesample (bucket 1 out of 2) s; select * from bucket0_mm tablesample (bucket 2 out of 2) s; -insert into table bucket0_mm select key, key from intermediate; +insert into table bucket0_mm select key, key from intermediate_n2; select * from bucket0_mm order by key, id; select * from bucket0_mm tablesample (bucket 1 out of 2) s; select * from bucket0_mm tablesample (bucket 2 out of 2) s; @@ -40,9 +40,9 @@ create table bucket1_mm(key int, id int) partitioned by (key2 int) clustered by (key) sorted by (key) into 2 buckets tblproperties("transactional"="true", "transactional_properties"="insert_only"); insert into table bucket1_mm partition (key2) -select key + 1, key, key - 1 from intermediate +select key + 1, key, key - 1 from intermediate_n2 union all -select key - 1, key, key + 1 from intermediate; +select key - 1, key, key + 1 from intermediate_n2; select * from bucket1_mm order by key, id; select * from bucket1_mm tablesample (bucket 1 out of 2) s order by key, id; select * from bucket1_mm tablesample (bucket 2 out of 2) s order by key, id; @@ -54,14 +54,14 @@ drop table bucket2_mm; create table bucket2_mm(key int, id int) clustered by (key) into 10 buckets tblproperties("transactional"="true", "transactional_properties"="insert_only"); -insert into table bucket2_mm select key, key from intermediate where key == 0; +insert into table bucket2_mm select key, key from intermediate_n2 where key == 0; select * from bucket2_mm order by key, id; select * from bucket2_mm tablesample (bucket 1 out of 10) s order by key, id; select * from bucket2_mm tablesample (bucket 4 out of 10) s order by key, id; -insert into table bucket2_mm select key, key from intermediate where key in (0, 103); +insert into table bucket2_mm select key, key from intermediate_n2 where key in (0, 103); select * from bucket2_mm; select * from bucket2_mm tablesample (bucket 1 out of 10) s order by key, id; select * from bucket2_mm tablesample (bucket 4 out of 10) s order by key, id; drop table bucket2_mm; -drop table intermediate; \ No newline at end of file +drop table intermediate_n2; \ No newline at end of file http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_cttas.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/mm_cttas.q b/ql/src/test/queries/clientpositive/mm_cttas.q index da8b84d..8f1274c 100644 --- a/ql/src/test/queries/clientpositive/mm_cttas.q +++ b/ql/src/test/queries/clientpositive/mm_cttas.q @@ -3,19 +3,19 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -drop table intermediate; -create table intermediate(key int) partitioned by (p int) stored as orc; -insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2; -insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2; -insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2; +drop table intermediate_n1; +create table intermediate_n1(key int) partitioned by (p int) stored as orc; +insert into table intermediate_n1 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2; +insert into table intermediate_n1 partition(p='456') select distinct key from src where key is not null order by key asc limit 2; +insert into table intermediate_n1 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2; drop table cttas1_mm; -create temporary table cttas1_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as select * from intermediate; +create temporary table cttas1_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as select * from intermediate_n1; select * from cttas1_mm; drop table cttas1_mm; -drop table intermediate; +drop table intermediate_n1; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_default.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/mm_default.q b/ql/src/test/queries/clientpositive/mm_default.q index 0fa46a9..8e34bf3 100644 --- a/ql/src/test/queries/clientpositive/mm_default.q +++ b/ql/src/test/queries/clientpositive/mm_default.q @@ -14,10 +14,10 @@ create table acid0 (key string) stored as ORC tblproperties("transactional"="tr set hive.create.as.insert.only=true; create table mm1 like non_mm0; create table mm2 like mm0; -create table acid1 like acid0; +create table acid1_n0 like acid0; create table mm3 as select key from src limit 1; create table mm4 (key string); -create table acid2 (key string) stored as ORC tblproperties("transactional"="true"); +create table acid2_n0 (key string) stored as ORC tblproperties("transactional"="true"); create table non_mm1 tblproperties("transactional"="false") as select key from src limit 1; @@ -28,8 +28,8 @@ desc formatted mm2; desc formatted mm3; desc formatted mm4; desc formatted non_mm1; -desc formatted acid1; -desc formatted acid2; +desc formatted acid1_n0; +desc formatted acid2_n0; drop table non_mm0; @@ -40,8 +40,8 @@ drop table mm2; drop table mm3; drop table mm4; drop table acid0; -drop table acid1; -drop table acid2; +drop table acid1_n0; +drop table acid2_n0; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_exim.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/mm_exim.q b/ql/src/test/queries/clientpositive/mm_exim.q index a2b6e08..9870bf4 100644 --- a/ql/src/test/queries/clientpositive/mm_exim.q +++ b/ql/src/test/queries/clientpositive/mm_exim.q @@ -10,27 +10,27 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -drop table intermediate; -create table intermediate(key int) partitioned by (p int) stored as orc tblproperties("transactional"="false"); -insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2; -insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2; -insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2; +drop table intermediate_n4; +create table intermediate_n4(key int) partitioned by (p int) stored as orc tblproperties("transactional"="false"); +insert into table intermediate_n4 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2; +insert into table intermediate_n4 partition(p='456') select distinct key from src where key is not null order by key asc limit 2; +insert into table intermediate_n4 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2; drop table intermediate_nonpart; drop table intermmediate_part; drop table intermmediate_nonpart; create table intermediate_nonpart(key int, p int) tblproperties("transactional"="false"); -insert into intermediate_nonpart select * from intermediate; +insert into intermediate_nonpart select * from intermediate_n4; create table intermmediate_nonpart(key int, p int) tblproperties("transactional"="true", "transactional_properties"="insert_only"); -insert into intermmediate_nonpart select * from intermediate; +insert into intermmediate_nonpart select * from intermediate_n4; create table intermmediate(key int) partitioned by (p int) tblproperties("transactional"="true", "transactional_properties"="insert_only"); -insert into table intermmediate partition(p) select key, p from intermediate; +insert into table intermmediate partition(p) select key, p from intermediate_n4; set hive.exim.test.mode=true; export table intermediate_nonpart to 'ql/test/data/exports/intermediate_nonpart'; export table intermmediate_nonpart to 'ql/test/data/exports/intermmediate_nonpart'; -export table intermediate to 'ql/test/data/exports/intermediate_part'; +export table intermediate_n4 to 'ql/test/data/exports/intermediate_part'; export table intermmediate to 'ql/test/data/exports/intermmediate_part'; drop table intermediate_nonpart; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_0.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/msck_repair_0.q b/ql/src/test/queries/clientpositive/msck_repair_0.q index cb291fe..aeb4820 100644 --- a/ql/src/test/queries/clientpositive/msck_repair_0.q +++ b/ql/src/test/queries/clientpositive/msck_repair_0.q @@ -1,36 +1,36 @@ set hive.msck.repair.batch.size=1; set hive.mv.files.thread=0; -DROP TABLE IF EXISTS repairtable; +DROP TABLE IF EXISTS repairtable_n5; -CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING); +CREATE TABLE repairtable_n5(col STRING) PARTITIONED BY (p1 STRING, p2 STRING); -MSCK TABLE repairtable; +MSCK TABLE repairtable_n5; -show partitions repairtable; +show partitions repairtable_n5; dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b; dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b/datafile; -MSCK TABLE default.repairtable; +MSCK TABLE default.repairtable_n5; -show partitions default.repairtable; +show partitions default.repairtable_n5; -MSCK REPAIR TABLE default.repairtable; +MSCK REPAIR TABLE default.repairtable_n5; -show partitions default.repairtable; +show partitions default.repairtable_n5; -MSCK TABLE repairtable; +MSCK TABLE repairtable_n5; -show partitions repairtable; +show partitions repairtable_n5; set hive.mapred.mode=strict; dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=e/p2=f/p3=g; dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=e/p2=f/p3=g/datafile; -MSCK REPAIR TABLE default.repairtable; +MSCK REPAIR TABLE default.repairtable_n5; -show partitions default.repairtable; +show partitions default.repairtable_n5; -DROP TABLE default.repairtable; +DROP TABLE default.repairtable_n5; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_2.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/msck_repair_2.q b/ql/src/test/queries/clientpositive/msck_repair_2.q index 77785e9..be745b2 100644 --- a/ql/src/test/queries/clientpositive/msck_repair_2.q +++ b/ql/src/test/queries/clientpositive/msck_repair_2.q @@ -1,25 +1,25 @@ set hive.msck.repair.batch.size=1; set hive.msck.path.validation=skip; -DROP TABLE IF EXISTS repairtable; +DROP TABLE IF EXISTS repairtable_n2; -CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING); +CREATE TABLE repairtable_n2(col STRING) PARTITIONED BY (p1 STRING, p2 STRING); -MSCK TABLE repairtable; +MSCK TABLE repairtable_n2; -show partitions repairtable; +show partitions repairtable_n2; dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b; dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b/datafile; dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=c/datafile; -MSCK TABLE default.repairtable; -show partitions repairtable; +MSCK TABLE default.repairtable_n2; +show partitions repairtable_n2; -MSCK REPAIR TABLE default.repairtable; -show partitions repairtable; +MSCK REPAIR TABLE default.repairtable_n2; +show partitions repairtable_n2; -MSCK TABLE repairtable; -show partitions repairtable; +MSCK TABLE repairtable_n2; +show partitions repairtable_n2; -DROP TABLE default.repairtable; +DROP TABLE default.repairtable_n2; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_3.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/msck_repair_3.q b/ql/src/test/queries/clientpositive/msck_repair_3.q index f42443f..140a690 100644 --- a/ql/src/test/queries/clientpositive/msck_repair_3.q +++ b/ql/src/test/queries/clientpositive/msck_repair_3.q @@ -1,21 +1,21 @@ set hive.msck.repair.batch.size=1; -DROP TABLE IF EXISTS repairtable; +DROP TABLE IF EXISTS repairtable_n3; -CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING); +CREATE TABLE repairtable_n3(col STRING) PARTITIONED BY (p1 STRING, p2 STRING); -MSCK TABLE repairtable; -show partitions repairtable; +MSCK TABLE repairtable_n3; +show partitions repairtable_n3; dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b; -MSCK TABLE default.repairtable; -show partitions repairtable; +MSCK TABLE default.repairtable_n3; +show partitions repairtable_n3; -MSCK REPAIR TABLE default.repairtable; -show partitions repairtable; +MSCK REPAIR TABLE default.repairtable_n3; +show partitions repairtable_n3; -MSCK TABLE repairtable; -show partitions repairtable; +MSCK TABLE repairtable_n3; +show partitions repairtable_n3; -DROP TABLE default.repairtable; +DROP TABLE default.repairtable_n3; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_batchsize.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/msck_repair_batchsize.q b/ql/src/test/queries/clientpositive/msck_repair_batchsize.q index a44c00e..5a7afcc 100644 --- a/ql/src/test/queries/clientpositive/msck_repair_batchsize.q +++ b/ql/src/test/queries/clientpositive/msck_repair_batchsize.q @@ -1,10 +1,10 @@ set hive.msck.repair.batch.size=2; -DROP TABLE IF EXISTS repairtable; +DROP TABLE IF EXISTS repairtable_n0; -CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING); +CREATE TABLE repairtable_n0(col STRING) PARTITIONED BY (p1 STRING, p2 STRING); -MSCK TABLE repairtable; +MSCK TABLE repairtable_n0; dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=a/p2=a; dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=b/p2=a; @@ -13,24 +13,24 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=a/p2=a/datafile; dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=b/p2=a/datafile; dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/datafile; -MSCK TABLE default.repairtable; -show partitions default.repairtable; +MSCK TABLE default.repairtable_n0; +show partitions default.repairtable_n0; -MSCK REPAIR TABLE default.repairtable; -show partitions default.repairtable; +MSCK REPAIR TABLE default.repairtable_n0; +show partitions default.repairtable_n0; -MSCK TABLE repairtable; -show partitions repairtable; +MSCK TABLE repairtable_n0; +show partitions repairtable_n0; -DROP TABLE default.repairtable; +DROP TABLE default.repairtable_n0; dfs ${system:test.dfs.mkdir} -p ${system:test.tmp.dir}/apps/hive/warehouse/test.db/repairtable/p1=c/p2=a/p3=b; -CREATE TABLE `repairtable`( `col` string) PARTITIONED BY ( `p1` string, `p2` string) location '${system:test.tmp.dir}/apps/hive/warehouse/test.db/repairtable/'; +CREATE TABLE `repairtable_n0`( `col` string) PARTITIONED BY ( `p1` string, `p2` string) location '${system:test.tmp.dir}/apps/hive/warehouse/test.db/repairtable/'; dfs -touchz ${system:test.tmp.dir}/apps/hive/warehouse/test.db/repairtable/p1=c/p2=a/p3=b/datafile; set hive.mv.files.thread=1; -MSCK TABLE repairtable; -show partitions repairtable; +MSCK TABLE repairtable_n0; +show partitions repairtable_n0; -DROP TABLE default.repairtable; +DROP TABLE default.repairtable_n0; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_drop.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/msck_repair_drop.q b/ql/src/test/queries/clientpositive/msck_repair_drop.q index bc14d98..9923fb5 100644 --- a/ql/src/test/queries/clientpositive/msck_repair_drop.q +++ b/ql/src/test/queries/clientpositive/msck_repair_drop.q @@ -1,9 +1,9 @@ set hive.mv.files.thread=0; -DROP TABLE IF EXISTS repairtable; +DROP TABLE IF EXISTS repairtable_n1; -CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING); +CREATE TABLE repairtable_n1(col STRING) PARTITIONED BY (p1 STRING, p2 STRING); --- repairtable will have partitions created with part keys p1=1, p1=2, p1=3, p1=4 and p1=5 +-- repairtable_n1 will have partitions created with part keys p1=1, p1=2, p1=3, p1=4 and p1=5 -- p1=2 will be used to test drop in 3 tests -- 1. each partition is dropped individually: set hive.msck.repair.batch.size=1; -- 2. partition are dropped in groups of 3: set hive.msck.repair.batch.size=3; @@ -37,20 +37,20 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=29/p3=291/datafile; dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101; dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101/datafile; -MSCK TABLE default.repairtable; -MSCK REPAIR TABLE default.repairtable; +MSCK TABLE default.repairtable_n1; +MSCK REPAIR TABLE default.repairtable_n1; -- Now all 12 partitions are in -show partitions default.repairtable; +show partitions default.repairtable_n1; -- Remove all p1=2 partitions from file system dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=2; -- test 1: each partition is dropped individually set hive.msck.repair.batch.size=1; -MSCK TABLE default.repairtable DROP PARTITIONS; -MSCK REPAIR TABLE default.repairtable DROP PARTITIONS; -show partitions default.repairtable; +MSCK TABLE default.repairtable_n1 DROP PARTITIONS; +MSCK REPAIR TABLE default.repairtable_n1 DROP PARTITIONS; +show partitions default.repairtable_n1; -- Recreate p1=2 partitions dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=21/p3=211; @@ -74,20 +74,20 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=29/p3=291/datafile; dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101; dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101/datafile; -MSCK TABLE default.repairtable; -MSCK REPAIR TABLE default.repairtable; +MSCK TABLE default.repairtable_n1; +MSCK REPAIR TABLE default.repairtable_n1; -- Now all 12 partitions are in -show partitions default.repairtable; +show partitions default.repairtable_n1; -- Remove all p1=2 partitions from file system dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=2; -- test 2: partition are dropped in groups of 3 set hive.msck.repair.batch.size=3; -MSCK TABLE default.repairtable DROP PARTITIONS; -MSCK REPAIR TABLE default.repairtable DROP PARTITIONS; -show partitions default.repairtable; +MSCK TABLE default.repairtable_n1 DROP PARTITIONS; +MSCK REPAIR TABLE default.repairtable_n1 DROP PARTITIONS; +show partitions default.repairtable_n1; -- Recreate p1=2 partitions dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=21/p3=211; @@ -111,20 +111,20 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=29/p3=291/datafile; dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101; dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101/datafile; -MSCK TABLE default.repairtable; -MSCK REPAIR TABLE default.repairtable; +MSCK TABLE default.repairtable_n1; +MSCK REPAIR TABLE default.repairtable_n1; -- Now all 12 partitions are in -show partitions default.repairtable; +show partitions default.repairtable_n1; -- Remove all p1=2 partitions from file system dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=2; -- test 3. all partitions are dropped in 1 shot set hive.msck.repair.batch.size=0; -MSCK TABLE default.repairtable DROP PARTITIONS; -MSCK REPAIR TABLE default.repairtable DROP PARTITIONS; -show partitions default.repairtable; +MSCK TABLE default.repairtable_n1 DROP PARTITIONS; +MSCK REPAIR TABLE default.repairtable_n1 DROP PARTITIONS; +show partitions default.repairtable_n1; -- test add parition keyword: begin dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=3/p2=31/p3=311; @@ -132,9 +132,9 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=3/p2=31/p3=311/datafile; dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=3/p2=32/p3=321; dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=3/p2=32/p3=321/datafile; -MSCK TABLE default.repairtable; -MSCK REPAIR TABLE default.repairtable; -show partitions default.repairtable; +MSCK TABLE default.repairtable_n1; +MSCK REPAIR TABLE default.repairtable_n1; +show partitions default.repairtable_n1; -- Create p1=4 in filesystem dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=4/p2=41/p3=411; @@ -148,9 +148,9 @@ dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=3; -- Status: p1=3 dropped from filesystem, but exists in metastore -- p1=4 exists in filesystem but not in metastore -- test add partition: only brings in p1=4 and doesn't remove p1=3 -MSCK TABLE default.repairtable ADD PARTITIONS; -MSCK REPAIR TABLE default.repairtable ADD PARTITIONS; -show partitions default.repairtable; +MSCK TABLE default.repairtable_n1 ADD PARTITIONS; +MSCK REPAIR TABLE default.repairtable_n1 ADD PARTITIONS; +show partitions default.repairtable_n1; -- test add partition keyword: end -- test drop partition keyword: begin @@ -162,9 +162,9 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=5/p2=52/p3=521/datafile; -- Status: p1=3 removed from filesystem, but exists in metastore (as part of add test) -- p1=5 exists in filesystem but not in metastore -- test drop partition: only removes p1=3 from metastore but doesn't update metadata for p1=5 -MSCK TABLE default.repairtable DROP PARTITIONS; -MSCK REPAIR TABLE default.repairtable DROP PARTITIONS; -show partitions default.repairtable; +MSCK TABLE default.repairtable_n1 DROP PARTITIONS; +MSCK REPAIR TABLE default.repairtable_n1 DROP PARTITIONS; +show partitions default.repairtable_n1; -- test drop partition keyword: end -- test sync partition keyword: begin @@ -174,7 +174,7 @@ dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=4; -- Status: p1=4 dropped from filesystem, but exists in metastore -- p1=5 exists in filesystem but not in metastore (as part of drop test) -- test sync partition: removes p1=4 from metastore and updates metadata for p1=5 -MSCK TABLE default.repairtable SYNC PARTITIONS; -MSCK REPAIR TABLE default.repairtable SYNC PARTITIONS; -show partitions default.repairtable; +MSCK TABLE default.repairtable_n1 SYNC PARTITIONS; +MSCK REPAIR TABLE default.repairtable_n1 SYNC PARTITIONS; +show partitions default.repairtable_n1; -- test sync partition keyword: end http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multiMapJoin2.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/multiMapJoin2.q b/ql/src/test/queries/clientpositive/multiMapJoin2.q index efaadcb..166cb09 100644 --- a/ql/src/test/queries/clientpositive/multiMapJoin2.q +++ b/ql/src/test/queries/clientpositive/multiMapJoin2.q @@ -179,18 +179,18 @@ GROUP BY tmp1.key ORDER BY key, cnt; -- Check if we can correctly handle partitioned table. -CREATE TABLE part_table(key string, value string) PARTITIONED BY (partitionId int); -INSERT OVERWRITE TABLE part_table PARTITION (partitionId=1) +CREATE TABLE part_table_n0(key string, value string) PARTITIONED BY (partitionId int); +INSERT OVERWRITE TABLE part_table_n0 PARTITION (partitionId=1) SELECT key, value FROM src ORDER BY key, value LIMIT 100; -INSERT OVERWRITE TABLE part_table PARTITION (partitionId=2) +INSERT OVERWRITE TABLE part_table_n0 PARTITION (partitionId=2) SELECT key, value FROM src1 ORDER BY key, value; EXPLAIN SELECT count(*) -FROM part_table x JOIN src1 y ON (x.key = y.key); +FROM part_table_n0 x JOIN src1 y ON (x.key = y.key); SELECT count(*) -FROM part_table x JOIN src1 y ON (x.key = y.key); +FROM part_table_n0 x JOIN src1 y ON (x.key = y.key); set hive.auto.convert.join.noconditionaltask.size=10000000; set hive.optimize.correlation=false; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_column_in.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/multi_column_in.q b/ql/src/test/queries/clientpositive/multi_column_in.q index 288406c..44fb1d9 100644 --- a/ql/src/test/queries/clientpositive/multi_column_in.q +++ b/ql/src/test/queries/clientpositive/multi_column_in.q @@ -1,72 +1,72 @@ set hive.mapred.mode=nonstrict; -drop table emps; +drop table emps_n1; -create table emps (empno int, deptno int, empname string); +create table emps_n1 (empno int, deptno int, empname string); -insert into table emps values (1,2,"11"),(1,2,"11"),(3,4,"33"),(1,3,"11"),(2,5,"22"),(2,5,"22"); +insert into table emps_n1 values (1,2,"11"),(1,2,"11"),(3,4,"33"),(1,3,"11"),(2,5,"22"),(2,5,"22"); -select * from emps; +select * from emps_n1; -select * from emps where (int(empno+deptno/2), int(deptno/3)) in ((2,0),(3,2)); +select * from emps_n1 where (int(empno+deptno/2), int(deptno/3)) in ((2,0),(3,2)); -select * from emps where (int(empno+deptno/2), int(deptno/3)) not in ((2,0),(3,2)); +select * from emps_n1 where (int(empno+deptno/2), int(deptno/3)) not in ((2,0),(3,2)); -select * from emps where (empno,deptno) in ((1,2),(3,2)); +select * from emps_n1 where (empno,deptno) in ((1,2),(3,2)); -select * from emps where (empno,deptno) not in ((1,2),(3,2)); +select * from emps_n1 where (empno,deptno) not in ((1,2),(3,2)); -select * from emps where (empno,deptno) in ((1,2),(1,3)); +select * from emps_n1 where (empno,deptno) in ((1,2),(1,3)); -select * from emps where (empno,deptno) not in ((1,2),(1,3)); +select * from emps_n1 where (empno,deptno) not in ((1,2),(1,3)); explain -select * from emps where (empno+1,deptno) in ((1,2),(3,2)); +select * from emps_n1 where (empno+1,deptno) in ((1,2),(3,2)); explain -select * from emps where (empno+1,deptno) not in ((1,2),(3,2)); +select * from emps_n1 where (empno+1,deptno) not in ((1,2),(3,2)); -select * from emps where empno in (1,2); +select * from emps_n1 where empno in (1,2); -select * from emps where empno in (1,2) and deptno > 2; +select * from emps_n1 where empno in (1,2) and deptno > 2; -select * from emps where (empno) in (1,2) and deptno > 2; +select * from emps_n1 where (empno) in (1,2) and deptno > 2; -select * from emps where ((empno) in (1,2) and deptno > 2); +select * from emps_n1 where ((empno) in (1,2) and deptno > 2); -explain select * from emps where ((empno*2)|1,deptno) in ((empno+1,2),(empno+2,2)); +explain select * from emps_n1 where ((empno*2)|1,deptno) in ((empno+1,2),(empno+2,2)); -select * from emps where ((empno*2)|1,deptno) in ((empno+1,2),(empno+2,2)); +select * from emps_n1 where ((empno*2)|1,deptno) in ((empno+1,2),(empno+2,2)); -select (empno*2)|1,substr(empname,1,1) from emps; +select (empno*2)|1,substr(empname,1,1) from emps_n1; -select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+2,'2')); +select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+2,'2')); -select * from emps where ((empno*2)|1,substr(empname,1,1)) not in ((empno+1,'2'),(empno+2,'2')); +select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) not in ((empno+1,'2'),(empno+2,'2')); -select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')); +select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')); -select * from emps where ((empno*2)|1,substr(empname,1,1)) not in ((empno+1,'2'),(empno+3,'2')); +select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) not in ((empno+1,'2'),(empno+3,'2')); -select sum(empno), empname from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')) +select sum(empno), empname from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')) group by empname; -select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')) +select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')) union -select * from emps where (empno,deptno) in ((1,2),(3,2)); +select * from emps_n1 where (empno,deptno) in ((1,2),(3,2)); -drop view v; +drop view v_n2; -create view v as +create view v_n2 as select * from( -select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')) +select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')) union -select * from emps where (empno,deptno) in ((1,2),(3,2)))subq order by empno desc; +select * from emps_n1 where (empno,deptno) in ((1,2),(3,2)))subq order by empno desc; -select * from v; +select * from v_n2; select subq.e1 from -(select (empno*2)|1 as e1, substr(empname,1,1) as n1 from emps)subq +(select (empno*2)|1 as e1, substr(empname,1,1) as n1 from emps_n1)subq join -(select empno as e2 from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')))subq2 +(select empno as e2 from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')))subq2 on e1=e2+1; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_column_in_single.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/multi_column_in_single.q b/ql/src/test/queries/clientpositive/multi_column_in_single.q index 7151fc4..32a2167 100644 --- a/ql/src/test/queries/clientpositive/multi_column_in_single.q +++ b/ql/src/test/queries/clientpositive/multi_column_in_single.q @@ -3,66 +3,66 @@ set hive.mapred.mode=nonstrict; select * from src where (key, value) in (('238','val_238')); -drop table emps; +drop table emps_n7; -create table emps (empno int, deptno int, empname string); +create table emps_n7 (empno int, deptno int, empname string); -insert into table emps values (1,2,"11"),(1,2,"11"),(3,4,"33"),(1,3,"11"),(2,5,"22"),(2,5,"22"); +insert into table emps_n7 values (1,2,"11"),(1,2,"11"),(3,4,"33"),(1,3,"11"),(2,5,"22"),(2,5,"22"); -select * from emps; +select * from emps_n7; -select * from emps where (int(empno+deptno/2), int(deptno/3)) in ((3,2)); +select * from emps_n7 where (int(empno+deptno/2), int(deptno/3)) in ((3,2)); -select * from emps where (int(empno+deptno/2), int(deptno/3)) not in ((3,2)); +select * from emps_n7 where (int(empno+deptno/2), int(deptno/3)) not in ((3,2)); -select * from emps where (empno,deptno) in ((3,2)); +select * from emps_n7 where (empno,deptno) in ((3,2)); -select * from emps where (empno,deptno) not in ((3,2)); +select * from emps_n7 where (empno,deptno) not in ((3,2)); -select * from emps where (empno,deptno) in ((1,3)); +select * from emps_n7 where (empno,deptno) in ((1,3)); -select * from emps where (empno,deptno) not in ((1,3)); +select * from emps_n7 where (empno,deptno) not in ((1,3)); explain -select * from emps where (empno+1,deptno) in ((3,2)); +select * from emps_n7 where (empno+1,deptno) in ((3,2)); explain -select * from emps where (empno+1,deptno) not in ((3,2)); +select * from emps_n7 where (empno+1,deptno) not in ((3,2)); -explain select * from emps where ((empno*2)|1,deptno) in ((empno+2,2)); +explain select * from emps_n7 where ((empno*2)|1,deptno) in ((empno+2,2)); -select * from emps where ((empno*2)|1,deptno) in ((empno+2,2)); +select * from emps_n7 where ((empno*2)|1,deptno) in ((empno+2,2)); -select (empno*2)|1,substr(empname,1,1) from emps; +select (empno*2)|1,substr(empname,1,1) from emps_n7; -select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+2,'2')); +select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+2,'2')); -select * from emps where ((empno*2)|1,substr(empname,1,1)) not in ((empno+2,'2')); +select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) not in ((empno+2,'2')); -select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')); +select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')); -select * from emps where ((empno*2)|1,substr(empname,1,1)) not in ((empno+3,'2')); +select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) not in ((empno+3,'2')); -select sum(empno), empname from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')) +select sum(empno), empname from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')) group by empname; -select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')) +select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')) union -select * from emps where (empno,deptno) in ((3,2)); +select * from emps_n7 where (empno,deptno) in ((3,2)); -drop view v; +drop view v_n11; -create view v as +create view v_n11 as select * from( -select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')) +select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')) union -select * from emps where (empno,deptno) in ((3,2)))subq order by empno desc; +select * from emps_n7 where (empno,deptno) in ((3,2)))subq order by empno desc; -select * from v; +select * from v_n11; select subq.e1 from -(select (empno*2)|1 as e1, substr(empname,1,1) as n1 from emps)subq +(select (empno*2)|1 as e1, substr(empname,1,1) as n1 from emps_n7)subq join -(select empno as e2 from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')))subq2 +(select empno as e2 from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')))subq2 on e1=e2+1; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_count_distinct.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/multi_count_distinct.q b/ql/src/test/queries/clientpositive/multi_count_distinct.q index 038be3e..6a9cbc9 100644 --- a/ql/src/test/queries/clientpositive/multi_count_distinct.q +++ b/ql/src/test/queries/clientpositive/multi_count_distinct.q @@ -1,36 +1,36 @@ SET hive.vectorized.execution.enabled=false; set hive.mapred.mode=nonstrict; -drop table employee; +drop table employee_n1; -create table employee (department_id int, gender varchar(10), education_level int); +create table employee_n1 (department_id int, gender varchar(10), education_level int); -insert into employee values (1, 'M', 1),(1, 'M', 1),(2, 'F', 1),(1, 'F', 3),(1, 'M', 2),(4, 'M', 1),(2, 'F', 1),(2, 'F', 3),(3, 'M', 2); +insert into employee_n1 values (1, 'M', 1),(1, 'M', 1),(2, 'F', 1),(1, 'F', 3),(1, 'M', 2),(4, 'M', 1),(2, 'F', 1),(2, 'F', 3),(3, 'M', 2); -explain select count(distinct department_id), count(distinct gender), count(distinct education_level) from employee; +explain select count(distinct department_id), count(distinct gender), count(distinct education_level) from employee_n1; -select count(distinct department_id), count(distinct gender), count(distinct education_level) from employee; +select count(distinct department_id), count(distinct gender), count(distinct education_level) from employee_n1; -select count(distinct department_id), count(distinct gender), count(distinct education_level), count(distinct education_level) from employee; +select count(distinct department_id), count(distinct gender), count(distinct education_level), count(distinct education_level) from employee_n1; select count(distinct department_id), count(distinct gender), count(distinct education_level), -count(distinct education_level, department_id) from employee; +count(distinct education_level, department_id) from employee_n1; select count(distinct gender), count(distinct department_id), count(distinct gender), count(distinct education_level), -count(distinct education_level, department_id), count(distinct department_id, education_level) from employee; +count(distinct education_level, department_id), count(distinct department_id, education_level) from employee_n1; explain select count(distinct gender), count(distinct department_id), count(distinct gender), count(distinct education_level), -count(distinct education_level, department_id), count(distinct department_id, education_level), count(distinct department_id, education_level, gender) from employee; +count(distinct education_level, department_id), count(distinct department_id, education_level), count(distinct department_id, education_level, gender) from employee_n1; select count(distinct gender), count(distinct department_id), count(distinct gender), count(distinct education_level), -count(distinct education_level, department_id), count(distinct department_id, education_level), count(distinct department_id, education_level, gender) from employee; +count(distinct education_level, department_id), count(distinct department_id, education_level), count(distinct department_id, education_level, gender) from employee_n1; select count(case i when 3 then 1 else null end) as c0, count(case i when 5 then 1 else null end) as c1, count(case i when 6 then 1 else null end) as c2 from (select grouping__id as i, department_id, gender, -education_level from employee group by department_id, gender, education_level grouping sets +education_level from employee_n1 group by department_id, gender, education_level grouping sets (department_id, gender, education_level))subq; -select grouping__id as i, department_id, gender, education_level from employee +select grouping__id as i, department_id, gender, education_level from employee_n1 group by department_id, gender, education_level grouping sets (department_id, gender, education_level, (education_level, department_id)); http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/multi_insert.q b/ql/src/test/queries/clientpositive/multi_insert.q index 9e33a84..634bdb4 100644 --- a/ql/src/test/queries/clientpositive/multi_insert.q +++ b/ql/src/test/queries/clientpositive/multi_insert.q @@ -1,23 +1,23 @@ --! qt:dataset:src -- SORT_QUERY_RESULTS -create table src_multi1 like src; -create table src_multi2 like src; +create table src_multi1_n5 like src; +create table src_multi2_n6 like src; set hive.merge.mapfiles=false; set hive.merge.mapredfiles=false; set hive.stats.dbclass=fs; explain from src -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; from src -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n5; +select * from src_multi2_n6; set hive.merge.mapfiles=true; @@ -25,45 +25,45 @@ set hive.merge.mapredfiles=false; explain from src -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; from src -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n5; +select * from src_multi2_n6; set hive.merge.mapfiles=false; set hive.merge.mapredfiles=true; explain from src -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; from src -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n5; +select * from src_multi2_n6; set hive.merge.mapfiles=true; set hive.merge.mapredfiles=true; explain from src -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; from src -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n5; +select * from src_multi2_n6; @@ -72,15 +72,15 @@ set hive.merge.mapredfiles=false; explain from src -insert overwrite table src_multi1 select * where key < 10 group by key, value -insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value; +insert overwrite table src_multi1_n5 select * where key < 10 group by key, value +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value; from src -insert overwrite table src_multi1 select * where key < 10 group by key, value -insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value; +insert overwrite table src_multi1_n5 select * where key < 10 group by key, value +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n5; +select * from src_multi2_n6; set hive.merge.mapfiles=false; @@ -88,30 +88,30 @@ set hive.merge.mapredfiles=true; explain from src -insert overwrite table src_multi1 select * where key < 10 group by key, value -insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value; +insert overwrite table src_multi1_n5 select * where key < 10 group by key, value +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value; from src -insert overwrite table src_multi1 select * where key < 10 group by key, value -insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value; +insert overwrite table src_multi1_n5 select * where key < 10 group by key, value +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n5; +select * from src_multi2_n6; set hive.merge.mapfiles=true; set hive.merge.mapredfiles=false; explain from src -insert overwrite table src_multi1 select * where key < 10 group by key, value -insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value; +insert overwrite table src_multi1_n5 select * where key < 10 group by key, value +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value; from src -insert overwrite table src_multi1 select * where key < 10 group by key, value -insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value; +insert overwrite table src_multi1_n5 select * where key < 10 group by key, value +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n5; +select * from src_multi2_n6; set hive.merge.mapfiles=true; @@ -119,15 +119,15 @@ set hive.merge.mapredfiles=true; explain from src -insert overwrite table src_multi1 select * where key < 10 group by key, value -insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value; +insert overwrite table src_multi1_n5 select * where key < 10 group by key, value +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value; from src -insert overwrite table src_multi1 select * where key < 10 group by key, value -insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value; +insert overwrite table src_multi1_n5 select * where key < 10 group by key, value +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n5; +select * from src_multi2_n6; @@ -137,60 +137,60 @@ set hive.merge.mapredfiles=false; explain from (select * from src union all select * from src) s -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; from (select * from src union all select * from src) s -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n5; +select * from src_multi2_n6; set hive.merge.mapfiles=true; set hive.merge.mapredfiles=false; explain from (select * from src union all select * from src) s -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; from (select * from src union all select * from src) s -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n5; +select * from src_multi2_n6; set hive.merge.mapfiles=false; set hive.merge.mapredfiles=true; explain from (select * from src union all select * from src) s -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; from (select * from src union all select * from src) s -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n5; +select * from src_multi2_n6; set hive.merge.mapfiles=true; set hive.merge.mapredfiles=true; explain from (select * from src union all select * from src) s -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; from (select * from src union all select * from src) s -insert overwrite table src_multi1 select * where key < 10 -insert overwrite table src_multi2 select * where key > 10 and key < 20; +insert overwrite table src_multi1_n5 select * where key < 10 +insert overwrite table src_multi2_n6 select * where key > 10 and key < 20; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n5; +select * from src_multi2_n6; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_gby.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/multi_insert_gby.q b/ql/src/test/queries/clientpositive/multi_insert_gby.q index 7c9e8de..0ec2b35 100644 --- a/ql/src/test/queries/clientpositive/multi_insert_gby.q +++ b/ql/src/test/queries/clientpositive/multi_insert_gby.q @@ -2,35 +2,35 @@ -- SORT_QUERY_RESULTS --HIVE-3699 Multiple insert overwrite into multiple tables query stores same results in all tables -create table e1 (key string, count int); -create table e2 (key string, count int); +create table e1_n0 (key string, count int); +create table e2_n1 (key string, count int); explain FROM src -INSERT OVERWRITE TABLE e1 +INSERT OVERWRITE TABLE e1_n0 SELECT key, COUNT(*) WHERE key>450 GROUP BY key -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n1 SELECT key, COUNT(*) WHERE key>500 GROUP BY key; FROM src -INSERT OVERWRITE TABLE e1 +INSERT OVERWRITE TABLE e1_n0 SELECT key, COUNT(*) WHERE key>450 GROUP BY key -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n1 SELECT key, COUNT(*) WHERE key>500 GROUP BY key; -select * from e1; -select * from e2; +select * from e1_n0; +select * from e2_n1; explain FROM src -INSERT OVERWRITE TABLE e1 +INSERT OVERWRITE TABLE e1_n0 SELECT key, COUNT(*) WHERE key>450 GROUP BY key -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n1 SELECT key, COUNT(*) GROUP BY key; FROM src -INSERT OVERWRITE TABLE e1 +INSERT OVERWRITE TABLE e1_n0 SELECT key, COUNT(*) WHERE key>450 GROUP BY key -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n1 SELECT key, COUNT(*) GROUP BY key; -select * from e1; -select * from e2; +select * from e1_n0; +select * from e2_n1; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_gby2.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/multi_insert_gby2.q b/ql/src/test/queries/clientpositive/multi_insert_gby2.q index a477f83..67077ba 100644 --- a/ql/src/test/queries/clientpositive/multi_insert_gby2.q +++ b/ql/src/test/queries/clientpositive/multi_insert_gby2.q @@ -2,20 +2,20 @@ set hive.mapred.mode=nonstrict; --HIVE-3699 Multiple insert overwrite into multiple tables query stores same results in all tables create table e1 (count int); -create table e2 (percentile double); +create table e2_n0 (percentile double); set hive.stats.dbclass=fs; explain FROM (select key, cast(key as double) as value from src order by key) a INSERT OVERWRITE TABLE e1 SELECT COUNT(*) -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n0 SELECT percentile_approx(value, 0.5); FROM (select key, cast(key as double) as value from src order by key) a INSERT OVERWRITE TABLE e1 SELECT COUNT(*) -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n0 SELECT percentile_approx(value, 0.5); select * from e1; -select * from e2; +select * from e2_n0; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_gby3.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/multi_insert_gby3.q b/ql/src/test/queries/clientpositive/multi_insert_gby3.q index 6bed476..de60fa7 100644 --- a/ql/src/test/queries/clientpositive/multi_insert_gby3.q +++ b/ql/src/test/queries/clientpositive/multi_insert_gby3.q @@ -1,45 +1,45 @@ --! qt:dataset:src set hive.mapred.mode=nonstrict; -- SORT_QUERY_RESULTS -create table e1 (key string, keyD double); -create table e2 (key string, keyD double, value string); +create table e1_n2 (key string, keyD double); +create table e2_n3 (key string, keyD double, value string); create table e3 (key string, keyD double); set hive.stats.dbclass=fs; explain FROM (select key, cast(key as double) as keyD, value from src order by key) a -INSERT OVERWRITE TABLE e1 +INSERT OVERWRITE TABLE e1_n2 SELECT key, COUNT(distinct value) group by key -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n3 SELECT key, sum(keyD), value group by key, value; explain FROM (select key, cast(key as double) as keyD, value from src order by key) a -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n3 SELECT key, sum(keyD), value group by key, value -INSERT OVERWRITE TABLE e1 +INSERT OVERWRITE TABLE e1_n2 SELECT key, COUNT(distinct value) group by key; FROM (select key, cast(key as double) as keyD, value from src order by key) a -INSERT OVERWRITE TABLE e1 +INSERT OVERWRITE TABLE e1_n2 SELECT key, COUNT(distinct value) group by key -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n3 SELECT key, sum(keyD), value group by key, value; -select * from e1; -select * from e2; +select * from e1_n2; +select * from e2_n3; FROM (select key, cast(key as double) as keyD, value from src order by key) a -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n3 SELECT key, sum(keyD), value group by key, value -INSERT OVERWRITE TABLE e1 +INSERT OVERWRITE TABLE e1_n2 SELECT key, COUNT(distinct value) group by key; -select * from e1; -select * from e2; +select * from e1_n2; +select * from e2_n3; explain from src -insert overwrite table e1 +insert overwrite table e1_n2 select key, count(distinct value) group by key insert overwrite table e3 select value, count(distinct key) group by value; @@ -47,9 +47,9 @@ select value, count(distinct key) group by value; explain FROM (select key, cast(key as double) as keyD, value from src order by key) a -INSERT OVERWRITE TABLE e1 +INSERT OVERWRITE TABLE e1_n2 SELECT key, COUNT(distinct value) group by key -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n3 SELECT key, sum(keyD), value group by key, value INSERT overwrite table e3 SELECT key, COUNT(distinct keyD) group by key, keyD, value; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_gby4.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/multi_insert_gby4.q b/ql/src/test/queries/clientpositive/multi_insert_gby4.q index d485813..b89d674 100644 --- a/ql/src/test/queries/clientpositive/multi_insert_gby4.q +++ b/ql/src/test/queries/clientpositive/multi_insert_gby4.q @@ -1,27 +1,27 @@ --! qt:dataset:src -- SORT_QUERY_RESULTS -create table e1 (key string, count int); -create table e2 (key string, count int); -create table e3 (key string, count int); +create table e1_n4 (key string, count int); +create table e2_n5 (key string, count int); +create table e3_n0 (key string, count int); explain FROM (SELECT key, value FROM src) a -INSERT OVERWRITE TABLE e1 +INSERT OVERWRITE TABLE e1_n4 SELECT key, COUNT(*) WHERE key>450 GROUP BY key -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n5 SELECT key, COUNT(*) WHERE key>500 GROUP BY key -INSERT OVERWRITE TABLE e3 +INSERT OVERWRITE TABLE e3_n0 SELECT key, COUNT(*) WHERE key>490 GROUP BY key; FROM (SELECT key, value FROM src) a -INSERT OVERWRITE TABLE e1 +INSERT OVERWRITE TABLE e1_n4 SELECT key, COUNT(*) WHERE key>450 GROUP BY key -INSERT OVERWRITE TABLE e2 +INSERT OVERWRITE TABLE e2_n5 SELECT key, COUNT(*) WHERE key>500 GROUP BY key -INSERT OVERWRITE TABLE e3 +INSERT OVERWRITE TABLE e3_n0 SELECT key, COUNT(*) WHERE key>490 GROUP BY key; -select * from e1; -select * from e2; -select * from e3; +select * from e1_n4; +select * from e2_n5; +select * from e3_n0; http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q b/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q index 481bdf2..3442e82 100644 --- a/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q +++ b/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q @@ -2,7 +2,7 @@ set hive.stats.dbclass=fs; -- SORT_QUERY_RESULTS -create table src_10 as select * from src limit 10; +create table src_10_n0 as select * from src limit 10; create table src_lv1 (key string, value string); create table src_lv2 (key string, value string); @@ -14,11 +14,11 @@ create table src_lv3 (key string, value string); -- -LVF[6]-SEL[7]-LVJ[10]-SEL[13]-FS[14] -- -SEL[8]-UDTF[9]-LVJ[10] explain -from src_10 +from src_10_n0 insert overwrite table src_lv1 select key, C lateral view explode(array(key+1, key+2)) A as C insert overwrite table src_lv2 select key, C lateral view explode(array(key+3, key+4)) A as C; -from src_10 +from src_10_n0 insert overwrite table src_lv1 select key, C lateral view explode(array(key+1, key+2)) A as C insert overwrite table src_lv2 select key, C lateral view explode(array(key+3, key+4)) A as C; @@ -31,11 +31,11 @@ select * from src_lv2; -- -LVF[6]-SEL[7]-LVJ[10]-SEL[17]-GBY[18]-RS[19]-GBY[20]-SEL[21]-FS[22] -- -SEL[8]-UDTF[9]-LVJ[10] explain -from src_10 +from src_10_n0 insert overwrite table src_lv1 select key, sum(C) lateral view explode(array(key+1, key+2)) A as C group by key insert overwrite table src_lv2 select key, sum(C) lateral view explode(array(key+3, key+4)) A as C group by key; -from src_10 +from src_10_n0 insert overwrite table src_lv1 select key, sum(C) lateral view explode(array(key+1, key+2)) A as C group by key insert overwrite table src_lv2 select key, sum(C) lateral view explode(array(key+3, key+4)) A as C group by key; @@ -48,12 +48,12 @@ select * from src_lv2; -- -FIL[12]-SEL[13]-RS[14]-FOR[15]-FIL[16]-GBY[17]-SEL[18]-FS[19] -- -FIL[20]-GBY[21]-SEL[22]-FS[23] explain -from src_10 +from src_10_n0 insert overwrite table src_lv1 select key, sum(C) lateral view explode(array(key+1, key+2)) A as C group by key insert overwrite table src_lv2 select key, count(value) where key > 200 group by key insert overwrite table src_lv3 select key, count(value) where key < 200 group by key; -from src_10 +from src_10_n0 insert overwrite table src_lv1 select key, sum(C) lateral view explode(array(key+1, key+2)) A as C group by key insert overwrite table src_lv2 select key, count(value) where key > 200 group by key insert overwrite table src_lv3 select key, count(value) where key < 200 group by key; @@ -70,12 +70,12 @@ select * from src_lv3; -- -SEL[8]-UDTF[9]-LVJ[10] -- -SEL[23]-GBY[24]-RS[25]-GBY[26]-SEL[27]-FS[28] explain -from src_10 +from src_10_n0 insert overwrite table src_lv1 select C, sum(distinct key) lateral view explode(array(key+1, key+2)) A as C group by C insert overwrite table src_lv2 select C, sum(distinct key) lateral view explode(array(key+3, key+4)) A as C group by C insert overwrite table src_lv3 select value, sum(distinct key) group by value; -from src_10 +from src_10_n0 insert overwrite table src_lv1 select C, sum(distinct key) lateral view explode(array(key+1, key+2)) A as C group by C insert overwrite table src_lv2 select C, sum(distinct key) lateral view explode(array(key+3, key+4)) A as C group by C insert overwrite table src_lv3 select value, sum(distinct key) group by value; @@ -88,13 +88,13 @@ create table src_lv4 (key string, value string); -- Common distincts optimization works across non-lateral view queries, but not across lateral view multi inserts explain -from src_10 +from src_10_n0 insert overwrite table src_lv1 select key, sum(distinct C) lateral view explode(array(key+1, key+2)) A as C group by key insert overwrite table src_lv2 select key, sum(distinct C) lateral view explode(array(key+3, key+4)) A as C group by key insert overwrite table src_lv3 select value, sum(distinct key) where key > 200 group by value insert overwrite table src_lv4 select value, sum(distinct key) where key < 200 group by value; -from src_10 +from src_10_n0 insert overwrite table src_lv1 select key, sum(distinct C) lateral view explode(array(key+1, key+2)) A as C group by key insert overwrite table src_lv2 select key, sum(distinct C) lateral view explode(array(key+3, key+4)) A as C group by key insert overwrite table src_lv3 select value, sum(distinct key) where key > 200 group by value http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_mixed.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/multi_insert_mixed.q b/ql/src/test/queries/clientpositive/multi_insert_mixed.q index a5c1e87..9c8fb2e 100644 --- a/ql/src/test/queries/clientpositive/multi_insert_mixed.q +++ b/ql/src/test/queries/clientpositive/multi_insert_mixed.q @@ -1,7 +1,7 @@ --! qt:dataset:src set hive.mapred.mode=nonstrict; -create table src_multi1 like src; -create table src_multi2 like src; +create table src_multi1_n2 like src; +create table src_multi2_n3 like src; create table src_multi3 like src; set hive.stats.dbclass=fs; -- Testing the case where a map work contains both shuffling (ReduceSinkOperator) @@ -9,15 +9,15 @@ set hive.stats.dbclass=fs; explain from src -insert overwrite table src_multi1 select key, count(1) group by key order by key -insert overwrite table src_multi2 select value, count(1) group by value order by value +insert overwrite table src_multi1_n2 select key, count(1) group by key order by key +insert overwrite table src_multi2_n3 select value, count(1) group by value order by value insert overwrite table src_multi3 select * where key < 10; from src -insert overwrite table src_multi1 select key, count(1) group by key order by key -insert overwrite table src_multi2 select value, count(1) group by value order by value +insert overwrite table src_multi1_n2 select key, count(1) group by key order by key +insert overwrite table src_multi2_n3 select value, count(1) group by value order by value insert overwrite table src_multi3 select * where key < 10; -select * from src_multi1; -select * from src_multi2; +select * from src_multi1_n2; +select * from src_multi2_n3; select * from src_multi3;