http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/accumulo-handler/src/test/queries/positive/accumulo_custom_key.q ---------------------------------------------------------------------- diff --git a/accumulo-handler/src/test/queries/positive/accumulo_custom_key.q b/accumulo-handler/src/test/queries/positive/accumulo_custom_key.q index 6684fd3..ca7e6c7 100644 --- a/accumulo-handler/src/test/queries/positive/accumulo_custom_key.q +++ b/accumulo-handler/src/test/queries/positive/accumulo_custom_key.q @@ -1,3 +1,4 @@ +--! qt:dataset:src CREATE TABLE accumulo_ck_1(key struct<col1:string,col2:string,col3:string>, value string) STORED BY 'org.apache.hadoop.hive.accumulo.AccumuloStorageHandler' WITH SERDEPROPERTIES (
http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/accumulo-handler/src/test/queries/positive/accumulo_custom_key2.q ---------------------------------------------------------------------- diff --git a/accumulo-handler/src/test/queries/positive/accumulo_custom_key2.q b/accumulo-handler/src/test/queries/positive/accumulo_custom_key2.q index 038633f..615e5f8 100644 --- a/accumulo-handler/src/test/queries/positive/accumulo_custom_key2.q +++ b/accumulo-handler/src/test/queries/positive/accumulo_custom_key2.q @@ -1,3 +1,4 @@ +--! qt:dataset:src CREATE TABLE accumulo_ck_3(key struct<col1:string,col2:string,col3:string>, value string) STORED BY 'org.apache.hadoop.hive.accumulo.AccumuloStorageHandler' WITH SERDEPROPERTIES ( http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/accumulo-handler/src/test/queries/positive/accumulo_joins.q ---------------------------------------------------------------------- diff --git a/accumulo-handler/src/test/queries/positive/accumulo_joins.q b/accumulo-handler/src/test/queries/positive/accumulo_joins.q index b72ec6b..a6b09ab 100644 --- a/accumulo-handler/src/test/queries/positive/accumulo_joins.q +++ b/accumulo-handler/src/test/queries/positive/accumulo_joins.q @@ -1,3 +1,4 @@ +--! qt:dataset:src DROP TABLE users; DROP TABLE states; DROP TABLE countries; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/accumulo-handler/src/test/queries/positive/accumulo_predicate_pushdown.q ---------------------------------------------------------------------- diff --git a/accumulo-handler/src/test/queries/positive/accumulo_predicate_pushdown.q b/accumulo-handler/src/test/queries/positive/accumulo_predicate_pushdown.q index 0f064af..4cb93b5 100644 --- a/accumulo-handler/src/test/queries/positive/accumulo_predicate_pushdown.q +++ b/accumulo-handler/src/test/queries/positive/accumulo_predicate_pushdown.q @@ -1,3 +1,5 @@ +--! qt:dataset:src +--! qt:dataset:part CREATE TABLE accumulo_pushdown(key string, value string) STORED BY 'org.apache.hadoop.hive.accumulo.AccumuloStorageHandler' WITH SERDEPROPERTIES ("accumulo.columns.mapping" = ":rowid,cf:string"); http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/accumulo-handler/src/test/queries/positive/accumulo_queries.q ---------------------------------------------------------------------- diff --git a/accumulo-handler/src/test/queries/positive/accumulo_queries.q b/accumulo-handler/src/test/queries/positive/accumulo_queries.q index 0aceaa0..bc93d8a 100644 --- a/accumulo-handler/src/test/queries/positive/accumulo_queries.q +++ b/accumulo-handler/src/test/queries/positive/accumulo_queries.q @@ -1,3 +1,4 @@ +--! qt:dataset:src -- remove these; after HIVE-18802 is fixed set hive.optimize.index.filter=false; set hive.optimize.ppd=false; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/accumulo-handler/src/test/queries/positive/accumulo_single_sourced_multi_insert.q ---------------------------------------------------------------------- diff --git a/accumulo-handler/src/test/queries/positive/accumulo_single_sourced_multi_insert.q b/accumulo-handler/src/test/queries/positive/accumulo_single_sourced_multi_insert.q index f904d3f..c12f962 100644 --- a/accumulo-handler/src/test/queries/positive/accumulo_single_sourced_multi_insert.q +++ b/accumulo-handler/src/test/queries/positive/accumulo_single_sourced_multi_insert.q @@ -1,3 +1,4 @@ +--! qt:dataset:src -- HIVE-4375 Single sourced multi insert consists of native and non-native table mixed throws NPE CREATE TABLE src_x1(key string, value string); CREATE TABLE src_x2(key string, value string) http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientnegative/case_with_row_sequence.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientnegative/case_with_row_sequence.q b/contrib/src/test/queries/clientnegative/case_with_row_sequence.q index 910ffda..9922092 100644 --- a/contrib/src/test/queries/clientnegative/case_with_row_sequence.q +++ b/contrib/src/test/queries/clientnegative/case_with_row_sequence.q @@ -1,3 +1,4 @@ +--! qt:dataset:src set hive.exec.submitviachild=true; set hive.exec.submit.local.task.via.child=true; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientnegative/invalid_row_sequence.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientnegative/invalid_row_sequence.q b/contrib/src/test/queries/clientnegative/invalid_row_sequence.q index 1de9d9e..b797fef 100644 --- a/contrib/src/test/queries/clientnegative/invalid_row_sequence.q +++ b/contrib/src/test/queries/clientnegative/invalid_row_sequence.q @@ -1,3 +1,4 @@ +--! qt:dataset:src -- Verify that a stateful UDF cannot be used outside of the SELECT list drop temporary function row_sequence; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientnegative/udtf_explode2.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientnegative/udtf_explode2.q b/contrib/src/test/queries/clientnegative/udtf_explode2.q index 29217eb..f202d77 100644 --- a/contrib/src/test/queries/clientnegative/udtf_explode2.q +++ b/contrib/src/test/queries/clientnegative/udtf_explode2.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; CREATE TEMPORARY FUNCTION explode2 AS 'org.apache.hadoop.hive.contrib.udtf.example.GenericUDTFExplode2'; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/dboutput.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/dboutput.q b/contrib/src/test/queries/clientpositive/dboutput.q index 2b20738..4798cb2 100644 --- a/contrib/src/test/queries/clientpositive/dboutput.q +++ b/contrib/src/test/queries/clientpositive/dboutput.q @@ -1,3 +1,4 @@ +--! qt:dataset:src set hive.mapred.mode=nonstrict; ADD JAR ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/fileformat_base64.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/fileformat_base64.q b/contrib/src/test/queries/clientpositive/fileformat_base64.q index 20bbfa8..f52669f 100644 --- a/contrib/src/test/queries/clientpositive/fileformat_base64.q +++ b/contrib/src/test/queries/clientpositive/fileformat_base64.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; DROP TABLE base64_test; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/java_mr_example.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/java_mr_example.q b/contrib/src/test/queries/clientpositive/java_mr_example.q index 06911dc..3b82c18 100644 --- a/contrib/src/test/queries/clientpositive/java_mr_example.q +++ b/contrib/src/test/queries/clientpositive/java_mr_example.q @@ -1,3 +1,4 @@ +--! qt:dataset:src FROM ( FROM src MAP value, key http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/lateral_view_explode2.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/lateral_view_explode2.q b/contrib/src/test/queries/clientpositive/lateral_view_explode2.q index 210946e..edfff44 100644 --- a/contrib/src/test/queries/clientpositive/lateral_view_explode2.q +++ b/contrib/src/test/queries/clientpositive/lateral_view_explode2.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; CREATE TEMPORARY FUNCTION explode2 AS 'org.apache.hadoop.hive.contrib.udtf.example.GenericUDTFExplode2'; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/serde_typedbytes.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/serde_typedbytes.q b/contrib/src/test/queries/clientpositive/serde_typedbytes.q index d0765fd..29e3ba1 100644 --- a/contrib/src/test/queries/clientpositive/serde_typedbytes.q +++ b/contrib/src/test/queries/clientpositive/serde_typedbytes.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; drop table dest1; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/serde_typedbytes2.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/serde_typedbytes2.q b/contrib/src/test/queries/clientpositive/serde_typedbytes2.q index a709558..069cfd4 100644 --- a/contrib/src/test/queries/clientpositive/serde_typedbytes2.q +++ b/contrib/src/test/queries/clientpositive/serde_typedbytes2.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; drop table dest1; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/serde_typedbytes3.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/serde_typedbytes3.q b/contrib/src/test/queries/clientpositive/serde_typedbytes3.q index 492c576..68917e6 100644 --- a/contrib/src/test/queries/clientpositive/serde_typedbytes3.q +++ b/contrib/src/test/queries/clientpositive/serde_typedbytes3.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; drop table dest1; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/serde_typedbytes4.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/serde_typedbytes4.q b/contrib/src/test/queries/clientpositive/serde_typedbytes4.q index bb1f257..49c9d7b 100644 --- a/contrib/src/test/queries/clientpositive/serde_typedbytes4.q +++ b/contrib/src/test/queries/clientpositive/serde_typedbytes4.q @@ -1,3 +1,4 @@ +--! qt:dataset:src set hive.mapred.mode=nonstrict; add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/serde_typedbytes5.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/serde_typedbytes5.q b/contrib/src/test/queries/clientpositive/serde_typedbytes5.q index f658c46..5978be1 100644 --- a/contrib/src/test/queries/clientpositive/serde_typedbytes5.q +++ b/contrib/src/test/queries/clientpositive/serde_typedbytes5.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; drop table dest1; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/serde_typedbytes_null.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/serde_typedbytes_null.q b/contrib/src/test/queries/clientpositive/serde_typedbytes_null.q index 59b757b..404dc65 100644 --- a/contrib/src/test/queries/clientpositive/serde_typedbytes_null.q +++ b/contrib/src/test/queries/clientpositive/serde_typedbytes_null.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; DROP TABLE table1; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/udaf_example_avg.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/udaf_example_avg.q b/contrib/src/test/queries/clientpositive/udaf_example_avg.q index 0f69926..d18b197 100644 --- a/contrib/src/test/queries/clientpositive/udaf_example_avg.q +++ b/contrib/src/test/queries/clientpositive/udaf_example_avg.q @@ -1,3 +1,4 @@ +--! qt:dataset:src set hive.mapred.mode=nonstrict; add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/udaf_example_group_concat.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/udaf_example_group_concat.q b/contrib/src/test/queries/clientpositive/udaf_example_group_concat.q index 817cf32..56e7164 100644 --- a/contrib/src/test/queries/clientpositive/udaf_example_group_concat.q +++ b/contrib/src/test/queries/clientpositive/udaf_example_group_concat.q @@ -1,3 +1,4 @@ +--! qt:dataset:src set hive.mapred.mode=nonstrict; add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/udaf_example_max.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/udaf_example_max.q b/contrib/src/test/queries/clientpositive/udaf_example_max.q index 7db78a7..2b72791 100644 --- a/contrib/src/test/queries/clientpositive/udaf_example_max.q +++ b/contrib/src/test/queries/clientpositive/udaf_example_max.q @@ -1,3 +1,4 @@ +--! qt:dataset:src set hive.mapred.mode=nonstrict; add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/udaf_example_max_n.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/udaf_example_max_n.q b/contrib/src/test/queries/clientpositive/udaf_example_max_n.q index 4cc2878..f4e5596 100644 --- a/contrib/src/test/queries/clientpositive/udaf_example_max_n.q +++ b/contrib/src/test/queries/clientpositive/udaf_example_max_n.q @@ -1,3 +1,4 @@ +--! qt:dataset:src set hive.mapred.mode=nonstrict; add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/udaf_example_min.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/udaf_example_min.q b/contrib/src/test/queries/clientpositive/udaf_example_min.q index bd94252..32ae4aa 100644 --- a/contrib/src/test/queries/clientpositive/udaf_example_min.q +++ b/contrib/src/test/queries/clientpositive/udaf_example_min.q @@ -1,3 +1,4 @@ +--! qt:dataset:src set hive.mapred.mode=nonstrict; add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/udaf_example_min_n.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/udaf_example_min_n.q b/contrib/src/test/queries/clientpositive/udaf_example_min_n.q index 12bf1ec..ca483eb 100644 --- a/contrib/src/test/queries/clientpositive/udaf_example_min_n.q +++ b/contrib/src/test/queries/clientpositive/udaf_example_min_n.q @@ -1,3 +1,4 @@ +--! qt:dataset:src set hive.mapred.mode=nonstrict; add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; CREATE TEMPORARY FUNCTION example_min_n AS 'org.apache.hadoop.hive.contrib.udaf.example.UDAFExampleMinN'; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/udf_example_add.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/udf_example_add.q b/contrib/src/test/queries/clientpositive/udf_example_add.q index fb7b5c9..e099935 100644 --- a/contrib/src/test/queries/clientpositive/udf_example_add.q +++ b/contrib/src/test/queries/clientpositive/udf_example_add.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; CREATE TEMPORARY FUNCTION example_add AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleAdd'; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/udf_example_arraymapstruct.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/udf_example_arraymapstruct.q b/contrib/src/test/queries/clientpositive/udf_example_arraymapstruct.q index a8ab751..f097958 100644 --- a/contrib/src/test/queries/clientpositive/udf_example_arraymapstruct.q +++ b/contrib/src/test/queries/clientpositive/udf_example_arraymapstruct.q @@ -1,3 +1,4 @@ +--! qt:dataset:src_thrift -- Suppress vectorization due to known bug. See HIVE-19110. set hive.vectorized.execution.enabled=false; set hive.test.vectorized.execution.enabled.override=disable; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/udf_example_format.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/udf_example_format.q b/contrib/src/test/queries/clientpositive/udf_example_format.q index edda00b..e42bab7 100644 --- a/contrib/src/test/queries/clientpositive/udf_example_format.q +++ b/contrib/src/test/queries/clientpositive/udf_example_format.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; CREATE TEMPORARY FUNCTION example_format AS 'org.apache.hadoop.hive.contrib.udf.example.UDFExampleFormat'; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/udf_row_sequence.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/udf_row_sequence.q b/contrib/src/test/queries/clientpositive/udf_row_sequence.q index 44099b1..7a3a386 100644 --- a/contrib/src/test/queries/clientpositive/udf_row_sequence.q +++ b/contrib/src/test/queries/clientpositive/udf_row_sequence.q @@ -1,3 +1,4 @@ +--! qt:dataset:src -- The ORDER BY on the outer query will typically have no effect, -- but there is really no guarantee that the ordering is preserved -- across various SQL operators. http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/udtf_explode2.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/udtf_explode2.q b/contrib/src/test/queries/clientpositive/udtf_explode2.q index 2bd1359..8c306cf 100644 --- a/contrib/src/test/queries/clientpositive/udtf_explode2.q +++ b/contrib/src/test/queries/clientpositive/udtf_explode2.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; CREATE TEMPORARY FUNCTION explode2 AS 'org.apache.hadoop.hive.contrib.udtf.example.GenericUDTFExplode2'; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/udtf_output_on_close.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/udtf_output_on_close.q b/contrib/src/test/queries/clientpositive/udtf_output_on_close.q index 87aec5e..08762aa 100644 --- a/contrib/src/test/queries/clientpositive/udtf_output_on_close.q +++ b/contrib/src/test/queries/clientpositive/udtf_output_on_close.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; CREATE TEMPORARY FUNCTION udtfCount2 AS 'org.apache.hadoop.hive.contrib.udtf.example.GenericUDTFCount2'; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/contrib/src/test/queries/clientpositive/url_hook.q ---------------------------------------------------------------------- diff --git a/contrib/src/test/queries/clientpositive/url_hook.q b/contrib/src/test/queries/clientpositive/url_hook.q index 1d56408..512e579 100644 --- a/contrib/src/test/queries/clientpositive/url_hook.q +++ b/contrib/src/test/queries/clientpositive/url_hook.q @@ -1,3 +1,4 @@ +--! qt:dataset:src add jar ${system:maven.local.repository}/org/apache/hive/hive-contrib/${system:hive.version}/hive-contrib-${system:hive.version}.jar; SHOW TABLES 'src'; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/alltypesorc/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/alltypesorc/load.hive.sql b/data/files/datasets/alltypesorc/load.hive.sql new file mode 100644 index 0000000..87f9d56 --- /dev/null +++ b/data/files/datasets/alltypesorc/load.hive.sql @@ -0,0 +1,21 @@ +CREATE TABLE alltypesorc( + ctinyint TINYINT, + csmallint SMALLINT, + cint INT, + cbigint BIGINT, + cfloat FLOAT, + cdouble DOUBLE, + cstring1 STRING, + cstring2 STRING, + ctimestamp1 TIMESTAMP, + ctimestamp2 TIMESTAMP, + cboolean1 BOOLEAN, + cboolean2 BOOLEAN) + STORED AS ORC; + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/alltypesorc" +OVERWRITE INTO TABLE alltypesorc; + +ANALYZE TABLE alltypesorc COMPUTE STATISTICS; + +ANALYZE TABLE alltypesorc COMPUTE STATISTICS FOR COLUMNS ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/alltypesparquet/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/alltypesparquet/load.hive.sql b/data/files/datasets/alltypesparquet/load.hive.sql new file mode 100644 index 0000000..45d3616 --- /dev/null +++ b/data/files/datasets/alltypesparquet/load.hive.sql @@ -0,0 +1,39 @@ +CREATE TABLE alltypesorc_to_parquet( + ctinyint TINYINT, + csmallint SMALLINT, + cint INT, + cbigint BIGINT, + cfloat FLOAT, + cdouble DOUBLE, + cstring1 STRING, + cstring2 STRING, + ctimestamp1 TIMESTAMP, + ctimestamp2 TIMESTAMP, + cboolean1 BOOLEAN, + cboolean2 BOOLEAN) + STORED AS ORC; + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/alltypesorc" +OVERWRITE INTO TABLE alltypesorc_to_parquet; + +CREATE TABLE alltypesparquet( + ctinyint TINYINT, + csmallint SMALLINT, + cint INT, + cbigint BIGINT, + cfloat FLOAT, + cdouble DOUBLE, + cstring1 STRING, + cstring2 STRING, + ctimestamp1 TIMESTAMP, + ctimestamp2 TIMESTAMP, + cboolean1 BOOLEAN, + cboolean2 BOOLEAN) + STORED AS PARQUET; + +INSERT OVERWRITE TABLE alltypesparquet SELECT * FROM alltypesorc_to_parquet; + +ANALYZE TABLE alltypesparquet COMPUTE STATISTICS; + +ANALYZE TABLE alltypesparquet COMPUTE STATISTICS FOR COLUMNS ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2; +DROP TABLE alltypesorc_to_parquet; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/cbo_t1/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/cbo_t1/load.hive.sql b/data/files/datasets/cbo_t1/load.hive.sql new file mode 100644 index 0000000..257f0bd --- /dev/null +++ b/data/files/datasets/cbo_t1/load.hive.sql @@ -0,0 +1,8 @@ +set hive.cbo.enable=true; + +create table cbo_t1(key string, value string, c_int int, c_float float, c_boolean boolean) partitioned by (dt string) row format delimited fields terminated by ',' STORED AS TEXTFILE; + +load data local inpath '${hiveconf:test.data.dir}/cbo_t1.txt' into table cbo_t1 partition (dt='2014'); + +analyze table cbo_t1 partition (dt) compute statistics; +analyze table cbo_t1 compute statistics for columns key, value, c_int, c_float, c_boolean; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/cbo_t2/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/cbo_t2/load.hive.sql b/data/files/datasets/cbo_t2/load.hive.sql new file mode 100644 index 0000000..293bae2 --- /dev/null +++ b/data/files/datasets/cbo_t2/load.hive.sql @@ -0,0 +1,8 @@ +set hive.cbo.enable=true; + +create table cbo_t2(key string, value string, c_int int, c_float float, c_boolean boolean) partitioned by (dt string) row format delimited fields terminated by ',' STORED AS TEXTFILE; + +load data local inpath '${hiveconf:test.data.dir}/cbo_t2.txt' into table cbo_t2 partition (dt='2014'); + +analyze table cbo_t2 partition (dt) compute statistics; +analyze table cbo_t2 compute statistics for columns key, value, c_int, c_float, c_boolean; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/cbo_t3/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/cbo_t3/load.hive.sql b/data/files/datasets/cbo_t3/load.hive.sql new file mode 100644 index 0000000..6c7c3e0 --- /dev/null +++ b/data/files/datasets/cbo_t3/load.hive.sql @@ -0,0 +1,8 @@ +set hive.cbo.enable=true; + +create table cbo_t3(key string, value string, c_int int, c_float float, c_boolean boolean) row format delimited fields terminated by ',' STORED AS TEXTFILE; + +load data local inpath '${hiveconf:test.data.dir}/cbo_t3.txt' into table cbo_t3; + +analyze table cbo_t3 compute statistics; +analyze table cbo_t3 compute statistics for columns key, value, c_int, c_float, c_boolean; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/lineitem/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/lineitem/load.hive.sql b/data/files/datasets/lineitem/load.hive.sql new file mode 100644 index 0000000..ab2f5fb --- /dev/null +++ b/data/files/datasets/lineitem/load.hive.sql @@ -0,0 +1,23 @@ +CREATE TABLE lineitem (L_ORDERKEY INT, + L_PARTKEY INT, + L_SUPPKEY INT, + L_LINENUMBER INT, + L_QUANTITY DOUBLE, + L_EXTENDEDPRICE DOUBLE, + L_DISCOUNT DOUBLE, + L_TAX DOUBLE, + L_RETURNFLAG STRING, + L_LINESTATUS STRING, + l_shipdate STRING, + L_COMMITDATE STRING, + L_RECEIPTDATE STRING, + L_SHIPINSTRUCT STRING, + L_SHIPMODE STRING, + L_COMMENT STRING) +ROW FORMAT DELIMITED +FIELDS TERMINATED BY '|'; + +LOAD DATA LOCAL INPATH '${hiveconf:test.data.dir}/lineitem.txt' OVERWRITE INTO TABLE lineitem; + +analyze table lineitem compute statistics; +analyze table lineitem compute statistics for columns; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/part/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/part/load.hive.sql b/data/files/datasets/part/load.hive.sql new file mode 100644 index 0000000..c24b505 --- /dev/null +++ b/data/files/datasets/part/load.hive.sql @@ -0,0 +1,16 @@ +CREATE TABLE part( + p_partkey INT, + p_name STRING, + p_mfgr STRING, + p_brand STRING, + p_type STRING, + p_size INT, + p_container STRING, + p_retailprice DOUBLE, + p_comment STRING +); + +LOAD DATA LOCAL INPATH '${hiveconf:test.data.dir}/part_tiny.txt' overwrite into table part; + +analyze table part compute statistics; +analyze table part compute statistics for columns; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/src/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/src/load.hive.sql b/data/files/datasets/src/load.hive.sql new file mode 100644 index 0000000..eca7e07 --- /dev/null +++ b/data/files/datasets/src/load.hive.sql @@ -0,0 +1,7 @@ +CREATE TABLE src (key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS TEXTFILE; + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" INTO TABLE src; + +ANALYZE TABLE src COMPUTE STATISTICS; + +ANALYZE TABLE src COMPUTE STATISTICS FOR COLUMNS key,value; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/src1/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/src1/load.hive.sql b/data/files/datasets/src1/load.hive.sql new file mode 100644 index 0000000..b64d45f --- /dev/null +++ b/data/files/datasets/src1/load.hive.sql @@ -0,0 +1,7 @@ +CREATE TABLE src1 (key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS TEXTFILE; + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv3.txt" INTO TABLE src1; + +ANALYZE TABLE src1 COMPUTE STATISTICS; + +ANALYZE TABLE src1 COMPUTE STATISTICS FOR COLUMNS key,value; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/src_cbo/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/src_cbo/load.hive.sql b/data/files/datasets/src_cbo/load.hive.sql new file mode 100644 index 0000000..570c316 --- /dev/null +++ b/data/files/datasets/src_cbo/load.hive.sql @@ -0,0 +1,8 @@ +set hive.cbo.enable=true; + +CREATE TABLE src_cbo (key STRING, value STRING) STORED AS TEXTFILE; + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" INTO TABLE src_cbo; + +analyze table src_cbo compute statistics; +analyze table src_cbo compute statistics for columns; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/src_hbase/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/src_hbase/load.hive.sql b/data/files/datasets/src_hbase/load.hive.sql new file mode 100644 index 0000000..5c7b5b7 --- /dev/null +++ b/data/files/datasets/src_hbase/load.hive.sql @@ -0,0 +1,14 @@ +CREATE TABLE src_hbase_tmp (key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS TEXTFILE; + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" INTO TABLE src_hbase_tmp; + +ANALYZE TABLE src_hbase_tmp COMPUTE STATISTICS; + +ANALYZE TABLE src_hbase_tmp COMPUTE STATISTICS FOR COLUMNS key,value; + +CREATE TABLE src_hbase (key INT, value STRING) +STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' +WITH SERDEPROPERTIES ('hbase.columns.mapping' = ':key,cf:val') +TBLPROPERTIES ('hbase.table.name' = 'src_hbase'); + +INSERT OVERWRITE TABLE src_hbase SELECT * FROM src_hbase_tmp http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/src_json/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/src_json/load.hive.sql b/data/files/datasets/src_json/load.hive.sql new file mode 100644 index 0000000..13e0b96 --- /dev/null +++ b/data/files/datasets/src_json/load.hive.sql @@ -0,0 +1,7 @@ +CREATE TABLE src_json (json STRING COMMENT 'default') STORED AS TEXTFILE; + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/json.txt" INTO TABLE src_json; + +ANALYZE TABLE src_json COMPUTE STATISTICS; + +ANALYZE TABLE src_json COMPUTE STATISTICS FOR COLUMNS json; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/src_sequencefile/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/src_sequencefile/load.hive.sql b/data/files/datasets/src_sequencefile/load.hive.sql new file mode 100644 index 0000000..6f10562 --- /dev/null +++ b/data/files/datasets/src_sequencefile/load.hive.sql @@ -0,0 +1,7 @@ +CREATE TABLE src_sequencefile (key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS SEQUENCEFILE; + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.seq" INTO TABLE src_sequencefile; + +ANALYZE TABLE src_sequencefile COMPUTE STATISTICS; + +ANALYZE TABLE src_sequencefile COMPUTE STATISTICS FOR COLUMNS key,value; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/src_thrift/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/src_thrift/load.hive.sql b/data/files/datasets/src_thrift/load.hive.sql new file mode 100644 index 0000000..1e1f1b5 --- /dev/null +++ b/data/files/datasets/src_thrift/load.hive.sql @@ -0,0 +1,10 @@ +CREATE TABLE src_thrift +ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer' +WITH SERDEPROPERTIES ( + 'serialization.class' = 'org.apache.hadoop.hive.serde2.thrift.test.Complex', + 'serialization.format' = 'org.apache.thrift.protocol.TBinaryProtocol') +STORED AS SEQUENCEFILE; + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/complex.seq" INTO TABLE src_thrift; + +ANALYZE TABLE src_thrift COMPUTE STATISTICS; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/srcbucket/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/srcbucket/load.hive.sql b/data/files/datasets/srcbucket/load.hive.sql new file mode 100644 index 0000000..4008115 --- /dev/null +++ b/data/files/datasets/srcbucket/load.hive.sql @@ -0,0 +1,13 @@ +CREATE TABLE srcbucket_tmp (key INT, value STRING) STORED AS TEXTFILE; +CREATE TABLE srcbucket (key INT, value STRING) +CLUSTERED BY (key) INTO 2 BUCKETS +STORED AS TEXTFILE; + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/bucketed_files/000000_0" INTO TABLE srcbucket_tmp; +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/bucketed_files/000001_0" INTO TABLE srcbucket_tmp; +INSERT INTO srcbucket SELECT * FROM srcbucket_tmp; +DROP TABLE srcbucket_tmp; + +ANALYZE TABLE srcbucket COMPUTE STATISTICS; + +ANALYZE TABLE srcbucket COMPUTE STATISTICS FOR COLUMNS key,value; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/srcbucket2/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/srcbucket2/load.hive.sql b/data/files/datasets/srcbucket2/load.hive.sql new file mode 100644 index 0000000..de00b2f --- /dev/null +++ b/data/files/datasets/srcbucket2/load.hive.sql @@ -0,0 +1,15 @@ +CREATE TABLE srcbucket_tmp (key INT, value STRING); +CREATE TABLE srcbucket2 (key INT, value STRING) +CLUSTERED BY (key) INTO 4 BUCKETS +STORED AS TEXTFILE; + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/bmj/000000_0" INTO TABLE srcbucket_tmp; +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/bmj/000001_0" INTO TABLE srcbucket_tmp; +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/bmj/000002_0" INTO TABLE srcbucket_tmp; +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/bmj/000003_0" INTO TABLE srcbucket_tmp; +INSERT INTO srcbucket2 SELECT * FROM srcbucket_tmp; +DROP TABLE srcbucket_tmp; + +ANALYZE TABLE srcbucket2 COMPUTE STATISTICS; + +ANALYZE TABLE srcbucket2 COMPUTE STATISTICS FOR COLUMNS key,value; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/files/datasets/srcpart/load.hive.sql ---------------------------------------------------------------------- diff --git a/data/files/datasets/srcpart/load.hive.sql b/data/files/datasets/srcpart/load.hive.sql new file mode 100644 index 0000000..f94d2a3 --- /dev/null +++ b/data/files/datasets/srcpart/load.hive.sql @@ -0,0 +1,19 @@ +CREATE TABLE srcpart (key STRING COMMENT 'default', value STRING COMMENT 'default') +PARTITIONED BY (ds STRING, hr STRING) +STORED AS TEXTFILE; + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" +OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-08", hr="11"); + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" +OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-08", hr="12"); + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" +OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-09", hr="11"); + +LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" +OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-09", hr="12"); + +ANALYZE TABLE srcpart PARTITION(ds, hr) COMPUTE STATISTICS; + +ANALYZE TABLE srcpart PARTITION(ds, hr) COMPUTE STATISTICS FOR COLUMNS key,value; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/scripts/q_test_druid_init.sql ---------------------------------------------------------------------- diff --git a/data/scripts/q_test_druid_init.sql b/data/scripts/q_test_druid_init.sql index ee025f1..e69de29 100644 --- a/data/scripts/q_test_druid_init.sql +++ b/data/scripts/q_test_druid_init.sql @@ -1,29 +0,0 @@ -set hive.stats.dbclass=fs; --- --- Table alltypesorc --- -DROP TABLE IF EXISTS alltypesorc; -CREATE TABLE alltypesorc( - ctinyint TINYINT, - csmallint SMALLINT, - cint INT, - cbigint BIGINT, - cfloat FLOAT, - cdouble DOUBLE, - cstring1 STRING, - cstring2 STRING, - ctimestamp1 TIMESTAMP, - ctimestamp2 TIMESTAMP, - cboolean1 BOOLEAN, - cboolean2 BOOLEAN) - STORED AS ORC; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/alltypesorc" -OVERWRITE INTO TABLE alltypesorc; - -ANALYZE TABLE alltypesorc COMPUTE STATISTICS; - -ANALYZE TABLE alltypesorc COMPUTE STATISTICS FOR COLUMNS ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2; - --- Druid Table - http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/scripts/q_test_init.sql ---------------------------------------------------------------------- diff --git a/data/scripts/q_test_init.sql b/data/scripts/q_test_init.sql index 5d36b6f..a269c55 100644 --- a/data/scripts/q_test_init.sql +++ b/data/scripts/q_test_init.sql @@ -1,224 +1,4 @@ set hive.stats.dbclass=fs; --- --- Table src --- -DROP TABLE IF EXISTS src; - -CREATE TABLE src (key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" INTO TABLE src; - -ANALYZE TABLE src COMPUTE STATISTICS; - -ANALYZE TABLE src COMPUTE STATISTICS FOR COLUMNS key,value; - --- --- Table src1 --- -DROP TABLE IF EXISTS src1; - -CREATE TABLE src1 (key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv3.txt" INTO TABLE src1; - -ANALYZE TABLE src1 COMPUTE STATISTICS; - -ANALYZE TABLE src1 COMPUTE STATISTICS FOR COLUMNS key,value; - --- --- Table src_json --- -DROP TABLE IF EXISTS src_json; - -CREATE TABLE src_json (json STRING COMMENT 'default') STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/json.txt" INTO TABLE src_json; - -ANALYZE TABLE src_json COMPUTE STATISTICS; - -ANALYZE TABLE src_json COMPUTE STATISTICS FOR COLUMNS json; - --- --- Table src_sequencefile --- -DROP TABLE IF EXISTS src_sequencefile; - -CREATE TABLE src_sequencefile (key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS SEQUENCEFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.seq" INTO TABLE src_sequencefile; - -ANALYZE TABLE src_sequencefile COMPUTE STATISTICS; - -ANALYZE TABLE src_sequencefile COMPUTE STATISTICS FOR COLUMNS key,value; - --- --- Table src_thrift --- -DROP TABLE IF EXISTS src_thrift; - -CREATE TABLE src_thrift -ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer' -WITH SERDEPROPERTIES ( - 'serialization.class' = 'org.apache.hadoop.hive.serde2.thrift.test.Complex', - 'serialization.format' = 'org.apache.thrift.protocol.TBinaryProtocol') -STORED AS SEQUENCEFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/complex.seq" INTO TABLE src_thrift; - -ANALYZE TABLE src_thrift COMPUTE STATISTICS; - --- --- Table srcbucket --- -DROP TABLE IF EXISTS srcbucket_tmp; -DROP TABLE IF EXISTS srcbucket; - -CREATE TABLE srcbucket_tmp (key INT, value STRING) STORED AS TEXTFILE; -CREATE TABLE srcbucket (key INT, value STRING) -CLUSTERED BY (key) INTO 2 BUCKETS -STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/bucketed_files/000000_0" INTO TABLE srcbucket_tmp; -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/bucketed_files/000001_0" INTO TABLE srcbucket_tmp; -INSERT INTO srcbucket SELECT * FROM srcbucket_tmp; -DROP TABLE srcbucket_tmp; - -ANALYZE TABLE srcbucket COMPUTE STATISTICS; - -ANALYZE TABLE srcbucket COMPUTE STATISTICS FOR COLUMNS key,value; - --- --- Table srcbucket2 --- -DROP TABLE IF EXISTS srcbucket_tmp; -DROP TABLE IF EXISTS srcbucket2; - -CREATE TABLE srcbucket_tmp (key INT, value STRING); -CREATE TABLE srcbucket2 (key INT, value STRING) -CLUSTERED BY (key) INTO 4 BUCKETS -STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/bmj/000000_0" INTO TABLE srcbucket_tmp; -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/bmj/000001_0" INTO TABLE srcbucket_tmp; -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/bmj/000002_0" INTO TABLE srcbucket_tmp; -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/bmj/000003_0" INTO TABLE srcbucket_tmp; -INSERT INTO srcbucket2 SELECT * FROM srcbucket_tmp; -DROP TABLE srcbucket_tmp; - -ANALYZE TABLE srcbucket2 COMPUTE STATISTICS; - -ANALYZE TABLE srcbucket2 COMPUTE STATISTICS FOR COLUMNS key,value; - --- --- Table srcpart --- -DROP TABLE IF EXISTS srcpart; - -CREATE TABLE srcpart (key STRING COMMENT 'default', value STRING COMMENT 'default') -PARTITIONED BY (ds STRING, hr STRING) -STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" -OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-08", hr="11"); - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" -OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-08", hr="12"); - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" -OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-09", hr="11"); - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" -OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-09", hr="12"); - -ANALYZE TABLE srcpart PARTITION(ds, hr) COMPUTE STATISTICS; - -ANALYZE TABLE srcpart PARTITION(ds, hr) COMPUTE STATISTICS FOR COLUMNS key,value; - --- --- Table alltypesorc --- -DROP TABLE IF EXISTS alltypesorc; -CREATE TABLE alltypesorc( - ctinyint TINYINT, - csmallint SMALLINT, - cint INT, - cbigint BIGINT, - cfloat FLOAT, - cdouble DOUBLE, - cstring1 STRING, - cstring2 STRING, - ctimestamp1 TIMESTAMP, - ctimestamp2 TIMESTAMP, - cboolean1 BOOLEAN, - cboolean2 BOOLEAN) - STORED AS ORC; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/alltypesorc" -OVERWRITE INTO TABLE alltypesorc; - -ANALYZE TABLE alltypesorc COMPUTE STATISTICS; - -ANALYZE TABLE alltypesorc COMPUTE STATISTICS FOR COLUMNS ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2; - --- --- Table alltypesparquet --- -DROP TABLE IF EXISTS alltypesparquet; -CREATE TABLE alltypesparquet( - ctinyint TINYINT, - csmallint SMALLINT, - cint INT, - cbigint BIGINT, - cfloat FLOAT, - cdouble DOUBLE, - cstring1 STRING, - cstring2 STRING, - ctimestamp1 TIMESTAMP, - ctimestamp2 TIMESTAMP, - cboolean1 BOOLEAN, - cboolean2 BOOLEAN) - STORED AS PARQUET; - -INSERT OVERWRITE TABLE alltypesparquet SELECT * FROM alltypesorc; - -ANALYZE TABLE alltypesparquet COMPUTE STATISTICS; - -ANALYZE TABLE alltypesparquet COMPUTE STATISTICS FOR COLUMNS ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2; - - --- --- Table primitives --- -DROP TABLE IF EXISTS primitives; -CREATE TABLE primitives ( - id INT COMMENT 'default', - bool_col BOOLEAN COMMENT 'default', - tinyint_col TINYINT COMMENT 'default', - smallint_col SMALLINT COMMENT 'default', - int_col INT COMMENT 'default', - bigint_col BIGINT COMMENT 'default', - float_col FLOAT COMMENT 'default', - double_col DOUBLE COMMENT 'default', - date_string_col STRING COMMENT 'default', - string_col STRING COMMENT 'default', - timestamp_col TIMESTAMP COMMENT 'default') -PARTITIONED BY (year INT COMMENT 'default', month INT COMMENT 'default') -ROW FORMAT DELIMITED - FIELDS TERMINATED BY ',' - ESCAPED BY '\\' -STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/types/primitives/090101.txt" -OVERWRITE INTO TABLE primitives PARTITION(year=2009, month=1); - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/types/primitives/090201.txt" -OVERWRITE INTO TABLE primitives PARTITION(year=2009, month=2); - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/types/primitives/090301.txt" -OVERWRITE INTO TABLE primitives PARTITION(year=2009, month=3); - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/types/primitives/090401.txt" -OVERWRITE INTO TABLE primitives PARTITION(year=2009, month=4); -- -- Function qtest_get_java_boolean @@ -278,78 +58,5 @@ STORED AS INPUTFORMAT 'org.apache.hadoop.mapred.SequenceFileInputFormat' OUTPUTFORMAT 'org.apache.hadoop.mapred.SequenceFileOutputFormat'; - --- --- CBO tables --- - -drop table if exists cbo_t1; -drop table if exists cbo_t2; -drop table if exists cbo_t3; -drop table if exists src_cbo; -drop table if exists part; -drop table if exists lineitem; - -set hive.cbo.enable=true; - -create table cbo_t1(key string, value string, c_int int, c_float float, c_boolean boolean) partitioned by (dt string) row format delimited fields terminated by ',' STORED AS TEXTFILE; -create table cbo_t2(key string, value string, c_int int, c_float float, c_boolean boolean) partitioned by (dt string) row format delimited fields terminated by ',' STORED AS TEXTFILE; -create table cbo_t3(key string, value string, c_int int, c_float float, c_boolean boolean) row format delimited fields terminated by ',' STORED AS TEXTFILE; - -load data local inpath '${hiveconf:test.data.dir}/cbo_t1.txt' into table cbo_t1 partition (dt='2014'); -load data local inpath '${hiveconf:test.data.dir}/cbo_t2.txt' into table cbo_t2 partition (dt='2014'); -load data local inpath '${hiveconf:test.data.dir}/cbo_t3.txt' into table cbo_t3; - -CREATE TABLE part( - p_partkey INT, - p_name STRING, - p_mfgr STRING, - p_brand STRING, - p_type STRING, - p_size INT, - p_container STRING, - p_retailprice DOUBLE, - p_comment STRING -); - -LOAD DATA LOCAL INPATH '${hiveconf:test.data.dir}/part_tiny.txt' overwrite into table part; - -CREATE TABLE lineitem (L_ORDERKEY INT, - L_PARTKEY INT, - L_SUPPKEY INT, - L_LINENUMBER INT, - L_QUANTITY DOUBLE, - L_EXTENDEDPRICE DOUBLE, - L_DISCOUNT DOUBLE, - L_TAX DOUBLE, - L_RETURNFLAG STRING, - L_LINESTATUS STRING, - l_shipdate STRING, - L_COMMITDATE STRING, - L_RECEIPTDATE STRING, - L_SHIPINSTRUCT STRING, - L_SHIPMODE STRING, - L_COMMENT STRING) -ROW FORMAT DELIMITED -FIELDS TERMINATED BY '|'; - -LOAD DATA LOCAL INPATH '${hiveconf:test.data.dir}/lineitem.txt' OVERWRITE INTO TABLE lineitem; - -create table src_cbo as select * from src; - - -analyze table cbo_t1 partition (dt) compute statistics; -analyze table cbo_t1 compute statistics for columns key, value, c_int, c_float, c_boolean; -analyze table cbo_t2 partition (dt) compute statistics; -analyze table cbo_t2 compute statistics for columns key, value, c_int, c_float, c_boolean; -analyze table cbo_t3 compute statistics; -analyze table cbo_t3 compute statistics for columns key, value, c_int, c_float, c_boolean; -analyze table src_cbo compute statistics; -analyze table src_cbo compute statistics for columns; -analyze table part compute statistics; -analyze table part compute statistics for columns; -analyze table lineitem compute statistics; -analyze table lineitem compute statistics for columns; - reset; set hive.stats.dbclass=fs; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/scripts/q_test_init_compare.sql ---------------------------------------------------------------------- diff --git a/data/scripts/q_test_init_compare.sql b/data/scripts/q_test_init_compare.sql index c554250..e69de29 100644 --- a/data/scripts/q_test_init_compare.sql +++ b/data/scripts/q_test_init_compare.sql @@ -1,26 +0,0 @@ -set hive.stats.dbclass=fs; --- --- Table alltypesorc --- -DROP TABLE IF EXISTS alltypesorc; -CREATE TABLE alltypesorc( - ctinyint TINYINT, - csmallint SMALLINT, - cint INT, - cbigint BIGINT, - cfloat FLOAT, - cdouble DOUBLE, - cstring1 STRING, - cstring2 STRING, - ctimestamp1 TIMESTAMP, - ctimestamp2 TIMESTAMP, - cboolean1 BOOLEAN, - cboolean2 BOOLEAN) - STORED AS ORC; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/alltypesorc" -OVERWRITE INTO TABLE alltypesorc; - -ANALYZE TABLE alltypesorc COMPUTE STATISTICS; - -ANALYZE TABLE alltypesorc COMPUTE STATISTICS FOR COLUMNS ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/scripts/q_test_init_contrib.sql ---------------------------------------------------------------------- diff --git a/data/scripts/q_test_init_contrib.sql b/data/scripts/q_test_init_contrib.sql index b7bb5ab..e69de29 100644 --- a/data/scripts/q_test_init_contrib.sql +++ b/data/scripts/q_test_init_contrib.sql @@ -1,29 +0,0 @@ -set hive.stats.dbclass=fs; --- --- Table src --- -DROP TABLE IF EXISTS src; - -CREATE TABLE src (key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" INTO TABLE src; - -ANALYZE TABLE src COMPUTE STATISTICS; - -ANALYZE TABLE src COMPUTE STATISTICS FOR COLUMNS key,value; - --- --- Table src_thrift --- -DROP TABLE IF EXISTS src_thrift; - -CREATE TABLE src_thrift -ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer' -WITH SERDEPROPERTIES ( - 'serialization.class' = 'org.apache.hadoop.hive.serde2.thrift.test.Complex', - 'serialization.format' = 'org.apache.thrift.protocol.TBinaryProtocol') -STORED AS SEQUENCEFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/complex.seq" INTO TABLE src_thrift; - -ANALYZE TABLE src_thrift COMPUTE STATISTICS; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/scripts/q_test_init_for_minimr.sql ---------------------------------------------------------------------- diff --git a/data/scripts/q_test_init_for_minimr.sql b/data/scripts/q_test_init_for_minimr.sql index ab2e1d7..e69de29 100644 --- a/data/scripts/q_test_init_for_minimr.sql +++ b/data/scripts/q_test_init_for_minimr.sql @@ -1,48 +0,0 @@ --- --- Table src --- -DROP TABLE IF EXISTS src; - -CREATE TABLE src (key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" INTO TABLE src; - --- --- Table srcpart --- -DROP TABLE IF EXISTS srcpart; - -CREATE TABLE srcpart (key STRING COMMENT 'default', value STRING COMMENT 'default') -PARTITIONED BY (ds STRING, hr STRING) -STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" -OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-08", hr="11"); - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" -OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-08", hr="12"); - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" -OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-09", hr="11"); - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" -OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-09", hr="12"); - --- --- Table part --- -drop table if exists part; - -CREATE TABLE part( - p_partkey INT, - p_name STRING, - p_mfgr STRING, - p_brand STRING, - p_type STRING, - p_size INT, - p_container STRING, - p_retailprice DOUBLE, - p_comment STRING -); - -LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/scripts/q_test_init_src.sql ---------------------------------------------------------------------- diff --git a/data/scripts/q_test_init_src.sql b/data/scripts/q_test_init_src.sql index 765d6c7..e69de29 100644 --- a/data/scripts/q_test_init_src.sql +++ b/data/scripts/q_test_init_src.sql @@ -1,9 +0,0 @@ -DROP TABLE IF EXISTS src PURGE; - -CREATE TABLE src(key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" OVERWRITE INTO TABLE src; - -ANALYZE TABLE src COMPUTE STATISTICS; - -ANALYZE TABLE src COMPUTE STATISTICS FOR COLUMNS key,value; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/scripts/q_test_init_src_with_stats.sql ---------------------------------------------------------------------- diff --git a/data/scripts/q_test_init_src_with_stats.sql b/data/scripts/q_test_init_src_with_stats.sql index 765d6c7..e69de29 100644 --- a/data/scripts/q_test_init_src_with_stats.sql +++ b/data/scripts/q_test_init_src_with_stats.sql @@ -1,9 +0,0 @@ -DROP TABLE IF EXISTS src PURGE; - -CREATE TABLE src(key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" OVERWRITE INTO TABLE src; - -ANALYZE TABLE src COMPUTE STATISTICS; - -ANALYZE TABLE src COMPUTE STATISTICS FOR COLUMNS key,value; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/data/scripts/q_test_init_tez.sql ---------------------------------------------------------------------- diff --git a/data/scripts/q_test_init_tez.sql b/data/scripts/q_test_init_tez.sql index 4e6176b..e69de29 100644 --- a/data/scripts/q_test_init_tez.sql +++ b/data/scripts/q_test_init_tez.sql @@ -1,78 +0,0 @@ -set hive.stats.dbclass=fs; - --- --- Table src --- -DROP TABLE IF EXISTS src; - -CREATE TABLE src(key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" OVERWRITE INTO TABLE src; - -ANALYZE TABLE src COMPUTE STATISTICS; - -ANALYZE TABLE src COMPUTE STATISTICS FOR COLUMNS key,value; - --- --- Table src1 --- -DROP TABLE IF EXISTS src1; - -CREATE TABLE src1 (key STRING COMMENT 'default', value STRING COMMENT 'default') STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv3.txt" INTO TABLE src1; - -ANALYZE TABLE src1 COMPUTE STATISTICS; - -ANALYZE TABLE src1 COMPUTE STATISTICS FOR COLUMNS key,value; - --- --- Table srcpart --- -DROP TABLE IF EXISTS srcpart; - -CREATE TABLE srcpart (key STRING COMMENT 'default', value STRING COMMENT 'default') -PARTITIONED BY (ds STRING, hr STRING) -STORED AS TEXTFILE; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" -OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-08", hr="11"); - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" -OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-08", hr="12"); - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" -OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-09", hr="11"); - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/kv1.txt" -OVERWRITE INTO TABLE srcpart PARTITION (ds="2008-04-09", hr="12"); - -ANALYZE TABLE srcpart PARTITION(ds, hr) COMPUTE STATISTICS; - -ANALYZE TABLE srcpart PARTITION(ds, hr) COMPUTE STATISTICS FOR COLUMNS key,value; - --- --- Table alltypesorc --- -DROP TABLE IF EXISTS alltypesorc; -CREATE TABLE alltypesorc( - ctinyint TINYINT, - csmallint SMALLINT, - cint INT, - cbigint BIGINT, - cfloat FLOAT, - cdouble DOUBLE, - cstring1 STRING, - cstring2 STRING, - ctimestamp1 TIMESTAMP, - ctimestamp2 TIMESTAMP, - cboolean1 BOOLEAN, - cboolean2 BOOLEAN) - STORED AS ORC; - -LOAD DATA LOCAL INPATH "${hiveconf:test.data.dir}/alltypesorc" -OVERWRITE INTO TABLE alltypesorc; - -ANALYZE TABLE alltypesorc COMPUTE STATISTICS; - -ANALYZE TABLE alltypesorc COMPUTE STATISTICS FOR COLUMNS ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/negative/generatehfiles_require_family_path.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/negative/generatehfiles_require_family_path.q b/hbase-handler/src/test/queries/negative/generatehfiles_require_family_path.q index 6844fbc..e6ff587 100644 --- a/hbase-handler/src/test/queries/negative/generatehfiles_require_family_path.q +++ b/hbase-handler/src/test/queries/negative/generatehfiles_require_family_path.q @@ -1,3 +1,4 @@ +--! qt:dataset:src -- -*- mode:sql -*- DROP TABLE IF EXISTS hbase_bulk; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/external_table_ppd.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/external_table_ppd.q b/hbase-handler/src/test/queries/positive/external_table_ppd.q index fbef4bb..c6f80cb 100644 --- a/hbase-handler/src/test/queries/positive/external_table_ppd.q +++ b/hbase-handler/src/test/queries/positive/external_table_ppd.q @@ -1,3 +1,4 @@ +--! qt:dataset:src DROP TABLE t_hbase; CREATE TABLE t_hbase(key STRING, http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_binary_map_queries.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_binary_map_queries.q b/hbase-handler/src/test/queries/positive/hbase_binary_map_queries.q index 255a2c7..7d9e16e 100644 --- a/hbase-handler/src/test/queries/positive/hbase_binary_map_queries.q +++ b/hbase-handler/src/test/queries/positive/hbase_binary_map_queries.q @@ -1,3 +1,4 @@ +--! qt:dataset:src DROP TABLE hbase_src; CREATE TABLE hbase_src(key STRING, http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_binary_map_queries_prefix.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_binary_map_queries_prefix.q b/hbase-handler/src/test/queries/positive/hbase_binary_map_queries_prefix.q index 9ff4366..c205210 100644 --- a/hbase-handler/src/test/queries/positive/hbase_binary_map_queries_prefix.q +++ b/hbase-handler/src/test/queries/positive/hbase_binary_map_queries_prefix.q @@ -1,3 +1,4 @@ +--! qt:dataset:src DROP TABLE hbase_src; CREATE TABLE hbase_src(key STRING, http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_binary_storage_queries.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_binary_storage_queries.q b/hbase-handler/src/test/queries/positive/hbase_binary_storage_queries.q index b048871..86ead91 100644 --- a/hbase-handler/src/test/queries/positive/hbase_binary_storage_queries.q +++ b/hbase-handler/src/test/queries/positive/hbase_binary_storage_queries.q @@ -1,3 +1,4 @@ +--! qt:dataset:src DROP TABLE t_hbase; CREATE TABLE t_hbase(key STRING, http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_bulk.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_bulk.q b/hbase-handler/src/test/queries/positive/hbase_bulk.q index 475aafc..d2e62dc 100644 --- a/hbase-handler/src/test/queries/positive/hbase_bulk.q +++ b/hbase-handler/src/test/queries/positive/hbase_bulk.q @@ -1,3 +1,4 @@ +--! qt:dataset:src drop table hbsort; drop table hbpartition; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_custom_key.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_custom_key.q b/hbase-handler/src/test/queries/positive/hbase_custom_key.q index 9dbb2a0..87fbf4a 100644 --- a/hbase-handler/src/test/queries/positive/hbase_custom_key.q +++ b/hbase-handler/src/test/queries/positive/hbase_custom_key.q @@ -1,3 +1,4 @@ +--! qt:dataset:src CREATE TABLE hbase_ck_1(key struct<col1:string,col2:string,col3:string>, value string) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES ( http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_custom_key2.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_custom_key2.q b/hbase-handler/src/test/queries/positive/hbase_custom_key2.q index 9fba4f6..5116475 100644 --- a/hbase-handler/src/test/queries/positive/hbase_custom_key2.q +++ b/hbase-handler/src/test/queries/positive/hbase_custom_key2.q @@ -1,3 +1,4 @@ +--! qt:dataset:src CREATE TABLE hbase_ck_4(key struct<col1:string,col2:string,col3:string>, value string) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES ( http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_custom_key3.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_custom_key3.q b/hbase-handler/src/test/queries/positive/hbase_custom_key3.q index 22d2c9e..488a32f 100644 --- a/hbase-handler/src/test/queries/positive/hbase_custom_key3.q +++ b/hbase-handler/src/test/queries/positive/hbase_custom_key3.q @@ -1,3 +1,4 @@ +--! qt:dataset:src CREATE TABLE hbase_ck_5(key struct<col1:string,col2:string,col3:string>, value string) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES ( http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_ddl.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_ddl.q b/hbase-handler/src/test/queries/positive/hbase_ddl.q index a8bae75..dcec8d6 100644 --- a/hbase-handler/src/test/queries/positive/hbase_ddl.q +++ b/hbase-handler/src/test/queries/positive/hbase_ddl.q @@ -1,3 +1,4 @@ +--! qt:dataset:src DROP TABLE hbase_table_1; CREATE TABLE hbase_table_1(key int comment 'It is a column key', value string comment 'It is the column string value') STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_handler_bulk.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_handler_bulk.q b/hbase-handler/src/test/queries/positive/hbase_handler_bulk.q index d02a61f..1eadfce 100644 --- a/hbase-handler/src/test/queries/positive/hbase_handler_bulk.q +++ b/hbase-handler/src/test/queries/positive/hbase_handler_bulk.q @@ -1,3 +1,4 @@ +--! qt:dataset:src -- -*- mode:sql -*- drop table if exists hb_target; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_handler_snapshot.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_handler_snapshot.q b/hbase-handler/src/test/queries/positive/hbase_handler_snapshot.q index dd4fe0f..e429071 100644 --- a/hbase-handler/src/test/queries/positive/hbase_handler_snapshot.q +++ b/hbase-handler/src/test/queries/positive/hbase_handler_snapshot.q @@ -1,3 +1,4 @@ +--! qt:dataset:src_hbase set hive.stats.column.autogather=true; SET hive.hbase.snapshot.name=src_hbase_snapshot; SET hive.hbase.snapshot.restoredir=/tmp; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_joins.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_joins.q b/hbase-handler/src/test/queries/positive/hbase_joins.q index 1616adc..7be9f6a 100644 --- a/hbase-handler/src/test/queries/positive/hbase_joins.q +++ b/hbase-handler/src/test/queries/positive/hbase_joins.q @@ -1,3 +1,4 @@ +--! qt:dataset:src DROP TABLE users; DROP TABLE states; DROP TABLE countries; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_ppd_join.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_ppd_join.q b/hbase-handler/src/test/queries/positive/hbase_ppd_join.q index 2436c19..b7e2a3b 100644 --- a/hbase-handler/src/test/queries/positive/hbase_ppd_join.q +++ b/hbase-handler/src/test/queries/positive/hbase_ppd_join.q @@ -1,3 +1,4 @@ +--! qt:dataset:src --create hive hbase table 1 drop table if exists hive1_tbl_data_hbase1; drop table if exists hive1_tbl_data_hbase2; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_ppd_key_range.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_ppd_key_range.q b/hbase-handler/src/test/queries/positive/hbase_ppd_key_range.q index 59e724d..cf1d1d4 100644 --- a/hbase-handler/src/test/queries/positive/hbase_ppd_key_range.q +++ b/hbase-handler/src/test/queries/positive/hbase_ppd_key_range.q @@ -1,3 +1,5 @@ +--! qt:dataset:src +--! qt:dataset:part CREATE TABLE hbase_pushdown(key string, value string) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf:string"); http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_pushdown.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_pushdown.q b/hbase-handler/src/test/queries/positive/hbase_pushdown.q index 0d29c82..8e366af 100644 --- a/hbase-handler/src/test/queries/positive/hbase_pushdown.q +++ b/hbase-handler/src/test/queries/positive/hbase_pushdown.q @@ -1,3 +1,5 @@ +--! qt:dataset:src +--! qt:dataset:part CREATE TABLE hbase_pushdown(key int, value string) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf:string"); http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_queries.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_queries.q b/hbase-handler/src/test/queries/positive/hbase_queries.q index 4604f3e..a4ea0c5 100644 --- a/hbase-handler/src/test/queries/positive/hbase_queries.q +++ b/hbase-handler/src/test/queries/positive/hbase_queries.q @@ -1,3 +1,4 @@ +--! qt:dataset:src DROP TABLE hbase_table_1; CREATE TABLE hbase_table_1(key int comment 'It is a column key', value string comment 'It is the column string value') STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_scan_params.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_scan_params.q b/hbase-handler/src/test/queries/positive/hbase_scan_params.q index 14d7d15..522960c 100644 --- a/hbase-handler/src/test/queries/positive/hbase_scan_params.q +++ b/hbase-handler/src/test/queries/positive/hbase_scan_params.q @@ -1,3 +1,4 @@ +--! qt:dataset:src CREATE TABLE hbase_pushdown(key int, value string) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf:string", http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_single_sourced_multi_insert.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_single_sourced_multi_insert.q b/hbase-handler/src/test/queries/positive/hbase_single_sourced_multi_insert.q index 96fec0e..bd4672b 100644 --- a/hbase-handler/src/test/queries/positive/hbase_single_sourced_multi_insert.q +++ b/hbase-handler/src/test/queries/positive/hbase_single_sourced_multi_insert.q @@ -1,3 +1,4 @@ +--! qt:dataset:src -- HIVE-4375 Single sourced multi insert consists of native and non-native table mixed throws NPE CREATE TABLE src_x1(key string, value string); CREATE TABLE src_x2(key string, value string) http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_timestamp.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_timestamp.q b/hbase-handler/src/test/queries/positive/hbase_timestamp.q index 6ae2c30..46d7529 100644 --- a/hbase-handler/src/test/queries/positive/hbase_timestamp.q +++ b/hbase-handler/src/test/queries/positive/hbase_timestamp.q @@ -1,3 +1,4 @@ +--! qt:dataset:src DROP TABLE hbase_table; CREATE TABLE hbase_table (key string, value string, `time` timestamp) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbase_timestamp_format.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbase_timestamp_format.q b/hbase-handler/src/test/queries/positive/hbase_timestamp_format.q index a8d5501..75820e1 100644 --- a/hbase-handler/src/test/queries/positive/hbase_timestamp_format.q +++ b/hbase-handler/src/test/queries/positive/hbase_timestamp_format.q @@ -1,3 +1,4 @@ +--! qt:dataset:src create table hbase_str(rowkey string,mytime string,mystr string) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/hbasestats.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/hbasestats.q b/hbase-handler/src/test/queries/positive/hbasestats.q index 52e11c9..0185002 100644 --- a/hbase-handler/src/test/queries/positive/hbasestats.q +++ b/hbase-handler/src/test/queries/positive/hbasestats.q @@ -1,3 +1,4 @@ +--! qt:dataset:src DROP TABLE users; CREATE TABLE users(key string, state string, country string, country_id int) http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/hbase-handler/src/test/queries/positive/ppd_key_ranges.q ---------------------------------------------------------------------- diff --git a/hbase-handler/src/test/queries/positive/ppd_key_ranges.q b/hbase-handler/src/test/queries/positive/ppd_key_ranges.q index 0497d25..8b70c04 100644 --- a/hbase-handler/src/test/queries/positive/ppd_key_ranges.q +++ b/hbase-handler/src/test/queries/positive/ppd_key_ranges.q @@ -1,3 +1,4 @@ +--! qt:dataset:src CREATE TABLE hbase_ppd_keyrange(key int, value string) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key#binary,cf:string"); http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreAccumuloCliDriver.java ---------------------------------------------------------------------- diff --git a/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreAccumuloCliDriver.java b/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreAccumuloCliDriver.java index d21fcaf..9c9ba18 100644 --- a/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreAccumuloCliDriver.java +++ b/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreAccumuloCliDriver.java @@ -43,16 +43,7 @@ public class CoreAccumuloCliDriver extends CliAdapter { @BeforeClass public void beforeClass() { setup = new AccumuloTestSetup(); - } - @Override - @AfterClass - public void shutdown() throws Exception { - setup.tearDown(); - } - @Override - @Before - public void setUp() { - + MiniClusterType miniMR = cliConfig.getClusterType(); String initScript = cliConfig.getInitScript(); String cleanupScript = cliConfig.getCleanupScript(); @@ -60,14 +51,20 @@ public class CoreAccumuloCliDriver extends CliAdapter { try { qt = new AccumuloQTestUtil(cliConfig.getResultsDir(), cliConfig.getLogDir(), miniMR, setup, initScript, cleanupScript); + + // do a one time initialization + qt.cleanUp(); + qt.createSources(); } catch (Exception e) { throw new RuntimeException("Unexpected exception in setUp",e); } } - + @Override - @After - public void tearDown() { + @AfterClass + public void shutdown() throws Exception { + setup.tearDown(); + try { qt.shutdown(); } @@ -75,6 +72,15 @@ public class CoreAccumuloCliDriver extends CliAdapter { throw new RuntimeException("Unexpected exception in tearDown",e); } } + @Override + @Before + public void setUp() { + } + + @Override + @After + public void tearDown() { + } @Override public void runTest(String tname, String fname, String fpath) throws Exception { @@ -89,7 +95,7 @@ public class CoreAccumuloCliDriver extends CliAdapter { return; } - qt.cliInit(new File(fpath)); + qt.cliInit(new File(fpath), false); qt.clearTestSideEffects(); int ecode = qt.executeClient(fname); if (ecode != 0) { http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreBeeLineDriver.java ---------------------------------------------------------------------- diff --git a/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreBeeLineDriver.java b/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreBeeLineDriver.java index 1ec1e5e..99748bd 100644 --- a/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreBeeLineDriver.java +++ b/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreBeeLineDriver.java @@ -19,11 +19,27 @@ package org.apache.hadoop.hive.cli.control; import static org.junit.Assert.fail; -import com.google.common.base.Strings; +import java.io.File; +import java.io.IOException; +import java.sql.SQLException; +import java.util.ArrayList; +import java.util.Arrays; +import java.util.HashMap; +import java.util.HashSet; +import java.util.List; +import java.util.Map; +import java.util.Set; +import java.util.concurrent.Callable; +import java.util.stream.Stream; + +import org.apache.commons.io.FileUtils; import org.apache.commons.lang.StringUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConfUtil; import org.apache.hadoop.hive.ql.QTestProcessExecResult; +import org.apache.hadoop.hive.ql.dataset.Dataset; +import org.apache.hadoop.hive.ql.dataset.DatasetCollection; +import org.apache.hadoop.hive.ql.dataset.DatasetParser; import org.apache.hadoop.hive.ql.hooks.PreExecutePrinter; import org.apache.hive.beeline.ConvertedOutputFile.Converter; import org.apache.hive.beeline.QFile; @@ -34,10 +50,8 @@ import org.apache.hive.jdbc.miniHS2.MiniHS2; import org.junit.AfterClass; import org.junit.BeforeClass; -import java.io.File; -import java.io.IOException; -import java.sql.SQLException; -import java.util.HashMap; +import com.google.common.base.Strings; +import com.google.common.collect.ObjectArrays; public class CoreBeeLineDriver extends CliAdapter { private final File hiveRootDirectory = new File(AbstractCliConfig.HIVE_ROOT); @@ -48,11 +62,13 @@ public class CoreBeeLineDriver extends CliAdapter { private final File cleanupScript; private final File testDataDirectory; private final File testScriptDirectory; + private final File datasetDirectory; private boolean overwrite = false; private boolean useSharedDatabase = false; private MiniHS2 miniHS2; private QFileClientBuilder clientBuilder; private QFileBuilder fileBuilder; + private final Map<String, Set<String>> datasets = new HashMap<String, Set<String>>(); public CoreBeeLineDriver(AbstractCliConfig testCliConfig) { super(testCliConfig); @@ -71,6 +87,7 @@ public class CoreBeeLineDriver extends CliAdapter { testDataDirectory = new File(testDataDirectoryName); } testScriptDirectory = new File(hiveRootDirectory, "data" + File.separator + "scripts"); + datasetDirectory = new File(testDataDirectory, "datasets"); String initScriptFileName = System.getProperty("test.init.script"); if (initScriptFileName != null) { initScript = new File(testScriptDirectory, initScriptFileName); @@ -156,6 +173,23 @@ public class CoreBeeLineDriver extends CliAdapter { + "\nCheck the following logs for details:\n - " + beeLineOutput + "\n - " + log, e); } } + + protected void runInfraScript(String[] commands, File beeLineOutput, File log) + throws IOException, SQLException { + try (QFileBeeLineClient beeLineClient = clientBuilder.getClient(beeLineOutput)) { + String[] preCommands = + new String[] { "set hive.exec.pre.hooks=" + PreExecutePrinter.class.getName() + ";", + "set test.data.dir=" + testDataDirectory + ";", + "set test.script.dir=" + testScriptDirectory + ";" }; + + String[] allCommands = + Stream.concat(Arrays.stream(preCommands), Arrays.stream(commands)).toArray(String[]::new); + beeLineClient.execute(allCommands, log, Converter.NONE); + } catch (Exception e) { + throw new SQLException("Error running infra commands, " + + "\nCheck the following logs for details:\n - " + beeLineOutput + "\n - " + log, e); + } + } @Override @AfterClass @@ -167,12 +201,12 @@ public class CoreBeeLineDriver extends CliAdapter { } } - public void runTest(QFile qFile) throws Exception { + private void runTest(QFile qFile, List<Callable<Void>> preCommands) throws Exception { try (QFileBeeLineClient beeLineClient = clientBuilder.getClient(qFile.getLogFile())) { long startTime = System.currentTimeMillis(); System.err.println(">>> STARTED " + qFile.getName()); - beeLineClient.execute(qFile); + beeLineClient.execute(qFile, preCommands); long queryEndTime = System.currentTimeMillis(); System.err.println(">>> EXECUTED " + qFile.getName() + ": " + (queryEndTime - startTime) @@ -207,6 +241,11 @@ public class CoreBeeLineDriver extends CliAdapter { throw new Exception("Exception running or analyzing the results of the query file: " + qFile + "\n" + qFile.getDebugHint(), e); } + + } + + public void runTest(QFile qFile) throws Exception { + runTest(qFile, null); } @Override @@ -220,6 +259,58 @@ public class CoreBeeLineDriver extends CliAdapter { @Override public void runTest(String name, String name2, String absolutePath) throws Exception { QFile qFile = fileBuilder.getQFile(name); - runTest(qFile); + List<Callable<Void>> commands = initDataSetForTest(qFile); + runTest(qFile, commands); + } + + private List<Callable<Void>> initDataSetForTest(QFile qFile) throws Exception { + DatasetParser parser = new DatasetParser(); + parser.parse(qFile.getInputFile()); + + List<Callable<Void>> commands = new ArrayList<>(); + + DatasetCollection datasets = parser.getDatasets(); + for (String table : datasets.getTables()) { + Callable<Void> command = initDataset(table, qFile); + if (command != null) { + commands.add(command); + } + } + + return commands; + } + + protected Callable<Void> initDataset(String table, QFile qFile) throws Exception { + if (datasetInitialized(table, qFile)) { + return null; + } + + Callable<Void> command = new Callable<Void>() { + @Override + public Void call() throws Exception { + File tableFile = new File(new File(datasetDirectory, table), Dataset.INIT_FILE_NAME); + List<String> datasetLines = FileUtils.readLines(tableFile); + String[] datasetCommands = datasetLines.toArray(new String[datasetLines.size()]); + + runInfraScript( + ObjectArrays.concat(String.format("use %s;", qFile.getDatabaseName()), datasetCommands), + new File(logDirectory, "dataset.beeline"), new File(logDirectory, "dataset.raw")); + + return null; + } + }; + + datasets.get(qFile.getDatabaseName()).add(table); + + return command; + } + + private boolean datasetInitialized(String table, QFile qFile) { + if (datasets.get(qFile.getDatabaseName()) == null) { + datasets.put(qFile.getDatabaseName(), new HashSet<String>()); + return false; + } + + return datasets.get(qFile.getDatabaseName()).contains(table); } } http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreCompareCliDriver.java ---------------------------------------------------------------------- diff --git a/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreCompareCliDriver.java b/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreCompareCliDriver.java index fbca424..c36d231 100644 --- a/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreCompareCliDriver.java +++ b/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreCompareCliDriver.java @@ -133,6 +133,9 @@ public class CoreCompareCliDriver extends CliAdapter{ } int ecode = 0; + + qt.cliInit(new File(fpath), false); + List<String> outputs = new ArrayList<>(versionFiles.size()); for (String versionFile : versionFiles) { // 1 for "_" after tname; 3 for ".qv" at the end. Version is in between. http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreHBaseCliDriver.java ---------------------------------------------------------------------- diff --git a/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreHBaseCliDriver.java b/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreHBaseCliDriver.java index 24bdc58..b40b8d7 100644 --- a/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreHBaseCliDriver.java +++ b/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/CoreHBaseCliDriver.java @@ -17,7 +17,6 @@ */ package org.apache.hadoop.hive.cli.control; -import static org.apache.hadoop.hive.cli.control.AbstractCliConfig.HIVE_ROOT; import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; http://git-wip-us.apache.org/repos/asf/hive/blob/86b678f5/itests/util/src/main/java/org/apache/hadoop/hive/hbase/HBaseQTestUtil.java ---------------------------------------------------------------------- diff --git a/itests/util/src/main/java/org/apache/hadoop/hive/hbase/HBaseQTestUtil.java b/itests/util/src/main/java/org/apache/hadoop/hive/hbase/HBaseQTestUtil.java index 1873bfe..07df0c9 100644 --- a/itests/util/src/main/java/org/apache/hadoop/hive/hbase/HBaseQTestUtil.java +++ b/itests/util/src/main/java/org/apache/hadoop/hive/hbase/HBaseQTestUtil.java @@ -68,14 +68,7 @@ public class HBaseQTestUtil extends QTestUtil { conf.setBoolean("hive.test.init.phase", true); - // create and load the input data into the hbase table - runCreateTableCmd( - "CREATE TABLE " + HBASE_SRC_NAME + "(key INT, value STRING)" - + " STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'" - + " WITH SERDEPROPERTIES ('hbase.columns.mapping' = ':key,cf:val')" - + " TBLPROPERTIES ('hbase.table.name' = '" + HBASE_SRC_NAME + "')" - ); - runCmd("INSERT OVERWRITE TABLE " + HBASE_SRC_NAME + " SELECT * FROM src"); + initDataset(HBASE_SRC_NAME); // create a snapshot Admin admin = null; @@ -93,9 +86,6 @@ public class HBaseQTestUtil extends QTestUtil { public void cleanUp(String tname) throws Exception { super.cleanUp(tname); - // drop in case leftover from unsuccessful run - db.dropTable(Warehouse.DEFAULT_DATABASE_NAME, HBASE_SRC_NAME); - Admin admin = null; try { admin = conn.getAdmin();