HIVE-13549: Remove jdk version specific out files from Hive2 (Mohit Sabharwal, reviewed by Sergio Pena)
Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/9687dcc9 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/9687dcc9 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/9687dcc9 Branch: refs/heads/java8 Commit: 9687dcc94b2ced7755f5173b56113af906c43466 Parents: 8694435 Author: Mohit Sabharwal <mohitsabhar...@cloudera.com> Authored: Fri May 20 11:14:13 2016 -0500 Committer: Sergio Pena <sergio.p...@cloudera.com> Committed: Thu May 26 10:39:18 2016 -0500 ---------------------------------------------------------------------- .../columnstats_partlvl_invalid_values.q | 1 - .../clientpositive/authorization_explain.q | 1 - ql/src/test/queries/clientpositive/avro_date.q | 1 - .../clientpositive/avro_deserialize_map_null.q | 1 - .../clientpositive/avro_nullable_fields.q | 1 - .../queries/clientpositive/avro_timestamp.q | 1 - .../clientpositive/cbo_rp_outer_join_ppr.q | 1 - ql/src/test/queries/clientpositive/char_udf1.q | 1 - ql/src/test/queries/clientpositive/input4.q | 1 - ql/src/test/queries/clientpositive/join0.q | 1 - .../queries/clientpositive/list_bucket_dml_10.q | 1 - .../queries/clientpositive/list_bucket_dml_11.q | 1 - .../queries/clientpositive/list_bucket_dml_12.q | 1 - .../queries/clientpositive/list_bucket_dml_13.q | 1 - .../queries/clientpositive/list_bucket_dml_2.q | 1 - .../queries/clientpositive/list_bucket_dml_4.q | 1 - .../queries/clientpositive/list_bucket_dml_5.q | 1 - .../queries/clientpositive/list_bucket_dml_6.q | 1 - .../queries/clientpositive/list_bucket_dml_8.q | 1 - .../queries/clientpositive/list_bucket_dml_9.q | 1 - .../queries/clientpositive/outer_join_ppr.q | 1 - .../queries/clientpositive/parquet_map_null.q | 1 - ql/src/test/queries/clientpositive/plan_json.q | 1 - .../queries/clientpositive/stats_list_bucket.q | 1 - ql/src/test/queries/clientpositive/str_to_map.q | 1 - .../clientpositive/subquery_multiinsert.q | 1 - .../clientpositive/subquery_notin_having.q | 1 - .../test/queries/clientpositive/varchar_udf1.q | 1 - .../clientpositive/vector_cast_constant.q | 1 - ...mnstats_partlvl_invalid_values.q.java1.7.out | 73 -- ...mnstats_partlvl_invalid_values.q.java1.8.out | 73 -- .../columnstats_partlvl_invalid_values.q.out | 69 ++ .../authorization_explain.q.java1.7.out | 44 - .../authorization_explain.q.java1.8.out | 47 - .../clientpositive/authorization_explain.q.out | 40 + .../clientpositive/avro_date.q.java1.7.out | 130 -- .../clientpositive/avro_date.q.java1.8.out | 130 -- .../test/results/clientpositive/avro_date.q.out | 126 ++ .../avro_deserialize_map_null.q.java1.7.out | 57 - .../avro_deserialize_map_null.q.java1.8.out | 57 - .../avro_deserialize_map_null.q.out | 55 + .../avro_nullable_fields.q.java1.7.out | 179 --- .../avro_nullable_fields.q.java1.8.out | 179 --- .../clientpositive/avro_nullable_fields.q.out | 177 +++ .../clientpositive/avro_timestamp.q.java1.7.out | 134 --- .../clientpositive/avro_timestamp.q.java1.8.out | 134 --- .../results/clientpositive/avro_timestamp.q.out | 132 +++ .../cbo_rp_outer_join_ppr.q.java1.7.out | 693 ----------- .../clientpositive/cbo_rp_outer_join_ppr.q.out | 691 +++++++++++ .../clientpositive/char_udf1.q.java1.7.out | 463 -------- .../clientpositive/char_udf1.q.java1.8.out | 457 ------- .../test/results/clientpositive/char_udf1.q.out | 459 +++++++ .../results/clientpositive/input4.q.java1.7.out | 559 --------- .../results/clientpositive/input4.q.java1.8.out | 559 --------- ql/src/test/results/clientpositive/input4.q.out | 555 +++++++++ .../results/clientpositive/join0.q.java1.7.out | 240 ---- .../results/clientpositive/join0.q.java1.8.out | 240 ---- ql/src/test/results/clientpositive/join0.q.out | 238 ++++ .../list_bucket_dml_10.q.java1.7.out | 361 ------ .../list_bucket_dml_10.q.java1.8.out | 389 ------ .../clientpositive/list_bucket_dml_10.q.out | 359 ++++++ .../list_bucket_dml_11.q.java1.7.out | 329 ----- .../list_bucket_dml_11.q.java1.8.out | 424 ------- .../clientpositive/list_bucket_dml_11.q.out | 327 +++++ .../list_bucket_dml_12.q.java1.7.out | 426 ------- .../list_bucket_dml_12.q.java1.8.out | 596 ---------- .../clientpositive/list_bucket_dml_12.q.out | 424 +++++++ .../list_bucket_dml_13.q.java1.7.out | 337 ------ .../list_bucket_dml_13.q.java1.8.out | 439 ------- .../clientpositive/list_bucket_dml_13.q.out | 335 ++++++ .../list_bucket_dml_2.q.java1.7.out | 591 --------- .../list_bucket_dml_2.q.java1.8.out | 692 ----------- .../clientpositive/list_bucket_dml_2.q.out | 589 +++++++++ .../list_bucket_dml_4.q.java1.7.out | 813 ------------- .../list_bucket_dml_4.q.java1.8.out | 915 -------------- .../clientpositive/list_bucket_dml_4.q.out | 811 +++++++++++++ .../list_bucket_dml_5.q.java1.7.out | 506 -------- .../list_bucket_dml_5.q.java1.8.out | 617 ---------- .../clientpositive/list_bucket_dml_5.q.out | 504 ++++++++ .../list_bucket_dml_6.q.java1.7.out | 1007 ---------------- .../list_bucket_dml_6.q.java1.8.out | 1119 ------------------ .../clientpositive/list_bucket_dml_6.q.out | 1005 ++++++++++++++++ .../list_bucket_dml_8.q.java1.7.out | 641 ---------- .../list_bucket_dml_8.q.java1.8.out | 712 ----------- .../clientpositive/list_bucket_dml_8.q.out | 639 ++++++++++ .../list_bucket_dml_9.q.java1.7.out | 813 ------------- .../list_bucket_dml_9.q.java1.8.out | 915 -------------- .../clientpositive/list_bucket_dml_9.q.out | 811 +++++++++++++ .../clientpositive/llap/join0.q.java1.7.out | 242 ---- .../clientpositive/llap/join0.q.java1.8.out | 242 ---- .../results/clientpositive/llap/join0.q.out | 243 ++++ .../llap/vector_cast_constant.q.java1.7.out | 217 ---- .../llap/vector_cast_constant.q.java1.8.out | 217 ---- .../llap/vector_cast_constant.q.out | 216 ++++ .../clientpositive/outer_join_ppr.q.java1.7.out | 685 ----------- .../clientpositive/outer_join_ppr.q.java1.8.out | 855 ------------- .../results/clientpositive/outer_join_ppr.q.out | 683 +++++++++++ .../parquet_map_null.q.java1.7.out | 70 -- .../parquet_map_null.q.java1.8.out | 70 -- .../clientpositive/parquet_map_null.q.out | 68 ++ .../clientpositive/plan_json.q.java1.7.out | 13 - .../clientpositive/plan_json.q.java1.8.out | 13 - .../test/results/clientpositive/plan_json.q.out | 11 + .../clientpositive/spark/join0.q.java1.7.out | 238 ---- .../clientpositive/spark/join0.q.java1.8.out | 238 ---- .../results/clientpositive/spark/join0.q.out | 20 +- .../spark/list_bucket_dml_10.q.java1.7.out | 252 ---- .../spark/list_bucket_dml_10.q.java1.8.out | 280 ----- .../spark/list_bucket_dml_10.q.out | 250 ++++ .../spark/list_bucket_dml_2.q.java1.7.out | 591 --------- .../spark/list_bucket_dml_2.q.java1.8.out | 663 ----------- .../spark/list_bucket_dml_2.q.out | Bin 28667 -> 27128 bytes .../spark/outer_join_ppr.q.java1.7.out | 709 ----------- .../spark/outer_join_ppr.q.java1.8.out | 879 -------------- .../clientpositive/spark/outer_join_ppr.q.out | 490 ++------ .../spark/subquery_multiinsert.q.java1.7.out | 886 -------------- .../spark/subquery_multiinsert.q.java1.8.out | 890 -------------- .../spark/subquery_multiinsert.q.out | 56 +- .../spark/vector_cast_constant.q.java1.7.out | 217 ---- .../spark/vector_cast_constant.q.java1.8.out | 203 ---- .../spark/vector_cast_constant.q.out | 54 +- .../stats_list_bucket.q.java1.7.out | 191 --- .../stats_list_bucket.q.java1.8.out | 193 --- .../clientpositive/stats_list_bucket.q.out | 189 +++ .../clientpositive/str_to_map.q.java1.7.out | 220 ---- .../clientpositive/str_to_map.q.java1.8.out | 219 ---- .../results/clientpositive/str_to_map.q.out | 216 ++++ .../subquery_multiinsert.q.java1.7.out | 999 ---------------- .../subquery_multiinsert.q.java1.8.out | 999 ---------------- .../clientpositive/subquery_multiinsert.q.out | 997 ++++++++++++++++ .../subquery_notin_having.q.java1.7.out | 766 ------------ .../subquery_notin_having.q.java1.8.out | 762 ------------ .../clientpositive/subquery_notin_having.q.out | 764 ++++++++++++ .../clientpositive/tez/join0.q.java1.7.out | 239 ---- .../clientpositive/tez/join0.q.java1.8.out | 236 ---- .../test/results/clientpositive/tez/join0.q.out | 237 ++++ .../tez/vector_cast_constant.q.java1.7.out | 218 ---- .../tez/vector_cast_constant.q.java1.8.out | 216 ---- .../tez/vector_cast_constant.q.out | 214 ++++ .../clientpositive/varchar_udf1.q.java1.7.out | 457 ------- .../clientpositive/varchar_udf1.q.java1.8.out | 457 ------- .../results/clientpositive/varchar_udf1.q.out | 453 +++++++ .../vector_cast_constant.q.java1.7.out | 220 ---- .../vector_cast_constant.q.java1.8.out | 197 --- .../clientpositive/vector_cast_constant.q.out | 53 +- 145 files changed, 13120 insertions(+), 32818 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientnegative/columnstats_partlvl_invalid_values.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientnegative/columnstats_partlvl_invalid_values.q b/ql/src/test/queries/clientnegative/columnstats_partlvl_invalid_values.q index 712ece7..8521631 100644 --- a/ql/src/test/queries/clientnegative/columnstats_partlvl_invalid_values.q +++ b/ql/src/test/queries/clientnegative/columnstats_partlvl_invalid_values.q @@ -1,4 +1,3 @@ --- JAVA_VERSION_SPECIFIC_OUTPUT DROP TABLE Employee_Part; http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/authorization_explain.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/authorization_explain.q b/ql/src/test/queries/clientpositive/authorization_explain.q index 6a9475c..d429704 100644 --- a/ql/src/test/queries/clientpositive/authorization_explain.q +++ b/ql/src/test/queries/clientpositive/authorization_explain.q @@ -2,7 +2,6 @@ set hive.security.authorization.manager=org.apache.hadoop.hive.ql.security.autho set hive.mapred.mode=nonstrict; set hive.security.authorization.enabled=true; --- JAVA_VERSION_SPECIFIC_OUTPUT explain authorization select * from src join srcpart; explain formatted authorization select * from src join srcpart; http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/avro_date.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/avro_date.q b/ql/src/test/queries/clientpositive/avro_date.q index 15c07de..7169822 100644 --- a/ql/src/test/queries/clientpositive/avro_date.q +++ b/ql/src/test/queries/clientpositive/avro_date.q @@ -1,5 +1,4 @@ set hive.mapred.mode=nonstrict; --- JAVA_VERSION_SPECIFIC_OUTPUT DROP TABLE avro_date_staging; DROP TABLE avro_date; http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/avro_deserialize_map_null.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/avro_deserialize_map_null.q b/ql/src/test/queries/clientpositive/avro_deserialize_map_null.q index 962e649..42258d9 100644 --- a/ql/src/test/queries/clientpositive/avro_deserialize_map_null.q +++ b/ql/src/test/queries/clientpositive/avro_deserialize_map_null.q @@ -4,7 +4,6 @@ -- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"] -- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}] --- JAVA_VERSION_SPECIFIC_OUTPUT DROP TABLE IF EXISTS avro_table; http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/avro_nullable_fields.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/avro_nullable_fields.q b/ql/src/test/queries/clientpositive/avro_nullable_fields.q index 9ba7441..cb398d6 100644 --- a/ql/src/test/queries/clientpositive/avro_nullable_fields.q +++ b/ql/src/test/queries/clientpositive/avro_nullable_fields.q @@ -1,6 +1,5 @@ -- Verify that nullable fields properly work --- JAVA_VERSION_SPECIFIC_OUTPUT CREATE TABLE test_serializer(string1 STRING, int1 INT, http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/avro_timestamp.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/avro_timestamp.q b/ql/src/test/queries/clientpositive/avro_timestamp.q index 7bf0dc8..847f250 100644 --- a/ql/src/test/queries/clientpositive/avro_timestamp.q +++ b/ql/src/test/queries/clientpositive/avro_timestamp.q @@ -1,7 +1,6 @@ set hive.mapred.mode=nonstrict; -- Exclude test on Windows due to space character being escaped in Hive paths on Windows. -- EXCLUDE_OS_WINDOWS --- JAVA_VERSION_SPECIFIC_OUTPUT DROP TABLE avro_timestamp_staging; DROP TABLE avro_timestamp; http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/cbo_rp_outer_join_ppr.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/cbo_rp_outer_join_ppr.q b/ql/src/test/queries/clientpositive/cbo_rp_outer_join_ppr.q index c497ce9..d8f726e 100644 --- a/ql/src/test/queries/clientpositive/cbo_rp_outer_join_ppr.q +++ b/ql/src/test/queries/clientpositive/cbo_rp_outer_join_ppr.q @@ -4,7 +4,6 @@ set hive.cbo.returnpath.hiveop=true; set hive.optimize.ppd=true; -- SORT_QUERY_RESULTS --- JAVA_VERSION_SPECIFIC_OUTPUT EXPLAIN EXTENDED FROM http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/char_udf1.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/char_udf1.q b/ql/src/test/queries/clientpositive/char_udf1.q index 09012b4..39aa0e0 100644 --- a/ql/src/test/queries/clientpositive/char_udf1.q +++ b/ql/src/test/queries/clientpositive/char_udf1.q @@ -4,7 +4,6 @@ create table char_udf_1 (c1 string, c2 string, c3 char(10), c4 char(20)); insert overwrite table char_udf_1 select key, value, key, value from src where key = '238' limit 1; --- JAVA_VERSION_SPECIFIC_OUTPUT -- UDFs with char support select http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/input4.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/input4.q b/ql/src/test/queries/clientpositive/input4.q index 83edbe2..90fcbdd 100644 --- a/ql/src/test/queries/clientpositive/input4.q +++ b/ql/src/test/queries/clientpositive/input4.q @@ -1,4 +1,3 @@ --- JAVA_VERSION_SPECIFIC_OUTPUT CREATE TABLE INPUT4(KEY STRING, VALUE STRING) STORED AS TEXTFILE; EXPLAIN http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/join0.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/join0.q b/ql/src/test/queries/clientpositive/join0.q index 66f2ef3..3252847 100644 --- a/ql/src/test/queries/clientpositive/join0.q +++ b/ql/src/test/queries/clientpositive/join0.q @@ -1,6 +1,5 @@ set hive.mapred.mode=nonstrict; set hive.explain.user=false; --- JAVA_VERSION_SPECIFIC_OUTPUT -- SORT_QUERY_RESULTS EXPLAIN http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/list_bucket_dml_10.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_10.q b/ql/src/test/queries/clientpositive/list_bucket_dml_10.q index 5d3dade..f25c174 100644 --- a/ql/src/test/queries/clientpositive/list_bucket_dml_10.q +++ b/ql/src/test/queries/clientpositive/list_bucket_dml_10.q @@ -1,7 +1,6 @@ set mapred.input.dir.recursive=true; -- run this test case in minimr to ensure it works in cluster --- JAVA_VERSION_SPECIFIC_OUTPUT -- list bucketing DML: static partition. multiple skewed columns. -- ds=2008-04-08/hr=11/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME: http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/list_bucket_dml_11.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_11.q b/ql/src/test/queries/clientpositive/list_bucket_dml_11.q index 2d22d66..8ac1627 100644 --- a/ql/src/test/queries/clientpositive/list_bucket_dml_11.q +++ b/ql/src/test/queries/clientpositive/list_bucket_dml_11.q @@ -6,7 +6,6 @@ set hive.merge.mapredfiles=false; -- Ensure it works if skewed column is not the first column in the table columns -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23) --- JAVA_VERSION_SPECIFIC_OUTPUT -- list bucketing DML: static partition. multiple skewed columns. http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/list_bucket_dml_12.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_12.q b/ql/src/test/queries/clientpositive/list_bucket_dml_12.q index ac063cc..9facfa5 100644 --- a/ql/src/test/queries/clientpositive/list_bucket_dml_12.q +++ b/ql/src/test/queries/clientpositive/list_bucket_dml_12.q @@ -7,7 +7,6 @@ set hive.merge.mapredfiles=false; -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23) -- SORT_QUERY_RESULTS --- JAVA_VERSION_SPECIFIC_OUTPUT -- test where the skewed values are more than 1 say columns no. 2 and 4 in a table with 5 columns create table list_bucketing_mul_col (col1 String, col2 String, col3 String, col4 String, col5 string) http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/list_bucket_dml_13.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_13.q b/ql/src/test/queries/clientpositive/list_bucket_dml_13.q index d68ca93..0fe7f61 100644 --- a/ql/src/test/queries/clientpositive/list_bucket_dml_13.q +++ b/ql/src/test/queries/clientpositive/list_bucket_dml_13.q @@ -7,7 +7,6 @@ set hive.merge.mapredfiles=false; -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23) -- SORT_QUERY_RESULTS --- JAVA_VERSION_SPECIFIC_OUTPUT -- test where the skewed values are more than 1 say columns no. 2 and 4 in a table with 5 columns create table list_bucketing_mul_col (col1 String, col2 String, col3 String, col4 String, col5 string) http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/list_bucket_dml_2.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_2.q b/ql/src/test/queries/clientpositive/list_bucket_dml_2.q index 263a002..c6dceab 100644 --- a/ql/src/test/queries/clientpositive/list_bucket_dml_2.q +++ b/ql/src/test/queries/clientpositive/list_bucket_dml_2.q @@ -10,7 +10,6 @@ set hive.stats.reliable=true; -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23) -- SORT_QUERY_RESULTS --- JAVA_VERSION_SPECIFIC_OUTPUT -- list bucketing DML: static partition. multiple skewed columns. -- ds=2008-04-08/hr=11/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME: http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/list_bucket_dml_4.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_4.q b/ql/src/test/queries/clientpositive/list_bucket_dml_4.q index 86ff342..950409d 100644 --- a/ql/src/test/queries/clientpositive/list_bucket_dml_4.q +++ b/ql/src/test/queries/clientpositive/list_bucket_dml_4.q @@ -9,7 +9,6 @@ set hive.merge.mapredfiles=false; -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23) -- SORT_QUERY_RESULTS --- JAVA_VERSION_SPECIFIC_OUTPUT -- list bucketing DML: static partition. multiple skewed columns. merge. -- ds=2008-04-08/hr=11/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME: http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/list_bucket_dml_5.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_5.q b/ql/src/test/queries/clientpositive/list_bucket_dml_5.q index ace7ba9..fce8e2e 100644 --- a/ql/src/test/queries/clientpositive/list_bucket_dml_5.q +++ b/ql/src/test/queries/clientpositive/list_bucket_dml_5.q @@ -10,7 +10,6 @@ set mapred.input.dir.recursive=true; -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23) -- SORT_QUERY_RESULTS --- JAVA_VERSION_SPECIFIC_OUTPUT -- create a skewed table create table list_bucketing_dynamic_part (key String, value String) http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/list_bucket_dml_6.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_6.q b/ql/src/test/queries/clientpositive/list_bucket_dml_6.q index 5684788..631c938 100644 --- a/ql/src/test/queries/clientpositive/list_bucket_dml_6.q +++ b/ql/src/test/queries/clientpositive/list_bucket_dml_6.q @@ -47,7 +47,6 @@ set hive.merge.mapredfiles=false; -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23) -- SORT_QUERY_RESULTS --- JAVA_VERSION_SPECIFIC_OUTPUT -- create a skewed table create table list_bucketing_dynamic_part (key String, value String) http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/list_bucket_dml_8.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_8.q b/ql/src/test/queries/clientpositive/list_bucket_dml_8.q index d904543..6d73896 100644 --- a/ql/src/test/queries/clientpositive/list_bucket_dml_8.q +++ b/ql/src/test/queries/clientpositive/list_bucket_dml_8.q @@ -48,7 +48,6 @@ set hive.merge.mapredfiles=false; -- 118 000002_0 -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23) --- JAVA_VERSION_SPECIFIC_OUTPUT -- create a skewed table create table list_bucketing_dynamic_part (key String, value String) http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/list_bucket_dml_9.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_9.q b/ql/src/test/queries/clientpositive/list_bucket_dml_9.q index 620750c..d2e24af 100644 --- a/ql/src/test/queries/clientpositive/list_bucket_dml_9.q +++ b/ql/src/test/queries/clientpositive/list_bucket_dml_9.q @@ -9,7 +9,6 @@ set hive.merge.mapredfiles=false; -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23) -- SORT_QUERY_RESULTS --- JAVA_VERSION_SPECIFIC_OUTPUT -- list bucketing DML: static partition. multiple skewed columns. merge. -- ds=2008-04-08/hr=11/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME: http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/outer_join_ppr.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/outer_join_ppr.q b/ql/src/test/queries/clientpositive/outer_join_ppr.q index 497a4d1..60a06ae 100644 --- a/ql/src/test/queries/clientpositive/outer_join_ppr.q +++ b/ql/src/test/queries/clientpositive/outer_join_ppr.q @@ -2,7 +2,6 @@ set hive.mapred.mode=nonstrict; set hive.optimize.ppd=true; -- SORT_QUERY_RESULTS --- JAVA_VERSION_SPECIFIC_OUTPUT EXPLAIN EXTENDED FROM http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/parquet_map_null.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/parquet_map_null.q b/ql/src/test/queries/clientpositive/parquet_map_null.q index 61058f3..e154159 100644 --- a/ql/src/test/queries/clientpositive/parquet_map_null.q +++ b/ql/src/test/queries/clientpositive/parquet_map_null.q @@ -1,5 +1,4 @@ -- This test attempts to write a parquet table from an avro table that contains map null values --- JAVA_VERSION_SPECIFIC_OUTPUT DROP TABLE IF EXISTS avro_table; DROP TABLE IF EXISTS parquet_table; http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/plan_json.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/plan_json.q b/ql/src/test/queries/clientpositive/plan_json.q index aa2b134..503b55d 100644 --- a/ql/src/test/queries/clientpositive/plan_json.q +++ b/ql/src/test/queries/clientpositive/plan_json.q @@ -1,5 +1,4 @@ -- explain plan json: the query gets the formatted json output of the query plan of the hive query --- JAVA_VERSION_SPECIFIC_OUTPUT EXPLAIN FORMATTED SELECT count(1) FROM src; http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/stats_list_bucket.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/stats_list_bucket.q b/ql/src/test/queries/clientpositive/stats_list_bucket.q index 51137a8..536702c 100644 --- a/ql/src/test/queries/clientpositive/stats_list_bucket.q +++ b/ql/src/test/queries/clientpositive/stats_list_bucket.q @@ -1,6 +1,5 @@ -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23) --- JAVA_VERSION_SPECIFIC_OUTPUT drop table stats_list_bucket; drop table stats_list_bucket_1; http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/str_to_map.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/str_to_map.q b/ql/src/test/queries/clientpositive/str_to_map.q index f2993b1..3280d89 100644 --- a/ql/src/test/queries/clientpositive/str_to_map.q +++ b/ql/src/test/queries/clientpositive/str_to_map.q @@ -1,7 +1,6 @@ set hive.mapred.mode=nonstrict; set hive.fetch.task.conversion=more; --- JAVA_VERSION_SPECIFIC_OUTPUT desc function str_to_map; desc function extended str_to_map; http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/subquery_multiinsert.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/subquery_multiinsert.q b/ql/src/test/queries/clientpositive/subquery_multiinsert.q index bea2e13..9d70f51 100644 --- a/ql/src/test/queries/clientpositive/subquery_multiinsert.q +++ b/ql/src/test/queries/clientpositive/subquery_multiinsert.q @@ -2,7 +2,6 @@ set hive.mapred.mode=nonstrict; set hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.PostExecutePrinter,org.apache.hadoop.hive.ql.hooks.PrintCompletedTasksHook; -- SORT_QUERY_RESULTS --- JAVA_VERSION_SPECIFIC_OUTPUT CREATE TABLE src_4( key STRING, http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/subquery_notin_having.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/subquery_notin_having.q b/ql/src/test/queries/clientpositive/subquery_notin_having.q index 8b2914d..05148df 100644 --- a/ql/src/test/queries/clientpositive/subquery_notin_having.q +++ b/ql/src/test/queries/clientpositive/subquery_notin_having.q @@ -1,6 +1,5 @@ set hive.mapred.mode=nonstrict; -- non agg, non corr --- JAVA_VERSION_SPECIFIC_OUTPUT explain select key, count(*) http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/varchar_udf1.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/varchar_udf1.q b/ql/src/test/queries/clientpositive/varchar_udf1.q index ff40b31..4d1f884 100644 --- a/ql/src/test/queries/clientpositive/varchar_udf1.q +++ b/ql/src/test/queries/clientpositive/varchar_udf1.q @@ -4,7 +4,6 @@ create table varchar_udf_1 (c1 string, c2 string, c3 varchar(10), c4 varchar(20) insert overwrite table varchar_udf_1 select key, value, key, value from src where key = '238' limit 1; --- JAVA_VERSION_SPECIFIC_OUTPUT -- UDFs with varchar support select http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/queries/clientpositive/vector_cast_constant.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/vector_cast_constant.q b/ql/src/test/queries/clientpositive/vector_cast_constant.q index c50dd8f..94bee09 100644 --- a/ql/src/test/queries/clientpositive/vector_cast_constant.q +++ b/ql/src/test/queries/clientpositive/vector_cast_constant.q @@ -2,7 +2,6 @@ set hive.mapred.mode=nonstrict; set hive.explain.user=false; SET hive.vectorized.execution.enabled=true; --- JAVA_VERSION_SPECIFIC_OUTPUT DROP TABLE over1k; DROP TABLE over1korc; http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.7.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.7.out b/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.7.out deleted file mode 100644 index 4ea70e3..0000000 --- a/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.7.out +++ /dev/null @@ -1,73 +0,0 @@ -PREHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT - -DROP TABLE Employee_Part -PREHOOK: type: DROPTABLE -POSTHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT - -DROP TABLE Employee_Part -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string) -row format delimited fields terminated by '|' stored as textfile -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@Employee_Part -POSTHOOK: query: CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string) -row format delimited fields terminated by '|' stored as textfile -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@Employee_Part -PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@employee_part -POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@employee_part -POSTHOOK: Output: default@employee_part@employeesalary=2000.0/country=USA -PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@employee_part -POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@employee_part -POSTHOOK: Output: default@employee_part@employeesalary=2000.0/country=UK -PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@employee_part -POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@employee_part -POSTHOOK: Output: default@employee_part@employeesalary=3000.0/country=USA -PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@employee_part -POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@employee_part -POSTHOOK: Output: default@employee_part@employeesalary=4000.0/country=USA -PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@employee_part -POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@employee_part -POSTHOOK: Output: default@employee_part@employeesalary=3500.0/country=UK -PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@employee_part -POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@employee_part -POSTHOOK: Output: default@employee_part@employeesalary=3000.0/country=UK -FAILED: SemanticException [Error 30007]: Invalid partitioning key/value specified in ANALYZE statement : {employeesalary=4000.0, country=Canada} http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.8.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.8.out b/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.8.out deleted file mode 100644 index 7cae55e..0000000 --- a/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.java1.8.out +++ /dev/null @@ -1,73 +0,0 @@ -PREHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT - -DROP TABLE Employee_Part -PREHOOK: type: DROPTABLE -POSTHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT - -DROP TABLE Employee_Part -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string) -row format delimited fields terminated by '|' stored as textfile -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@Employee_Part -POSTHOOK: query: CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string) -row format delimited fields terminated by '|' stored as textfile -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@Employee_Part -PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@employee_part -POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@employee_part -POSTHOOK: Output: default@employee_part@employeesalary=2000.0/country=USA -PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@employee_part -POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@employee_part -POSTHOOK: Output: default@employee_part@employeesalary=2000.0/country=UK -PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@employee_part -POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@employee_part -POSTHOOK: Output: default@employee_part@employeesalary=3000.0/country=USA -PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@employee_part -POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@employee_part -POSTHOOK: Output: default@employee_part@employeesalary=4000.0/country=USA -PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@employee_part -POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@employee_part -POSTHOOK: Output: default@employee_part@employeesalary=3500.0/country=UK -PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@employee_part -POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@employee_part -POSTHOOK: Output: default@employee_part@employeesalary=3000.0/country=UK -FAILED: SemanticException [Error 30007]: Invalid partitioning key/value specified in ANALYZE statement : {country=Canada, employeesalary=4000.0} http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.out b/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.out new file mode 100644 index 0000000..3261f78 --- /dev/null +++ b/ql/src/test/results/clientnegative/columnstats_partlvl_invalid_values.q.out @@ -0,0 +1,69 @@ +PREHOOK: query: DROP TABLE Employee_Part +PREHOOK: type: DROPTABLE +POSTHOOK: query: DROP TABLE Employee_Part +POSTHOOK: type: DROPTABLE +PREHOOK: query: CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string) +row format delimited fields terminated by '|' stored as textfile +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@Employee_Part +POSTHOOK: query: CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string) +row format delimited fields terminated by '|' stored as textfile +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@Employee_Part +PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@employee_part +POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@employee_part +POSTHOOK: Output: default@employee_part@employeesalary=2000.0/country=USA +PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@employee_part +POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@employee_part +POSTHOOK: Output: default@employee_part@employeesalary=2000.0/country=UK +PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@employee_part +POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@employee_part +POSTHOOK: Output: default@employee_part@employeesalary=3000.0/country=USA +PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@employee_part +POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@employee_part +POSTHOOK: Output: default@employee_part@employeesalary=4000.0/country=USA +PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@employee_part +POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@employee_part +POSTHOOK: Output: default@employee_part@employeesalary=3500.0/country=UK +PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@employee_part +POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@employee_part +POSTHOOK: Output: default@employee_part@employeesalary=3000.0/country=UK +FAILED: SemanticException [Error 30007]: Invalid partitioning key/value specified in ANALYZE statement : {employeesalary=4000.0, country=Canada} http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/results/clientpositive/authorization_explain.q.java1.7.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/authorization_explain.q.java1.7.out b/ql/src/test/results/clientpositive/authorization_explain.q.java1.7.out deleted file mode 100644 index fefb50c..0000000 --- a/ql/src/test/results/clientpositive/authorization_explain.q.java1.7.out +++ /dev/null @@ -1,44 +0,0 @@ -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT - -explain authorization select * from src join srcpart -PREHOOK: type: QUERY -POSTHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT - -explain authorization select * from src join srcpart -POSTHOOK: type: QUERY -INPUTS: - default@src - default@srcpart - default@srcpart@ds=2008-04-08/hr=11 - default@srcpart@ds=2008-04-08/hr=12 - default@srcpart@ds=2008-04-09/hr=11 - default@srcpart@ds=2008-04-09/hr=12 -OUTPUTS: -#### A masked pattern was here #### -CURRENT_USER: - hive_test_user -OPERATION: - QUERY -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: explain formatted authorization select * from src join srcpart -PREHOOK: type: QUERY -POSTHOOK: query: explain formatted authorization select * from src join srcpart -POSTHOOK: type: QUERY -#### A masked pattern was here #### -PREHOOK: query: explain authorization use default -PREHOOK: type: SWITCHDATABASE -POSTHOOK: query: explain authorization use default -POSTHOOK: type: SWITCHDATABASE -INPUTS: - database:default -OUTPUTS: -CURRENT_USER: - hive_test_user -OPERATION: - SWITCHDATABASE -PREHOOK: query: explain formatted authorization use default -PREHOOK: type: SWITCHDATABASE -POSTHOOK: query: explain formatted authorization use default -POSTHOOK: type: SWITCHDATABASE -{"INPUTS":["database:default"],"OUTPUTS":[],"CURRENT_USER":"hive_test_user","OPERATION":"SWITCHDATABASE"} http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/results/clientpositive/authorization_explain.q.java1.8.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/authorization_explain.q.java1.8.out b/ql/src/test/results/clientpositive/authorization_explain.q.java1.8.out deleted file mode 100644 index b7ec209..0000000 --- a/ql/src/test/results/clientpositive/authorization_explain.q.java1.8.out +++ /dev/null @@ -1,47 +0,0 @@ -Warning: Shuffle Join JOIN[7][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT - -explain authorization select * from src join srcpart -PREHOOK: type: QUERY -POSTHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT - -explain authorization select * from src join srcpart -POSTHOOK: type: QUERY -INPUTS: - default@src - default@srcpart - default@srcpart@ds=2008-04-08/hr=11 - default@srcpart@ds=2008-04-08/hr=12 - default@srcpart@ds=2008-04-09/hr=11 - default@srcpart@ds=2008-04-09/hr=12 -OUTPUTS: -#### A masked pattern was here #### -CURRENT_USER: - hive_test_user -OPERATION: - QUERY -AUTHORIZATION_FAILURES: - No privilege 'Select' found for inputs { database:default, table:src, columnName:key} - No privilege 'Select' found for inputs { database:default, table:srcpart, columnName:key} -Warning: Shuffle Join JOIN[7][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: explain formatted authorization select * from src join srcpart -PREHOOK: type: QUERY -POSTHOOK: query: explain formatted authorization select * from src join srcpart -POSTHOOK: type: QUERY -#### A masked pattern was here #### -PREHOOK: query: explain authorization use default -PREHOOK: type: SWITCHDATABASE -POSTHOOK: query: explain authorization use default -POSTHOOK: type: SWITCHDATABASE -INPUTS: - database:default -OUTPUTS: -CURRENT_USER: - hive_test_user -OPERATION: - SWITCHDATABASE -PREHOOK: query: explain formatted authorization use default -PREHOOK: type: SWITCHDATABASE -POSTHOOK: query: explain formatted authorization use default -POSTHOOK: type: SWITCHDATABASE -{"INPUTS":["database:default"],"OUTPUTS":[],"CURRENT_USER":"hive_test_user","OPERATION":"SWITCHDATABASE"} http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/results/clientpositive/authorization_explain.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/authorization_explain.q.out b/ql/src/test/results/clientpositive/authorization_explain.q.out new file mode 100644 index 0000000..851b845 --- /dev/null +++ b/ql/src/test/results/clientpositive/authorization_explain.q.out @@ -0,0 +1,40 @@ +Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product +PREHOOK: query: explain authorization select * from src join srcpart +PREHOOK: type: QUERY +POSTHOOK: query: explain authorization select * from src join srcpart +POSTHOOK: type: QUERY +INPUTS: + default@src + default@srcpart + default@srcpart@ds=2008-04-08/hr=11 + default@srcpart@ds=2008-04-08/hr=12 + default@srcpart@ds=2008-04-09/hr=11 + default@srcpart@ds=2008-04-09/hr=12 +OUTPUTS: +#### A masked pattern was here #### +CURRENT_USER: + hive_test_user +OPERATION: + QUERY +Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product +PREHOOK: query: explain formatted authorization select * from src join srcpart +PREHOOK: type: QUERY +POSTHOOK: query: explain formatted authorization select * from src join srcpart +POSTHOOK: type: QUERY +#### A masked pattern was here #### +PREHOOK: query: explain authorization use default +PREHOOK: type: SWITCHDATABASE +POSTHOOK: query: explain authorization use default +POSTHOOK: type: SWITCHDATABASE +INPUTS: + database:default +OUTPUTS: +CURRENT_USER: + hive_test_user +OPERATION: + SWITCHDATABASE +PREHOOK: query: explain formatted authorization use default +PREHOOK: type: SWITCHDATABASE +POSTHOOK: query: explain formatted authorization use default +POSTHOOK: type: SWITCHDATABASE +{"INPUTS":["database:default"],"OUTPUTS":[],"CURRENT_USER":"hive_test_user","OPERATION":"SWITCHDATABASE"} http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/results/clientpositive/avro_date.q.java1.7.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/avro_date.q.java1.7.out b/ql/src/test/results/clientpositive/avro_date.q.java1.7.out deleted file mode 100644 index 501b983..0000000 --- a/ql/src/test/results/clientpositive/avro_date.q.java1.7.out +++ /dev/null @@ -1,130 +0,0 @@ -PREHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT - -DROP TABLE avro_date_staging -PREHOOK: type: DROPTABLE -POSTHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT - -DROP TABLE avro_date_staging -POSTHOOK: type: DROPTABLE -PREHOOK: query: DROP TABLE avro_date -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE avro_date -POSTHOOK: type: DROPTABLE -PREHOOK: query: DROP TABLE avro_date_casts -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE avro_date_casts -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE avro_date_staging (d date, m1 map<string, date>, l1 array<date>) - ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' - COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':' - STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@avro_date_staging -POSTHOOK: query: CREATE TABLE avro_date_staging (d date, m1 map<string, date>, l1 array<date>) - ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' - COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':' - STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@avro_date_staging -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/avro_date.txt' OVERWRITE INTO TABLE avro_date_staging -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@avro_date_staging -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/avro_date.txt' OVERWRITE INTO TABLE avro_date_staging -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@avro_date_staging -PREHOOK: query: CREATE TABLE avro_date (d date, m1 map<string, date>, l1 array<date>) - PARTITIONED BY (p1 int, p2 date) - ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' - COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':' - STORED AS AVRO -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@avro_date -POSTHOOK: query: CREATE TABLE avro_date (d date, m1 map<string, date>, l1 array<date>) - PARTITIONED BY (p1 int, p2 date) - ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' - COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':' - STORED AS AVRO -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@avro_date -PREHOOK: query: INSERT OVERWRITE TABLE avro_date PARTITION(p1=2, p2='2014-09-26') SELECT * FROM avro_date_staging -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_date_staging -PREHOOK: Output: default@avro_date@p1=2/p2=2014-09-26 -POSTHOOK: query: INSERT OVERWRITE TABLE avro_date PARTITION(p1=2, p2='2014-09-26') SELECT * FROM avro_date_staging -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_date_staging -POSTHOOK: Output: default@avro_date@p1=2/p2=2014-09-26 -POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).d SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:d, type:date, comment:null), ] -POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).l1 SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:l1, type:array<date>, comment:null), ] -POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).m1 SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:m1, type:map<string,date>, comment:null), ] -PREHOOK: query: SELECT * FROM avro_date -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_date -PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM avro_date -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_date -POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -2012-02-21 {"foo":"1980-12-16","bar":"1998-05-07"} ["2011-09-04","2011-09-05"] 2 2014-09-26 -2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26 -1947-02-11 {"baz":"1921-12-16"} ["2011-09-05"] 2 2014-09-26 -8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26 -PREHOOK: query: SELECT d, COUNT(d) FROM avro_date GROUP BY d -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_date -PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -POSTHOOK: query: SELECT d, COUNT(d) FROM avro_date GROUP BY d -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_date -POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -1947-02-11 1 -2012-02-21 1 -2014-02-11 1 -8200-02-11 1 -PREHOOK: query: SELECT * FROM avro_date WHERE d!='1947-02-11' -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_date -PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM avro_date WHERE d!='1947-02-11' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_date -POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -2012-02-21 {"foo":"1980-12-16","bar":"1998-05-07"} ["2011-09-04","2011-09-05"] 2 2014-09-26 -2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26 -8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26 -PREHOOK: query: SELECT * FROM avro_date WHERE d<'2014-12-21' -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_date -PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM avro_date WHERE d<'2014-12-21' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_date -POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -2012-02-21 {"foo":"1980-12-16","bar":"1998-05-07"} ["2011-09-04","2011-09-05"] 2 2014-09-26 -2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26 -1947-02-11 {"baz":"1921-12-16"} ["2011-09-05"] 2 2014-09-26 -PREHOOK: query: SELECT * FROM avro_date WHERE d>'8000-12-01' -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_date -PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM avro_date WHERE d>'8000-12-01' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_date -POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26 http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/results/clientpositive/avro_date.q.java1.8.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/avro_date.q.java1.8.out b/ql/src/test/results/clientpositive/avro_date.q.java1.8.out deleted file mode 100644 index dea51c6..0000000 --- a/ql/src/test/results/clientpositive/avro_date.q.java1.8.out +++ /dev/null @@ -1,130 +0,0 @@ -PREHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT - -DROP TABLE avro_date_staging -PREHOOK: type: DROPTABLE -POSTHOOK: query: -- JAVA_VERSION_SPECIFIC_OUTPUT - -DROP TABLE avro_date_staging -POSTHOOK: type: DROPTABLE -PREHOOK: query: DROP TABLE avro_date -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE avro_date -POSTHOOK: type: DROPTABLE -PREHOOK: query: DROP TABLE avro_date_casts -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE avro_date_casts -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE avro_date_staging (d date, m1 map<string, date>, l1 array<date>) - ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' - COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':' - STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@avro_date_staging -POSTHOOK: query: CREATE TABLE avro_date_staging (d date, m1 map<string, date>, l1 array<date>) - ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' - COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':' - STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@avro_date_staging -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/avro_date.txt' OVERWRITE INTO TABLE avro_date_staging -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@avro_date_staging -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/avro_date.txt' OVERWRITE INTO TABLE avro_date_staging -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@avro_date_staging -PREHOOK: query: CREATE TABLE avro_date (d date, m1 map<string, date>, l1 array<date>) - PARTITIONED BY (p1 int, p2 date) - ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' - COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':' - STORED AS AVRO -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@avro_date -POSTHOOK: query: CREATE TABLE avro_date (d date, m1 map<string, date>, l1 array<date>) - PARTITIONED BY (p1 int, p2 date) - ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' - COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':' - STORED AS AVRO -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@avro_date -PREHOOK: query: INSERT OVERWRITE TABLE avro_date PARTITION(p1=2, p2='2014-09-26') SELECT * FROM avro_date_staging -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_date_staging -PREHOOK: Output: default@avro_date@p1=2/p2=2014-09-26 -POSTHOOK: query: INSERT OVERWRITE TABLE avro_date PARTITION(p1=2, p2='2014-09-26') SELECT * FROM avro_date_staging -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_date_staging -POSTHOOK: Output: default@avro_date@p1=2/p2=2014-09-26 -POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).d SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:d, type:date, comment:null), ] -POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).l1 SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:l1, type:array<date>, comment:null), ] -POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).m1 SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:m1, type:map<string,date>, comment:null), ] -PREHOOK: query: SELECT * FROM avro_date -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_date -PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM avro_date -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_date -POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -2012-02-21 {"bar":"1998-05-07","foo":"1980-12-16"} ["2011-09-04","2011-09-05"] 2 2014-09-26 -2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26 -1947-02-11 {"baz":"1921-12-16"} ["2011-09-05"] 2 2014-09-26 -8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26 -PREHOOK: query: SELECT d, COUNT(d) FROM avro_date GROUP BY d -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_date -PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -POSTHOOK: query: SELECT d, COUNT(d) FROM avro_date GROUP BY d -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_date -POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -1947-02-11 1 -2012-02-21 1 -2014-02-11 1 -8200-02-11 1 -PREHOOK: query: SELECT * FROM avro_date WHERE d!='1947-02-11' -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_date -PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM avro_date WHERE d!='1947-02-11' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_date -POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -2012-02-21 {"bar":"1998-05-07","foo":"1980-12-16"} ["2011-09-04","2011-09-05"] 2 2014-09-26 -2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26 -8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26 -PREHOOK: query: SELECT * FROM avro_date WHERE d<'2014-12-21' -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_date -PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM avro_date WHERE d<'2014-12-21' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_date -POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -2012-02-21 {"bar":"1998-05-07","foo":"1980-12-16"} ["2011-09-04","2011-09-05"] 2 2014-09-26 -2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26 -1947-02-11 {"baz":"1921-12-16"} ["2011-09-05"] 2 2014-09-26 -PREHOOK: query: SELECT * FROM avro_date WHERE d>'8000-12-01' -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_date -PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM avro_date WHERE d>'8000-12-01' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_date -POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 -#### A masked pattern was here #### -8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26 http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/results/clientpositive/avro_date.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/avro_date.q.out b/ql/src/test/results/clientpositive/avro_date.q.out new file mode 100644 index 0000000..32501cf --- /dev/null +++ b/ql/src/test/results/clientpositive/avro_date.q.out @@ -0,0 +1,126 @@ +PREHOOK: query: DROP TABLE avro_date_staging +PREHOOK: type: DROPTABLE +POSTHOOK: query: DROP TABLE avro_date_staging +POSTHOOK: type: DROPTABLE +PREHOOK: query: DROP TABLE avro_date +PREHOOK: type: DROPTABLE +POSTHOOK: query: DROP TABLE avro_date +POSTHOOK: type: DROPTABLE +PREHOOK: query: DROP TABLE avro_date_casts +PREHOOK: type: DROPTABLE +POSTHOOK: query: DROP TABLE avro_date_casts +POSTHOOK: type: DROPTABLE +PREHOOK: query: CREATE TABLE avro_date_staging (d date, m1 map<string, date>, l1 array<date>) + ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' + COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':' + STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@avro_date_staging +POSTHOOK: query: CREATE TABLE avro_date_staging (d date, m1 map<string, date>, l1 array<date>) + ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' + COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':' + STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@avro_date_staging +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/avro_date.txt' OVERWRITE INTO TABLE avro_date_staging +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@avro_date_staging +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/avro_date.txt' OVERWRITE INTO TABLE avro_date_staging +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@avro_date_staging +PREHOOK: query: CREATE TABLE avro_date (d date, m1 map<string, date>, l1 array<date>) + PARTITIONED BY (p1 int, p2 date) + ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' + COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':' + STORED AS AVRO +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@avro_date +POSTHOOK: query: CREATE TABLE avro_date (d date, m1 map<string, date>, l1 array<date>) + PARTITIONED BY (p1 int, p2 date) + ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' + COLLECTION ITEMS TERMINATED BY ',' MAP KEYS TERMINATED BY ':' + STORED AS AVRO +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@avro_date +PREHOOK: query: INSERT OVERWRITE TABLE avro_date PARTITION(p1=2, p2='2014-09-26') SELECT * FROM avro_date_staging +PREHOOK: type: QUERY +PREHOOK: Input: default@avro_date_staging +PREHOOK: Output: default@avro_date@p1=2/p2=2014-09-26 +POSTHOOK: query: INSERT OVERWRITE TABLE avro_date PARTITION(p1=2, p2='2014-09-26') SELECT * FROM avro_date_staging +POSTHOOK: type: QUERY +POSTHOOK: Input: default@avro_date_staging +POSTHOOK: Output: default@avro_date@p1=2/p2=2014-09-26 +POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).d SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:d, type:date, comment:null), ] +POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).l1 SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:l1, type:array<date>, comment:null), ] +POSTHOOK: Lineage: avro_date PARTITION(p1=2,p2=2014-09-26).m1 SIMPLE [(avro_date_staging)avro_date_staging.FieldSchema(name:m1, type:map<string,date>, comment:null), ] +PREHOOK: query: SELECT * FROM avro_date +PREHOOK: type: QUERY +PREHOOK: Input: default@avro_date +PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM avro_date +POSTHOOK: type: QUERY +POSTHOOK: Input: default@avro_date +POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 +#### A masked pattern was here #### +2012-02-21 {"bar":"1998-05-07","foo":"1980-12-16"} ["2011-09-04","2011-09-05"] 2 2014-09-26 +2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26 +1947-02-11 {"baz":"1921-12-16"} ["2011-09-05"] 2 2014-09-26 +8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26 +PREHOOK: query: SELECT d, COUNT(d) FROM avro_date GROUP BY d +PREHOOK: type: QUERY +PREHOOK: Input: default@avro_date +PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 +#### A masked pattern was here #### +POSTHOOK: query: SELECT d, COUNT(d) FROM avro_date GROUP BY d +POSTHOOK: type: QUERY +POSTHOOK: Input: default@avro_date +POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 +#### A masked pattern was here #### +1947-02-11 1 +2012-02-21 1 +2014-02-11 1 +8200-02-11 1 +PREHOOK: query: SELECT * FROM avro_date WHERE d!='1947-02-11' +PREHOOK: type: QUERY +PREHOOK: Input: default@avro_date +PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM avro_date WHERE d!='1947-02-11' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@avro_date +POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 +#### A masked pattern was here #### +2012-02-21 {"bar":"1998-05-07","foo":"1980-12-16"} ["2011-09-04","2011-09-05"] 2 2014-09-26 +2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26 +8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26 +PREHOOK: query: SELECT * FROM avro_date WHERE d<'2014-12-21' +PREHOOK: type: QUERY +PREHOOK: Input: default@avro_date +PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM avro_date WHERE d<'2014-12-21' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@avro_date +POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 +#### A masked pattern was here #### +2012-02-21 {"bar":"1998-05-07","foo":"1980-12-16"} ["2011-09-04","2011-09-05"] 2 2014-09-26 +2014-02-11 {"baz":"1981-12-16"} ["2011-09-05"] 2 2014-09-26 +1947-02-11 {"baz":"1921-12-16"} ["2011-09-05"] 2 2014-09-26 +PREHOOK: query: SELECT * FROM avro_date WHERE d>'8000-12-01' +PREHOOK: type: QUERY +PREHOOK: Input: default@avro_date +PREHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM avro_date WHERE d>'8000-12-01' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@avro_date +POSTHOOK: Input: default@avro_date@p1=2/p2=2014-09-26 +#### A masked pattern was here #### +8200-02-11 {"baz":"6981-12-16"} ["1039-09-05"] 2 2014-09-26 http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.7.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.7.out b/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.7.out deleted file mode 100644 index 8f8065e..0000000 --- a/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.7.out +++ /dev/null @@ -1,57 +0,0 @@ -PREHOOK: query: -- These test attempts to deserialize an Avro file that contains map null values, and the file schema --- vs record schema have the null values in different positions --- i.e. --- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"] --- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}] - --- JAVA_VERSION_SPECIFIC_OUTPUT - -DROP TABLE IF EXISTS avro_table -PREHOOK: type: DROPTABLE -POSTHOOK: query: -- These test attempts to deserialize an Avro file that contains map null values, and the file schema --- vs record schema have the null values in different positions --- i.e. --- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"] --- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}] - --- JAVA_VERSION_SPECIFIC_OUTPUT - -DROP TABLE IF EXISTS avro_table -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE avro_table (avreau_col_1 map<string,string>) STORED AS AVRO -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@avro_table -POSTHOOK: query: CREATE TABLE avro_table (avreau_col_1 map<string,string>) STORED AS AVRO -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@avro_table -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/map_null_val.avro' OVERWRITE INTO TABLE avro_table -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@avro_table -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/map_null_val.avro' OVERWRITE INTO TABLE avro_table -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@avro_table -PREHOOK: query: SELECT * FROM avro_table -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_table -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM avro_table -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_table -#### A masked pattern was here #### -{"key4":null,"key3":"val3"} -{"key4":null,"key3":"val3"} -{"key2":"val2","key1":null} -{"key4":null,"key3":"val3"} -{"key4":null,"key3":"val3"} -PREHOOK: query: DROP TABLE avro_table -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@avro_table -PREHOOK: Output: default@avro_table -POSTHOOK: query: DROP TABLE avro_table -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@avro_table -POSTHOOK: Output: default@avro_table http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.8.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.8.out b/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.8.out deleted file mode 100644 index 127d8b3..0000000 --- a/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.java1.8.out +++ /dev/null @@ -1,57 +0,0 @@ -PREHOOK: query: -- These test attempts to deserialize an Avro file that contains map null values, and the file schema --- vs record schema have the null values in different positions --- i.e. --- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"] --- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}] - --- JAVA_VERSION_SPECIFIC_OUTPUT - -DROP TABLE IF EXISTS avro_table -PREHOOK: type: DROPTABLE -POSTHOOK: query: -- These test attempts to deserialize an Avro file that contains map null values, and the file schema --- vs record schema have the null values in different positions --- i.e. --- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"] --- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}] - --- JAVA_VERSION_SPECIFIC_OUTPUT - -DROP TABLE IF EXISTS avro_table -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE avro_table (avreau_col_1 map<string,string>) STORED AS AVRO -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@avro_table -POSTHOOK: query: CREATE TABLE avro_table (avreau_col_1 map<string,string>) STORED AS AVRO -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@avro_table -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/map_null_val.avro' OVERWRITE INTO TABLE avro_table -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@avro_table -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/map_null_val.avro' OVERWRITE INTO TABLE avro_table -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@avro_table -PREHOOK: query: SELECT * FROM avro_table -PREHOOK: type: QUERY -PREHOOK: Input: default@avro_table -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM avro_table -POSTHOOK: type: QUERY -POSTHOOK: Input: default@avro_table -#### A masked pattern was here #### -{"key3":"val3","key4":null} -{"key3":"val3","key4":null} -{"key1":null,"key2":"val2"} -{"key3":"val3","key4":null} -{"key3":"val3","key4":null} -PREHOOK: query: DROP TABLE avro_table -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@avro_table -PREHOOK: Output: default@avro_table -POSTHOOK: query: DROP TABLE avro_table -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@avro_table -POSTHOOK: Output: default@avro_table http://git-wip-us.apache.org/repos/asf/hive/blob/9687dcc9/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.out b/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.out new file mode 100644 index 0000000..2d983f1 --- /dev/null +++ b/ql/src/test/results/clientpositive/avro_deserialize_map_null.q.out @@ -0,0 +1,55 @@ +PREHOOK: query: -- These test attempts to deserialize an Avro file that contains map null values, and the file schema +-- vs record schema have the null values in different positions +-- i.e. +-- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"] +-- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}] + + +DROP TABLE IF EXISTS avro_table +PREHOOK: type: DROPTABLE +POSTHOOK: query: -- These test attempts to deserialize an Avro file that contains map null values, and the file schema +-- vs record schema have the null values in different positions +-- i.e. +-- fileSchema = [{ "type" : "map", "values" : ["string","null"]}, "null"] +-- recordSchema = ["null", { "type" : "map", "values" : ["string","null"]}] + + +DROP TABLE IF EXISTS avro_table +POSTHOOK: type: DROPTABLE +PREHOOK: query: CREATE TABLE avro_table (avreau_col_1 map<string,string>) STORED AS AVRO +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@avro_table +POSTHOOK: query: CREATE TABLE avro_table (avreau_col_1 map<string,string>) STORED AS AVRO +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@avro_table +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/map_null_val.avro' OVERWRITE INTO TABLE avro_table +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@avro_table +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/map_null_val.avro' OVERWRITE INTO TABLE avro_table +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@avro_table +PREHOOK: query: SELECT * FROM avro_table +PREHOOK: type: QUERY +PREHOOK: Input: default@avro_table +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM avro_table +POSTHOOK: type: QUERY +POSTHOOK: Input: default@avro_table +#### A masked pattern was here #### +{"key3":"val3","key4":null} +{"key3":"val3","key4":null} +{"key1":null,"key2":"val2"} +{"key3":"val3","key4":null} +{"key3":"val3","key4":null} +PREHOOK: query: DROP TABLE avro_table +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@avro_table +PREHOOK: Output: default@avro_table +POSTHOOK: query: DROP TABLE avro_table +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@avro_table +POSTHOOK: Output: default@avro_table