This is an automated email from the ASF dual-hosted git repository.
zhangbutao pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hive.git
The following commit(s) were added to refs/heads/master by this push:
new 81b394104b0 HIVE-23558: Remove compute_stats UDAF (#4928)(Butao Zhang,
reviewed by Ayush Saxena)
81b394104b0 is described below
commit 81b394104b08efdc42d782d444462f4fdc12a8fd
Author: Butao Zhang <[email protected]>
AuthorDate: Sun Dec 24 21:23:03 2023 +0800
HIVE-23558: Remove compute_stats UDAF (#4928)(Butao Zhang, reviewed by
Ayush Saxena)
---
data/files/datasets/temp_udf/load.hive.sql | 1 +
.../hadoop/hive/ql/exec/FunctionRegistry.java | 1 -
.../queries/clientnegative/compute_stats_long.q | 7 -
ql/src/test/queries/clientpositive/char_udf1.q | 1 +
.../queries/clientpositive/compute_stats_binary.q | 9 --
.../queries/clientpositive/compute_stats_boolean.q | 9 --
.../queries/clientpositive/compute_stats_date.q | 28 ----
.../queries/clientpositive/compute_stats_decimal.q | 11 --
.../queries/clientpositive/compute_stats_double.q | 9 --
.../clientpositive/compute_stats_empty_table.q | 12 --
.../queries/clientpositive/compute_stats_long.q | 9 --
.../queries/clientpositive/compute_stats_string.q | 9 --
.../reduceSinkDeDuplication_pRS_key_empty.q | 1 +
ql/src/test/queries/clientpositive/varchar_udf1.q | 1 +
ql/src/test/queries/clientpositive/vector_udf1.q | 1 +
.../clientnegative/compute_stats_long.q.out | 57 -------
.../clientpositive/llap/compute_stats_binary.q.out | 34 ----
.../llap/compute_stats_boolean.q.out | 34 ----
.../clientpositive/llap/compute_stats_date.q.out | 176 ---------------------
.../llap/compute_stats_decimal.q.out | Bin 1512 -> 0 bytes
.../clientpositive/llap/compute_stats_double.q.out | Bin 1443 -> 0 bytes
.../llap/compute_stats_empty_table.q.out | 62 --------
.../clientpositive/llap/compute_stats_long.q.out | Bin 1376 -> 0 bytes
.../clientpositive/llap/compute_stats_string.q.out | Bin 1448 -> 0 bytes
.../clientpositive/llap/show_functions.q.out | 3 -
25 files changed, 5 insertions(+), 470 deletions(-)
diff --git a/data/files/datasets/temp_udf/load.hive.sql
b/data/files/datasets/temp_udf/load.hive.sql
new file mode 100644
index 00000000000..b8a21d37a23
--- /dev/null
+++ b/data/files/datasets/temp_udf/load.hive.sql
@@ -0,0 +1 @@
+CREATE FUNCTION compute_stats AS
'org.apache.hadoop.hive.ql.udf.generic.GenericUDAFComputeStats';
\ No newline at end of file
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
index c0b9519944d..6c1c5fb0113 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
@@ -563,7 +563,6 @@ public final class FunctionRegistry {
system.registerGenericUDAF("ngrams", new GenericUDAFnGrams());
system.registerGenericUDAF("context_ngrams", new
GenericUDAFContextNGrams());
- system.registerGenericUDAF("compute_stats", new GenericUDAFComputeStats());
system.registerGenericUDF("ndv_compute_bit_vector",
GenericUDFNDVComputeBitVector.class);
system.registerGenericUDAF("compute_bit_vector_hll", new
GenericUDAFComputeBitVectorHLL());
system.registerGenericUDAF("compute_bit_vector_fm", new
GenericUDAFComputeBitVectorFMSketch());
diff --git a/ql/src/test/queries/clientnegative/compute_stats_long.q
b/ql/src/test/queries/clientnegative/compute_stats_long.q
deleted file mode 100644
index bf87393abd0..00000000000
--- a/ql/src/test/queries/clientnegative/compute_stats_long.q
+++ /dev/null
@@ -1,7 +0,0 @@
-create table tab_int(a int);
-
--- insert some data
-LOAD DATA LOCAL INPATH "../../data/files/int.txt" INTO TABLE tab_int;
-
--- this should raise an error since the number of bit vectors has a hard limit
at 1024
-select compute_stats(a, 'fm', 10000) from tab_int;
diff --git a/ql/src/test/queries/clientpositive/char_udf1.q
b/ql/src/test/queries/clientpositive/char_udf1.q
index 20440436150..4a6ac4020c3 100644
--- a/ql/src/test/queries/clientpositive/char_udf1.q
+++ b/ql/src/test/queries/clientpositive/char_udf1.q
@@ -1,4 +1,5 @@
--! qt:dataset:src
+--! qt:dataset:temp_udf
drop table char_udf_1;
create table char_udf_1 (c1 string, c2 string, c3 char(10), c4 char(20));
diff --git a/ql/src/test/queries/clientpositive/compute_stats_binary.q
b/ql/src/test/queries/clientpositive/compute_stats_binary.q
deleted file mode 100644
index fd15634f202..00000000000
--- a/ql/src/test/queries/clientpositive/compute_stats_binary.q
+++ /dev/null
@@ -1,9 +0,0 @@
-create table tab_binary(a binary);
-
--- insert some data
-LOAD DATA LOCAL INPATH "../../data/files/binary.txt" INTO TABLE tab_binary;
-
-select count(*) from tab_binary;
-
--- compute statistical summary of data
-select compute_stats(a, 16) from tab_binary;
diff --git a/ql/src/test/queries/clientpositive/compute_stats_boolean.q
b/ql/src/test/queries/clientpositive/compute_stats_boolean.q
deleted file mode 100644
index cddb53f8f64..00000000000
--- a/ql/src/test/queries/clientpositive/compute_stats_boolean.q
+++ /dev/null
@@ -1,9 +0,0 @@
-create table tab_bool(a boolean);
-
--- insert some data
-LOAD DATA LOCAL INPATH "../../data/files/bool.txt" INTO TABLE tab_bool;
-
-select count(*) from tab_bool;
-
--- compute statistical summary of data
-select compute_stats(a, 16) from tab_bool;
diff --git a/ql/src/test/queries/clientpositive/compute_stats_date.q
b/ql/src/test/queries/clientpositive/compute_stats_date.q
deleted file mode 100644
index bf478526ba8..00000000000
--- a/ql/src/test/queries/clientpositive/compute_stats_date.q
+++ /dev/null
@@ -1,28 +0,0 @@
-
-create table tab_date (
- origin_city_name string,
- dest_city_name string,
- fl_date date,
- arr_delay float,
- fl_num int
-);
-
--- insert some data
-load data local inpath '../../data/files/flights_join.txt' overwrite into
table tab_date;
-
-select count(*) from tab_date;
-
--- compute statistical summary of data
-select compute_stats(fl_date, 'hll') from tab_date;
-
-explain
-analyze table tab_date compute statistics for columns fl_date;
-
-analyze table tab_date compute statistics for columns fl_date;
-
-describe formatted tab_date fl_date;
-
--- Update stats manually. Try both yyyy-mm-dd and integer value for high/low
value
-alter table tab_date update statistics for column fl_date set ('numDVs'='19',
'highValue'='2015-01-01', 'lowValue'='0');
-
-describe formatted tab_date fl_date;
diff --git a/ql/src/test/queries/clientpositive/compute_stats_decimal.q
b/ql/src/test/queries/clientpositive/compute_stats_decimal.q
deleted file mode 100644
index 2beafaf219a..00000000000
--- a/ql/src/test/queries/clientpositive/compute_stats_decimal.q
+++ /dev/null
@@ -1,11 +0,0 @@
-set hive.stats.autogather=true;
-
-create table tab_decimal(a decimal(35,3));
-
--- insert some data
-LOAD DATA LOCAL INPATH "../../data/files/decimal.txt" INTO TABLE tab_decimal;
-
-select count(*) from tab_decimal;
-
--- compute statistical summary of data
-select compute_stats(a, 'fm', 18) from tab_decimal;
diff --git a/ql/src/test/queries/clientpositive/compute_stats_double.q
b/ql/src/test/queries/clientpositive/compute_stats_double.q
deleted file mode 100644
index 6bae0643a89..00000000000
--- a/ql/src/test/queries/clientpositive/compute_stats_double.q
+++ /dev/null
@@ -1,9 +0,0 @@
-create table tab_double(a double);
-
--- insert some data
-LOAD DATA LOCAL INPATH "../../data/files/double.txt" INTO TABLE tab_double;
-
-select count(*) from tab_double;
-
--- compute statistical summary of data
-select compute_stats(a, 'fm', 16) from tab_double;
diff --git a/ql/src/test/queries/clientpositive/compute_stats_empty_table.q
b/ql/src/test/queries/clientpositive/compute_stats_empty_table.q
deleted file mode 100644
index d4ed93fa639..00000000000
--- a/ql/src/test/queries/clientpositive/compute_stats_empty_table.q
+++ /dev/null
@@ -1,12 +0,0 @@
-create table tab_empty(a boolean, b int, c double, d string, e binary);
-
-select count(*) from tab_empty;
-
--- compute statistical summary of data
-select compute_stats(a, 16) from tab_empty;
-select compute_stats(b, 16) from tab_empty;
-select compute_stats(c, 16) from tab_empty;
-select compute_stats(d, 16) from tab_empty;
-select compute_stats(e, 16) from tab_empty;
-
-
diff --git a/ql/src/test/queries/clientpositive/compute_stats_long.q
b/ql/src/test/queries/clientpositive/compute_stats_long.q
deleted file mode 100644
index 48f4ebb9791..00000000000
--- a/ql/src/test/queries/clientpositive/compute_stats_long.q
+++ /dev/null
@@ -1,9 +0,0 @@
-create table tab_int(a int);
-
--- insert some data
-LOAD DATA LOCAL INPATH "../../data/files/int.txt" INTO TABLE tab_int;
-
-select count(*) from tab_int;
-
--- compute statistical summary of data
-select compute_stats(a, 'fm', 16) from tab_int;
diff --git a/ql/src/test/queries/clientpositive/compute_stats_string.q
b/ql/src/test/queries/clientpositive/compute_stats_string.q
deleted file mode 100644
index 79a531e8ec4..00000000000
--- a/ql/src/test/queries/clientpositive/compute_stats_string.q
+++ /dev/null
@@ -1,9 +0,0 @@
-create table tab_string(a string);
-
--- insert some data
-LOAD DATA LOCAL INPATH "../../data/files/string.txt" INTO TABLE tab_string;
-
-select count(*) from tab_string;
-
--- compute statistical summary of data
-select compute_stats(a, 'fm', 16) from tab_string;
diff --git
a/ql/src/test/queries/clientpositive/reduceSinkDeDuplication_pRS_key_empty.q
b/ql/src/test/queries/clientpositive/reduceSinkDeDuplication_pRS_key_empty.q
index f92229560ae..8af3ff49385 100644
--- a/ql/src/test/queries/clientpositive/reduceSinkDeDuplication_pRS_key_empty.q
+++ b/ql/src/test/queries/clientpositive/reduceSinkDeDuplication_pRS_key_empty.q
@@ -1,4 +1,5 @@
--! qt:dataset:src
+--! qt:dataset:temp_udf
set hive.mapred.mode=nonstrict;
set hive.cbo.enable=false;
diff --git a/ql/src/test/queries/clientpositive/varchar_udf1.q
b/ql/src/test/queries/clientpositive/varchar_udf1.q
index 7aee1ce2251..80e184a8c19 100644
--- a/ql/src/test/queries/clientpositive/varchar_udf1.q
+++ b/ql/src/test/queries/clientpositive/varchar_udf1.q
@@ -1,4 +1,5 @@
--! qt:dataset:src
+--! qt:dataset:temp_udf
drop table varchar_udf_1;
create table varchar_udf_1 (c1 string, c2 string, c3 varchar(10), c4
varchar(20));
diff --git a/ql/src/test/queries/clientpositive/vector_udf1.q
b/ql/src/test/queries/clientpositive/vector_udf1.q
index a258025dc10..5a642dafb28 100644
--- a/ql/src/test/queries/clientpositive/vector_udf1.q
+++ b/ql/src/test/queries/clientpositive/vector_udf1.q
@@ -1,4 +1,5 @@
--! qt:dataset:src
+--! qt:dataset:temp_udf
SET hive.vectorized.execution.enabled=true;
set hive.fetch.task.conversion=none;
diff --git a/ql/src/test/results/clientnegative/compute_stats_long.q.out
b/ql/src/test/results/clientnegative/compute_stats_long.q.out
deleted file mode 100644
index 990a9417907..00000000000
--- a/ql/src/test/results/clientnegative/compute_stats_long.q.out
+++ /dev/null
@@ -1,57 +0,0 @@
-PREHOOK: query: create table tab_int(a int)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@tab_int
-POSTHOOK: query: create table tab_int(a int)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@tab_int
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/int.txt" INTO TABLE
tab_int
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@tab_int
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/int.txt" INTO TABLE
tab_int
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@tab_int
-PREHOOK: query: select compute_stats(a, 'fm', 10000) from tab_int
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_int
-#### A masked pattern was here ####
-Status: Failed
-Vertex failed, vertexName=Map 1, vertexId=vertex_#ID#, diagnostics=[Task
failed, taskId=task_#ID#, diagnostics=[TaskAttempt 0 failed, info=[Error: Error
while running task ( failure ) : attempt_#ID#:java.lang.RuntimeException:
java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException:
Hive Runtime Error while processing row
-#### A masked pattern was here ####
-Caused by: java.lang.RuntimeException:
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while
processing row
-#### A masked pattern was here ####
-Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime
Error while processing row
-#### A masked pattern was here ####
-Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: The maximum
allowed value for number of bit vectors is 1024, but was passed 10000 bit
vectors
-#### A masked pattern was here ####
-], TaskAttempt 1 failed, info=[Error: Error while running task ( failure ) :
attempt_#ID#:java.lang.RuntimeException: java.lang.RuntimeException:
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while
processing row
-#### A masked pattern was here ####
-Caused by: java.lang.RuntimeException:
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while
processing row
-#### A masked pattern was here ####
-Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime
Error while processing row
-#### A masked pattern was here ####
-Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: The maximum
allowed value for number of bit vectors is 1024, but was passed 10000 bit
vectors
-#### A masked pattern was here ####
-]], Vertex did not succeed due to OWN_TASK_FAILURE, failedTasks:1
killedTasks:0, Vertex vertex_#ID# [Map 1] killed/failed due to:OWN_TASK_FAILURE]
-[Masked Vertex killed due to OTHER_VERTEX_FAILURE]
-DAG did not succeed due to VERTEX_FAILURE. failedVertices:1 killedVertices:1
-FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.tez.TezTask. Vertex failed, vertexName=Map 1,
vertexId=vertex_#ID#, diagnostics=[Task failed, taskId=task_#ID#,
diagnostics=[TaskAttempt 0 failed, info=[Error: Error while running task (
failure ) : attempt_#ID#:java.lang.RuntimeException:
java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException:
Hive Runtime Error while processing row
-#### A masked pattern was here ####
-Caused by: java.lang.RuntimeException:
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while
processing row
-#### A masked pattern was here ####
-Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime
Error while processing row
-#### A masked pattern was here ####
-Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: The maximum
allowed value for number of bit vectors is 1024, but was passed 10000 bit
vectors
-#### A masked pattern was here ####
-], TaskAttempt 1 failed, info=[Error: Error while running task ( failure ) :
attempt_#ID#:java.lang.RuntimeException: java.lang.RuntimeException:
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while
processing row
-#### A masked pattern was here ####
-Caused by: java.lang.RuntimeException:
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while
processing row
-#### A masked pattern was here ####
-Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime
Error while processing row
-#### A masked pattern was here ####
-Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: The maximum
allowed value for number of bit vectors is 1024, but was passed 10000 bit
vectors
-#### A masked pattern was here ####
-]], Vertex did not succeed due to OWN_TASK_FAILURE, failedTasks:1
killedTasks:0, Vertex vertex_#ID# [Map 1] killed/failed due
to:OWN_TASK_FAILURE][Masked Vertex killed due to OTHER_VERTEX_FAILURE]DAG did
not succeed due to VERTEX_FAILURE. failedVertices:1 killedVertices:1
diff --git a/ql/src/test/results/clientpositive/llap/compute_stats_binary.q.out
b/ql/src/test/results/clientpositive/llap/compute_stats_binary.q.out
deleted file mode 100644
index fc90c894cd9..00000000000
--- a/ql/src/test/results/clientpositive/llap/compute_stats_binary.q.out
+++ /dev/null
@@ -1,34 +0,0 @@
-PREHOOK: query: create table tab_binary(a binary)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@tab_binary
-POSTHOOK: query: create table tab_binary(a binary)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@tab_binary
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/binary.txt" INTO
TABLE tab_binary
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@tab_binary
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/binary.txt" INTO
TABLE tab_binary
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@tab_binary
-PREHOOK: query: select count(*) from tab_binary
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_binary
-#### A masked pattern was here ####
-POSTHOOK: query: select count(*) from tab_binary
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@tab_binary
-#### A masked pattern was here ####
-10
-PREHOOK: query: select compute_stats(a, 16) from tab_binary
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_binary
-#### A masked pattern was here ####
-POSTHOOK: query: select compute_stats(a, 16) from tab_binary
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@tab_binary
-#### A masked pattern was here ####
-{"columntype":"Binary","maxlength":58,"avglength":32.5,"countnulls":0}
diff --git
a/ql/src/test/results/clientpositive/llap/compute_stats_boolean.q.out
b/ql/src/test/results/clientpositive/llap/compute_stats_boolean.q.out
deleted file mode 100644
index be90ee22273..00000000000
--- a/ql/src/test/results/clientpositive/llap/compute_stats_boolean.q.out
+++ /dev/null
@@ -1,34 +0,0 @@
-PREHOOK: query: create table tab_bool(a boolean)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@tab_bool
-POSTHOOK: query: create table tab_bool(a boolean)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@tab_bool
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/bool.txt" INTO TABLE
tab_bool
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@tab_bool
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/bool.txt" INTO TABLE
tab_bool
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@tab_bool
-PREHOOK: query: select count(*) from tab_bool
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_bool
-#### A masked pattern was here ####
-POSTHOOK: query: select count(*) from tab_bool
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@tab_bool
-#### A masked pattern was here ####
-33
-PREHOOK: query: select compute_stats(a, 16) from tab_bool
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_bool
-#### A masked pattern was here ####
-POSTHOOK: query: select compute_stats(a, 16) from tab_bool
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@tab_bool
-#### A masked pattern was here ####
-{"columntype":"Boolean","counttrues":13,"countfalses":19,"countnulls":1}
diff --git a/ql/src/test/results/clientpositive/llap/compute_stats_date.q.out
b/ql/src/test/results/clientpositive/llap/compute_stats_date.q.out
deleted file mode 100644
index b33be22d858..00000000000
--- a/ql/src/test/results/clientpositive/llap/compute_stats_date.q.out
+++ /dev/null
@@ -1,176 +0,0 @@
-PREHOOK: query: create table tab_date (
- origin_city_name string,
- dest_city_name string,
- fl_date date,
- arr_delay float,
- fl_num int
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@tab_date
-POSTHOOK: query: create table tab_date (
- origin_city_name string,
- dest_city_name string,
- fl_date date,
- arr_delay float,
- fl_num int
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@tab_date
-PREHOOK: query: load data local inpath '../../data/files/flights_join.txt'
overwrite into table tab_date
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@tab_date
-POSTHOOK: query: load data local inpath '../../data/files/flights_join.txt'
overwrite into table tab_date
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@tab_date
-PREHOOK: query: select count(*) from tab_date
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_date
-#### A masked pattern was here ####
-POSTHOOK: query: select count(*) from tab_date
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@tab_date
-#### A masked pattern was here ####
-20
-PREHOOK: query: select compute_stats(fl_date, 'hll') from tab_date
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_date
-#### A masked pattern was here ####
-POSTHOOK: query: select compute_stats(fl_date, 'hll') from tab_date
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@tab_date
-#### A masked pattern was here ####
-{"columntype":"Date","min":"2000-11-20","max":"2010-10-29","countnulls":0,"numdistinctvalues":19,"ndvbitvector":HLL������������t������������������������R��ո��������������Y������������������������������������������������犞}
-PREHOOK: query: explain
-analyze table tab_date compute statistics for columns fl_date
-PREHOOK: type: ANALYZE_TABLE
-PREHOOK: Input: default@tab_date
-PREHOOK: Output: default@tab_date
-#### A masked pattern was here ####
-POSTHOOK: query: explain
-analyze table tab_date compute statistics for columns fl_date
-POSTHOOK: type: ANALYZE_TABLE
-POSTHOOK: Input: default@tab_date
-POSTHOOK: Output: default@tab_date
-#### A masked pattern was here ####
-STAGE DEPENDENCIES:
- Stage-0 is a root stage
- Stage-2 depends on stages: Stage-0
-
-STAGE PLANS:
- Stage: Stage-0
- Tez
-#### A masked pattern was here ####
- Edges:
- Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: tab_date
- Statistics: Num rows: 1 Data size: 56 Basic stats: COMPLETE
Column stats: NONE
- Select Operator
- expressions: fl_date (type: date)
- outputColumnNames: fl_date
- Statistics: Num rows: 1 Data size: 56 Basic stats:
COMPLETE Column stats: NONE
- Group By Operator
- aggregations: min(fl_date), max(fl_date), count(1),
count(fl_date), compute_bit_vector_hll(fl_date)
- minReductionHashAggr: 0.99
- mode: hash
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 1 Data size: 328 Basic stats:
COMPLETE Column stats: NONE
- Reduce Output Operator
- null sort order:
- sort order:
- Statistics: Num rows: 1 Data size: 328 Basic stats:
COMPLETE Column stats: NONE
- value expressions: _col0 (type: date), _col1 (type:
date), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: binary)
- Execution mode: llap
- LLAP IO: all inputs
- Reducer 2
- Execution mode: vectorized, llap
- Reduce Operator Tree:
- Group By Operator
- aggregations: min(VALUE._col0), max(VALUE._col1),
count(VALUE._col2), count(VALUE._col3), compute_bit_vector_hll(VALUE._col4)
- mode: mergepartial
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 1 Data size: 328 Basic stats: COMPLETE
Column stats: NONE
- Select Operator
- expressions: 'DATE' (type: string), _col0 (type: date),
_col1 (type: date), (_col2 - _col3) (type: bigint),
COALESCE(ndv_compute_bit_vector(_col4),0) (type: bigint), _col4 (type: binary)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
- Statistics: Num rows: 1 Data size: 328 Basic stats: COMPLETE
Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 328 Basic stats:
COMPLETE Column stats: NONE
- table:
- input format:
org.apache.hadoop.mapred.SequenceFileInputFormat
- output format:
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde:
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
- Stage: Stage-2
- Stats Work
- Basic Stats Work:
- Column Stats Desc:
- Columns: fl_date
- Column Types: date
- Table: default.tab_date
-
-PREHOOK: query: analyze table tab_date compute statistics for columns fl_date
-PREHOOK: type: ANALYZE_TABLE
-PREHOOK: Input: default@tab_date
-PREHOOK: Output: default@tab_date
-#### A masked pattern was here ####
-POSTHOOK: query: analyze table tab_date compute statistics for columns fl_date
-POSTHOOK: type: ANALYZE_TABLE
-POSTHOOK: Input: default@tab_date
-POSTHOOK: Output: default@tab_date
-#### A masked pattern was here ####
-PREHOOK: query: describe formatted tab_date fl_date
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@tab_date
-POSTHOOK: query: describe formatted tab_date fl_date
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@tab_date
-col_name fl_date
-data_type date
-min 2000-11-20
-max 2010-10-29
-num_nulls 0
-distinct_count 19
-avg_col_len
-max_col_len
-num_trues
-num_falses
-bit_vector HL
-comment from deserializer
-COLUMN_STATS_ACCURATE
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"fl_date\":\"true\"}}
-PREHOOK: query: alter table tab_date update statistics for column fl_date set
('numDVs'='19', 'highValue'='2015-01-01', 'lowValue'='0')
-PREHOOK: type: ALTERTABLE_UPDATETABLESTATS
-PREHOOK: Input: default@tab_date
-PREHOOK: Output: default@tab_date
-POSTHOOK: query: alter table tab_date update statistics for column fl_date set
('numDVs'='19', 'highValue'='2015-01-01', 'lowValue'='0')
-POSTHOOK: type: ALTERTABLE_UPDATETABLESTATS
-POSTHOOK: Input: default@tab_date
-POSTHOOK: Output: default@tab_date
-PREHOOK: query: describe formatted tab_date fl_date
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@tab_date
-POSTHOOK: query: describe formatted tab_date fl_date
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@tab_date
-col_name fl_date
-data_type date
-min 1970-01-01
-max 2015-01-01
-num_nulls 0
-distinct_count 19
-avg_col_len
-max_col_len
-num_trues
-num_falses
-bit_vector HL
-comment from deserializer
-COLUMN_STATS_ACCURATE
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"fl_date\":\"true\"}}
diff --git
a/ql/src/test/results/clientpositive/llap/compute_stats_decimal.q.out
b/ql/src/test/results/clientpositive/llap/compute_stats_decimal.q.out
deleted file mode 100644
index 810a8b232af..00000000000
Binary files
a/ql/src/test/results/clientpositive/llap/compute_stats_decimal.q.out and
/dev/null differ
diff --git a/ql/src/test/results/clientpositive/llap/compute_stats_double.q.out
b/ql/src/test/results/clientpositive/llap/compute_stats_double.q.out
deleted file mode 100644
index c6a90202249..00000000000
Binary files
a/ql/src/test/results/clientpositive/llap/compute_stats_double.q.out and
/dev/null differ
diff --git
a/ql/src/test/results/clientpositive/llap/compute_stats_empty_table.q.out
b/ql/src/test/results/clientpositive/llap/compute_stats_empty_table.q.out
deleted file mode 100644
index 2ce83e14bf1..00000000000
--- a/ql/src/test/results/clientpositive/llap/compute_stats_empty_table.q.out
+++ /dev/null
@@ -1,62 +0,0 @@
-PREHOOK: query: create table tab_empty(a boolean, b int, c double, d string, e
binary)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@tab_empty
-POSTHOOK: query: create table tab_empty(a boolean, b int, c double, d string,
e binary)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@tab_empty
-PREHOOK: query: select count(*) from tab_empty
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_empty
-#### A masked pattern was here ####
-POSTHOOK: query: select count(*) from tab_empty
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@tab_empty
-#### A masked pattern was here ####
-0
-PREHOOK: query: select compute_stats(a, 16) from tab_empty
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_empty
-#### A masked pattern was here ####
-POSTHOOK: query: select compute_stats(a, 16) from tab_empty
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@tab_empty
-#### A masked pattern was here ####
-{"columntype":"Boolean","counttrues":0,"countfalses":0,"countnulls":0}
-PREHOOK: query: select compute_stats(b, 16) from tab_empty
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_empty
-#### A masked pattern was here ####
-POSTHOOK: query: select compute_stats(b, 16) from tab_empty
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@tab_empty
-#### A masked pattern was here ####
-{"columntype":"Long","min":null,"max":null,"countnulls":0,"numdistinctvalues":0,"ndvbitvector":}
-PREHOOK: query: select compute_stats(c, 16) from tab_empty
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_empty
-#### A masked pattern was here ####
-POSTHOOK: query: select compute_stats(c, 16) from tab_empty
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@tab_empty
-#### A masked pattern was here ####
-{"columntype":"Double","min":null,"max":null,"countnulls":0,"numdistinctvalues":0,"ndvbitvector":}
-PREHOOK: query: select compute_stats(d, 16) from tab_empty
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_empty
-#### A masked pattern was here ####
-POSTHOOK: query: select compute_stats(d, 16) from tab_empty
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@tab_empty
-#### A masked pattern was here ####
-{"columntype":"String","maxlength":0,"avglength":0.0,"countnulls":0,"numdistinctvalues":0,"ndvbitvector":}
-PREHOOK: query: select compute_stats(e, 16) from tab_empty
-PREHOOK: type: QUERY
-PREHOOK: Input: default@tab_empty
-#### A masked pattern was here ####
-POSTHOOK: query: select compute_stats(e, 16) from tab_empty
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@tab_empty
-#### A masked pattern was here ####
-{"columntype":"Binary","maxlength":0,"avglength":0.0,"countnulls":0}
diff --git a/ql/src/test/results/clientpositive/llap/compute_stats_long.q.out
b/ql/src/test/results/clientpositive/llap/compute_stats_long.q.out
deleted file mode 100644
index 138ee3f1047..00000000000
Binary files a/ql/src/test/results/clientpositive/llap/compute_stats_long.q.out
and /dev/null differ
diff --git a/ql/src/test/results/clientpositive/llap/compute_stats_string.q.out
b/ql/src/test/results/clientpositive/llap/compute_stats_string.q.out
deleted file mode 100644
index 135ac45e754..00000000000
Binary files
a/ql/src/test/results/clientpositive/llap/compute_stats_string.q.out and
/dev/null differ
diff --git a/ql/src/test/results/clientpositive/llap/show_functions.q.out
b/ql/src/test/results/clientpositive/llap/show_functions.q.out
index e5a24964711..9642bfbd8b0 100644
--- a/ql/src/test/results/clientpositive/llap/show_functions.q.out
+++ b/ql/src/test/results/clientpositive/llap/show_functions.q.out
@@ -83,7 +83,6 @@ collect_list
collect_set
compute_bit_vector_fm
compute_bit_vector_hll
-compute_stats
concat
concat_ws
context_ngrams
@@ -533,7 +532,6 @@ collect_list
collect_set
compute_bit_vector_fm
compute_bit_vector_hll
-compute_stats
concat
concat_ws
context_ngrams
@@ -713,7 +711,6 @@ collect_list
collect_set
compute_bit_vector_fm
compute_bit_vector_hll
-compute_stats
concat
concat_ws
context_ngrams