http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java index c0e56e7..375516d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java @@ -45,8 +45,8 @@ import org.apache.hadoop.hive.ql.optimizer.ReduceSinkMapJoinProc; import org.apache.hadoop.hive.ql.optimizer.RemoveDynamicPruningBySize; import org.apache.hadoop.hive.ql.optimizer.SetReducerParallelism; import org.apache.hadoop.hive.ql.optimizer.metainfo.annotation.AnnotateWithOpTraits; -import org.apache.hadoop.hive.ql.optimizer.physical.AnnotateRunTimeStatsOptimizer; import org.apache.hadoop.hive.ql.optimizer.physical.CrossProductCheck; +import org.apache.hadoop.hive.ql.optimizer.physical.AnnotateRunTimeStatsOptimizer; import org.apache.hadoop.hive.ql.optimizer.physical.LlapDecider; import org.apache.hadoop.hive.ql.optimizer.physical.MemoryDecider; import org.apache.hadoop.hive.ql.optimizer.physical.MetadataOnlyOptimizer;
http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/java/org/apache/hadoop/hive/ql/stats/StatsUtils.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/stats/StatsUtils.java b/ql/src/java/org/apache/hadoop/hive/ql/stats/StatsUtils.java index 0f1e3a3..7f3be26 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/stats/StatsUtils.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/stats/StatsUtils.java @@ -1341,7 +1341,7 @@ public class StatsUtils { || colType.startsWith(serdeConstants.MAP_TYPE_NAME) || colType.startsWith(serdeConstants.STRUCT_TYPE_NAME) || colType.startsWith(serdeConstants.UNION_TYPE_NAME)) { - avgColSize = getAvgColLenOfVariableLengthTypes(conf, oi, colType); + avgColSize = getAvgColLenOf(conf, oi, colType); } else { avgColSize = getAvgColLenOfFixedLengthTypes(colType); } http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/queries/clientpositive/explainanalyze_1.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/explainanalyze_1.q b/ql/src/test/queries/clientpositive/explainanalyze_1.q index a4b3dc5..0a33a4f 100644 --- a/ql/src/test/queries/clientpositive/explainanalyze_1.q +++ b/ql/src/test/queries/clientpositive/explainanalyze_1.q @@ -1,3 +1,4 @@ +set hive.map.aggr=false; set hive.mapred.mode=nonstrict; explain analyze select * from src a union all select * from src b limit 10; http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/queries/clientpositive/explainanalyze_2.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/explainanalyze_2.q b/ql/src/test/queries/clientpositive/explainanalyze_2.q index dfee826..7637101 100644 --- a/ql/src/test/queries/clientpositive/explainanalyze_2.q +++ b/ql/src/test/queries/clientpositive/explainanalyze_2.q @@ -1,199 +1,11 @@ -set hive.explain.user=true; -set hive.metastore.aggregate.stats.cache.enabled=false; - --- SORT_QUERY_RESULTS - -CREATE TABLE dest_j1(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE; - -CREATE TABLE ss(k1 STRING,v1 STRING,k2 STRING,v2 STRING,k3 STRING,v3 STRING) STORED AS TEXTFILE; - -CREATE TABLE sr(k1 STRING,v1 STRING,k2 STRING,v2 STRING,k3 STRING,v3 STRING) STORED AS TEXTFILE; - -CREATE TABLE cs(k1 STRING,v1 STRING,k2 STRING,v2 STRING,k3 STRING,v3 STRING) STORED AS TEXTFILE; - -INSERT OVERWRITE TABLE ss -SELECT x.key,x.value,y.key,y.value,z.key,z.value -FROM src1 x -JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11); - -INSERT OVERWRITE TABLE sr -SELECT x.key,x.value,y.key,y.value,z.key,z.value -FROM src1 x -JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=12); - -INSERT OVERWRITE TABLE cs -SELECT x.key,x.value,y.key,y.value,z.key,z.value -FROM src1 x -JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08'); - - -ANALYZE TABLE ss COMPUTE STATISTICS; -ANALYZE TABLE ss COMPUTE STATISTICS FOR COLUMNS k1,v1,k2,v2,k3,v3; - -ANALYZE TABLE sr COMPUTE STATISTICS; -ANALYZE TABLE sr COMPUTE STATISTICS FOR COLUMNS k1,v1,k2,v2,k3,v3; - -ANALYZE TABLE cs COMPUTE STATISTICS; -ANALYZE TABLE cs COMPUTE STATISTICS FOR COLUMNS k1,v1,k2,v2,k3,v3; - -set hive.auto.convert.join=false; - -explain analyze -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11); - -explain analyze -select -ss.k1,sr.k2,cs.k3,count(ss.v1),count(sr.v2),count(cs.v3) -FROM -ss,sr,cs,src d1,src d2,src d3,src1,srcpart -where - ss.k1 = d1.key -and sr.k1 = d2.key -and cs.k1 = d3.key -and ss.k2 = sr.k2 -and ss.k3 = sr.k3 -and ss.v1 = src1.value -and ss.v2 = srcpart.value -and sr.v2 = cs.v2 -and sr.v3 = cs.v3 -and ss.v3='ssv3' -and sr.v1='srv1' -and src1.key = 'src1key' -and srcpart.key = 'srcpartkey' -and d1.value = 'd1value' -and d2.value in ('2000Q1','2000Q2','2000Q3') -and d3.value in ('2000Q1','2000Q2','2000Q3') -group by -ss.k1,sr.k2,cs.k3 -order by -ss.k1,sr.k2,cs.k3 -limit 100; - -explain analyze -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select * from src1 union select * from src)z ON (x.value = z.value) -union -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select * from src1 union select * from src)z ON (x.value = z.value); - -explain analyze -SELECT x.key, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select * from src1 union select * from src)z ON (x.value = z.value) -union -SELECT x.key, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select key, value from src1 union select key, value from src union select key, value from src)z ON (x.value = z.value) -union -SELECT x.key, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select key, value from src1 union select key, value from src union select key, value from src union select key, value from src)z ON (x.value = z.value); - - -set hive.auto.convert.join=true; -set hive.auto.convert.join.noconditionaltask=true; -set hive.auto.convert.join.noconditionaltask.size=10000; -set hive.stats.fetch.column.stats=false; - - -explain analyze -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11); - -explain analyze -select -ss.k1,sr.k2,cs.k3,count(ss.v1),count(sr.v2),count(cs.v3) -FROM -ss,sr,cs,src d1,src d2,src d3,src1,srcpart -where - ss.k1 = d1.key -and sr.k1 = d2.key -and cs.k1 = d3.key -and ss.k2 = sr.k2 -and ss.k3 = sr.k3 -and ss.v1 = src1.value -and ss.v2 = srcpart.value -and sr.v2 = cs.v2 -and sr.v3 = cs.v3 -and ss.v3='ssv3' -and sr.v1='srv1' -and src1.key = 'src1key' -and srcpart.key = 'srcpartkey' -and d1.value = 'd1value' -and d2.value in ('2000Q1','2000Q2','2000Q3') -and d3.value in ('2000Q1','2000Q2','2000Q3') -group by -ss.k1,sr.k2,cs.k3 -order by -ss.k1,sr.k2,cs.k3 -limit 100; - -explain analyze -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select * from src1 union select * from src)z ON (x.value = z.value) -union -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select * from src1 union select * from src)z ON (x.value = z.value); - -explain analyze -SELECT x.key, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select * from src1 union select * from src)z ON (x.value = z.value) -union -SELECT x.key, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select key, value from src1 union select key, value from src union select key, value from src)z ON (x.value = z.value) -union -SELECT x.key, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select key, value from src1 union select key, value from src union select key, value from src union select key, value from src)z ON (x.value = z.value); - - -set hive.auto.convert.join=true; -set hive.auto.convert.join.noconditionaltask=true; -set hive.auto.convert.join.noconditionaltask.size=10000; -set hive.auto.convert.sortmerge.join.bigtable.selection.policy = org.apache.hadoop.hive.ql.optimizer.TableSizeBasedBigTableSelectorForAutoSMJ; - -CREATE TABLE srcbucket_mapjoin(key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE; -CREATE TABLE tab_part (key int, value string) PARTITIONED BY(ds STRING) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE; -CREATE TABLE srcbucket_mapjoin_part (key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE; - -load data local inpath '../../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08'); -load data local inpath '../../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08'); - -load data local inpath '../../data/files/srcbucket20.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08'); -load data local inpath '../../data/files/srcbucket21.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08'); -load data local inpath '../../data/files/srcbucket22.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08'); -load data local inpath '../../data/files/srcbucket23.txt' INTO TABLE srcbucket_mapjoin_part partition(ds='2008-04-08'); - +set hive.map.aggr=false; +set hive.explain.user=true; -set hive.optimize.bucketingsorting=false; -insert overwrite table tab_part partition (ds='2008-04-08') -select key,value from srcbucket_mapjoin_part; +CREATE TABLE tab(key int, value string) PARTITIONED BY(ds STRING); -CREATE TABLE tab(key int, value string) PARTITIONED BY(ds STRING) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE; insert overwrite table tab partition (ds='2008-04-08') -select key,value from srcbucket_mapjoin; - -CREATE TABLE tab2(key int, value string) PARTITIONED BY(ds STRING) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE; -insert overwrite table tab2 partition (ds='2008-04-08') -select key,value from srcbucket_mapjoin; - -set hive.convert.join.bucket.mapjoin.tez = false; -set hive.auto.convert.sortmerge.join = true; - -set hive.auto.convert.join.noconditionaltask.size=500; +select key,value from src; explain analyze select s1.key as key, s1.value as value from tab s1 join tab s3 on s1.key=s3.key; @@ -201,129 +13,3 @@ select s1.key as key, s1.value as value from tab s1 join tab s3 on s1.key=s3.key explain analyze select s1.key as key, s1.value as value from tab s1 join tab s3 on s1.key=s3.key join tab s2 on s1.value=s2.value; -explain analyze -select s1.key as key, s1.value as value from tab s1 join tab2 s3 on s1.key=s3.key; - -explain analyze -select s1.key as key, s1.value as value from tab s1 join tab2 s3 on s1.key=s3.key join tab2 s2 on s1.value=s2.value; - -explain analyze -select count(*) from (select s1.key as key, s1.value as value from tab s1 join tab s3 on s1.key=s3.key -UNION ALL -select s2.key as key, s2.value as value from tab s2 -) a join tab_part b on (a.key = b.key); - -explain analyze -select count(*) from (select s1.key as key, s1.value as value from tab s1 join tab s3 on s1.key=s3.key join tab s2 on s1.value=s2.value -UNION ALL -select s2.key as key, s2.value as value from tab s2 -) a join tab_part b on (a.key = b.key); - -explain analyze -SELECT x.key, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select * from src1 union all select * from src)z ON (x.value = z.value) -union all -SELECT x.key, y.value -FROM src x JOIN src y ON (x.key = y.key) -JOIN (select key, value from src1 union all select key, value from src union all select key, value from src)z ON (x.value = z.value) -union all -SELECT x.key, y.value -FROM src1 x JOIN src1 y ON (x.key = y.key) -JOIN (select key, value from src1 union all select key, value from src union all select key, value from src union all select key, value from src)z ON (x.value = z.value); - -explain analyze -SELECT x.key, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select * from src1 union select * from src)z ON (x.value = z.value) -union -SELECT x.key, y.value -FROM src x JOIN src y ON (x.key = y.key) -JOIN (select key, value from src1 union select key, value from src union select key, value from src)z ON (x.value = z.value) -union -SELECT x.key, y.value -FROM src1 x JOIN src1 y ON (x.key = y.key) -JOIN (select key, value from src1 union select key, value from src union select key, value from src union select key, value from src)z ON (x.value = z.value); - -CREATE TABLE a(key STRING, value STRING) STORED AS TEXTFILE; -CREATE TABLE b(key STRING, value STRING) STORED AS TEXTFILE; -CREATE TABLE c(key STRING, value STRING) STORED AS TEXTFILE; - -explain analyze -from -( -SELECT x.key, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN (select * from src1 union all select * from src)z ON (x.value = z.value) -union all -SELECT x.key, y.value -FROM src x JOIN src y ON (x.key = y.key) -JOIN (select key, value from src1 union all select key, value from src union all select key, value from src)z ON (x.value = z.value) -union all -SELECT x.key, y.value -FROM src1 x JOIN src1 y ON (x.key = y.key) -JOIN (select key, value from src1 union all select key, value from src union all select key, value from src union all select key, value from src)z ON (x.value = z.value) -) tmp -INSERT OVERWRITE TABLE a SELECT tmp.key, tmp.value -INSERT OVERWRITE TABLE b SELECT tmp.key, tmp.value -INSERT OVERWRITE TABLE c SELECT tmp.key, tmp.value; - -explain analyze -FROM -( -SELECT x.key as key, y.value as value from src1 x JOIN src y ON (x.key = y.key) -JOIN (select * from src1 union select * from src)z ON (x.value = z.value) -union -SELECT x.key as key, y.value as value from src x JOIN src y ON (x.key = y.key) -JOIN (select key, value from src1 union select key, value from src union select key, value from src)z ON (x.value = z.value) -union -SELECT x.key as key, y.value as value from src1 x JOIN src1 y ON (x.key = y.key) -JOIN (select key, value from src1 union select key, value from src union select key, value from src union select key, value from src)z ON (x.value = z.value) -) tmp -INSERT OVERWRITE TABLE a SELECT tmp.key, tmp.value -INSERT OVERWRITE TABLE b SELECT tmp.key, tmp.value -INSERT OVERWRITE TABLE c SELECT tmp.key, tmp.value; - - -CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE; -CREATE TABLE DEST2(key STRING, val1 STRING, val2 STRING) STORED AS TEXTFILE; - -explain analyze -FROM (select 'tst1' as key, cast(count(1) as string) as value from src s1 - UNION DISTINCT - select s2.key as key, s2.value as value from src s2) unionsrc -INSERT OVERWRITE TABLE DEST1 SELECT unionsrc.key, COUNT(DISTINCT SUBSTR(unionsrc.value,5)) GROUP BY unionsrc.key -INSERT OVERWRITE TABLE DEST2 SELECT unionsrc.key, unionsrc.value, COUNT(DISTINCT SUBSTR(unionsrc.value,5)) GROUP BY unionsrc.key, unionsrc.value; - -explain analyze FROM UNIQUEJOIN PRESERVE src a (a.key), PRESERVE src1 b (b.key), PRESERVE srcpart c (c.key) SELECT a.key, b.key, c.key; - -set hive.entity.capture.transform=true; - -explain analyze -SELECT -TRANSFORM(a.key, a.value) USING 'cat' AS (tkey, tvalue) -FROM src a join src b -on a.key = b.key; - -explain analyze -FROM ( - select key, value from ( - select 'tst1' as key, cast(count(1) as string) as value, 'tst1' as value2 from src s1 - UNION all - select s2.key as key, s2.value as value, 'tst1' as value2 from src s2) unionsub - UNION all - select key, value from src s0 - ) unionsrc -INSERT OVERWRITE TABLE DEST1 SELECT unionsrc.key, COUNT(DISTINCT SUBSTR(unionsrc.value,5)) GROUP BY unionsrc.key -INSERT OVERWRITE TABLE DEST2 SELECT unionsrc.key, unionsrc.value, COUNT(DISTINCT SUBSTR(unionsrc.value,5)) -GROUP BY unionsrc.key, unionsrc.value; - -explain analyze -FROM ( - select 'tst1' as key, cast(count(1) as string) as value, 'tst1' as value2 from src s1 - UNION all - select s2.key as key, s2.value as value, 'tst1' as value2 from src s2 - ) unionsrc -INSERT OVERWRITE TABLE DEST1 SELECT unionsrc.key, COUNT(DISTINCT SUBSTR(unionsrc.value,5)) GROUP BY unionsrc.key -INSERT OVERWRITE TABLE DEST2 SELECT unionsrc.key, unionsrc.value, COUNT(DISTINCT SUBSTR(unionsrc.value,5)) -GROUP BY unionsrc.key, unionsrc.value; http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/queries/clientpositive/explainanalyze_3.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/explainanalyze_3.q b/ql/src/test/queries/clientpositive/explainanalyze_3.q index 69f82e5..53dfabd 100644 --- a/ql/src/test/queries/clientpositive/explainanalyze_3.q +++ b/ql/src/test/queries/clientpositive/explainanalyze_3.q @@ -1,3 +1,5 @@ +set hive.map.aggr=false; + set hive.security.authorization.manager=org.apache.hadoop.hive.ql.security.authorization.DefaultHiveAuthorizationProvider; set hive.metastore.filter.hook=org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl; set hive.mapred.mode=nonstrict; @@ -40,9 +42,13 @@ use default; drop database newDB; -explain analyze analyze table src compute statistics; +drop table src_stats; + +create table src_stats as select * from src; + +explain analyze analyze table src_stats compute statistics; -explain analyze analyze table src compute statistics for columns; +explain analyze analyze table src_stats compute statistics for columns; explain analyze CREATE TEMPORARY MACRO SIGMOID (x DOUBLE) 1.0 / (1.0 + EXP(-x)); http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/queries/clientpositive/explainanalyze_4.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/explainanalyze_4.q b/ql/src/test/queries/clientpositive/explainanalyze_4.q index dad397b..498a128 100644 --- a/ql/src/test/queries/clientpositive/explainanalyze_4.q +++ b/ql/src/test/queries/clientpositive/explainanalyze_4.q @@ -1,3 +1,5 @@ +set hive.map.aggr=false; + set hive.mapred.mode=nonstrict; set hive.explain.user=true; http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/queries/clientpositive/explainanalyze_5.q ---------------------------------------------------------------------- diff --git a/ql/src/test/queries/clientpositive/explainanalyze_5.q b/ql/src/test/queries/clientpositive/explainanalyze_5.q index bb23e45..696064c 100644 --- a/ql/src/test/queries/clientpositive/explainanalyze_5.q +++ b/ql/src/test/queries/clientpositive/explainanalyze_5.q @@ -1,8 +1,14 @@ +set hive.map.aggr=false; + set hive.stats.column.autogather=true; -explain analyze analyze table src compute statistics; +drop table src_stats; + +create table src_stats as select * from src; + +explain analyze analyze table src_stats compute statistics; -explain analyze analyze table src compute statistics for columns; +explain analyze analyze table src_stats compute statistics for columns; drop table src_multi2; http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/annotate_stats_join_pkfk.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/annotate_stats_join_pkfk.q.out b/ql/src/test/results/clientpositive/annotate_stats_join_pkfk.q.out index 71da136..b4d46d2 100644 --- a/ql/src/test/results/clientpositive/annotate_stats_join_pkfk.q.out +++ b/ql/src/test/results/clientpositive/annotate_stats_join_pkfk.q.out @@ -444,19 +444,19 @@ STAGE PLANS: Statistics: Num rows: 12 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE TableScan alias: ss - Statistics: Num rows: 1000 Data size: 3856 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1000 Data size: 3860 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ss_store_sk is not null (type: boolean) - Statistics: Num rows: 964 Data size: 3716 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 964 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: ss_store_sk (type: int) outputColumnNames: _col0 - Statistics: Num rows: 964 Data size: 3716 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 964 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: UDFToLong(_col0) (type: bigint) sort order: + Map-reduce partition columns: UDFToLong(_col0) (type: bigint) - Statistics: Num rows: 964 Data size: 3716 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 964 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE Reduce Operator Tree: Join Operator condition map: http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/dynamic_rdd_cache.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/dynamic_rdd_cache.q.out b/ql/src/test/results/clientpositive/dynamic_rdd_cache.q.out index 3a8da0e..87166a7 100644 --- a/ql/src/test/results/clientpositive/dynamic_rdd_cache.q.out +++ b/ql/src/test/results/clientpositive/dynamic_rdd_cache.q.out @@ -679,7 +679,7 @@ STAGE PLANS: Select Operator expressions: 'tst1' (type: string), UDFToString(_col0) (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: true table: @@ -692,12 +692,12 @@ STAGE PLANS: Map Operator Tree: TableScan Union - Statistics: Num rows: 167 Data size: 1935 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 167 Data size: 2035 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 167 Data size: 1935 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 167 Data size: 2035 Basic stats: COMPLETE Column stats: PARTIAL value expressions: _col1 (type: string) TableScan alias: s2 @@ -710,21 +710,21 @@ STAGE PLANS: outputColumnNames: _col0, _col1 Statistics: Num rows: 166 Data size: 1763 Basic stats: COMPLETE Column stats: NONE Union - Statistics: Num rows: 167 Data size: 1935 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 167 Data size: 2035 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 167 Data size: 1935 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 167 Data size: 2035 Basic stats: COMPLETE Column stats: PARTIAL value expressions: _col1 (type: string) TableScan Union - Statistics: Num rows: 167 Data size: 1935 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 167 Data size: 2035 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 167 Data size: 1935 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 167 Data size: 2035 Basic stats: COMPLETE Column stats: PARTIAL value expressions: _col1 (type: string) TableScan alias: s4 @@ -737,12 +737,12 @@ STAGE PLANS: outputColumnNames: _col0, _col1 Statistics: Num rows: 166 Data size: 1763 Basic stats: COMPLETE Column stats: NONE Union - Statistics: Num rows: 167 Data size: 1935 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 167 Data size: 2035 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 167 Data size: 1935 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 167 Data size: 2035 Basic stats: COMPLETE Column stats: PARTIAL value expressions: _col1 (type: string) Reduce Operator Tree: Join Operator @@ -752,10 +752,10 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 27889 Data size: 9593816 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 27889 Data size: 15171616 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: true - Statistics: Num rows: 27889 Data size: 9593816 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 27889 Data size: 15171616 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -787,7 +787,7 @@ STAGE PLANS: Select Operator expressions: 'tst1' (type: string), UDFToString(_col0) (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: true table: http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out index 1f25d49..385b5cf 100644 --- a/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out +++ b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out @@ -284,10 +284,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=1000) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Execution mode: llap LLAP IO: all inputs @@ -332,10 +332,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=1000) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Stage: Stage-0 @@ -681,10 +681,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=2000) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Execution mode: llap LLAP IO: all inputs @@ -714,10 +714,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=1000) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Select Operator expressions: _col0 (type: string) @@ -727,10 +727,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=1000) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Execution mode: llap LLAP IO: all inputs @@ -797,10 +797,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=2000) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Reducer 6 Execution mode: llap @@ -809,10 +809,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=1000) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Reducer 7 Execution mode: llap @@ -821,10 +821,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=1000) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Stage: Stage-0 @@ -1056,10 +1056,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=1000) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Select Operator expressions: _col1 (type: string) @@ -1069,10 +1069,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=1000) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Execution mode: llap LLAP IO: all inputs @@ -1117,10 +1117,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=1000) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Reducer 6 Execution mode: llap @@ -1129,10 +1129,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=1000) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Stage: Stage-0 @@ -1367,10 +1367,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=1000) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Execution mode: llap LLAP IO: all inputs @@ -1418,10 +1418,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=2200) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Reducer 3 Execution mode: llap @@ -1464,10 +1464,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=2200) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Reducer 7 Execution mode: llap @@ -1476,10 +1476,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=1000) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Stage: Stage-0 @@ -1697,10 +1697,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=1000) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Execution mode: llap LLAP IO: all inputs @@ -1726,10 +1726,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=1000) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Stage: Stage-0 http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/llap/subquery_multi.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/llap/subquery_multi.q.out b/ql/src/test/results/clientpositive/llap/subquery_multi.q.out index 854fb78..0cc1781 100644 --- a/ql/src/test/results/clientpositive/llap/subquery_multi.q.out +++ b/ql/src/test/results/clientpositive/llap/subquery_multi.q.out @@ -3148,10 +3148,10 @@ STAGE PLANS: aggregations: avg(l_quantity) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: struct<count:bigint,sum:double,input:double>) Execution mode: llap LLAP IO: no inputs @@ -3206,10 +3206,10 @@ STAGE PLANS: aggregations: avg(l_quantity) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: struct<count:bigint,sum:double,input:double>) Execution mode: llap LLAP IO: no inputs @@ -3220,15 +3220,15 @@ STAGE PLANS: aggregations: avg(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(), count(_col0) mode: complete outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint), _col1 (type: bigint) Reducer 12 Execution mode: llap @@ -3237,17 +3237,17 @@ STAGE PLANS: aggregations: avg(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: _col0 (type: double) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: double) sort order: + Map-reduce partition columns: _col0 (type: double) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Reducer 13 Execution mode: llap Reduce Operator Tree: @@ -3255,16 +3255,16 @@ STAGE PLANS: keys: KEY._col0 (type: double) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: double), true (type: boolean) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: double) sort order: + Map-reduce partition columns: _col0 (type: double) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: boolean) Reducer 2 Execution mode: llap @@ -3328,12 +3328,12 @@ STAGE PLANS: 0 1 outputColumnNames: _col0, _col1, _col2, _col4, _col5 - Statistics: Num rows: 25 Data size: 3424 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 5224 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col2 (type: double) sort order: + Map-reduce partition columns: _col2 (type: double) - Statistics: Num rows: 25 Data size: 3424 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 5224 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: int), _col1 (type: int), _col4 (type: bigint), _col5 (type: bigint) Reducer 8 Execution mode: llap @@ -3345,24 +3345,24 @@ STAGE PLANS: 0 _col2 (type: double) 1 _col0 (type: double) outputColumnNames: _col0, _col1, _col2, _col4, _col5, _col7 - Statistics: Num rows: 27 Data size: 3766 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 5746 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (not CASE WHEN ((_col4 = 0)) THEN (false) WHEN (_col7 is not null) THEN (true) WHEN (_col2 is null) THEN (null) WHEN ((_col5 < _col4)) THEN (true) ELSE (false) END) (type: boolean) - Statistics: Num rows: 14 Data size: 1952 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 14 Data size: 2979 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), _col1 (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 14 Data size: 1952 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 14 Data size: 2979 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: _col0 (type: int), _col1 (type: int) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 14 Data size: 1952 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 14 Data size: 2979 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: int) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 14 Data size: 1952 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 14 Data size: 2979 Basic stats: COMPLETE Column stats: NONE Stage: Stage-0 Fetch Operator http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/llap/subquery_scalar.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/llap/subquery_scalar.q.out b/ql/src/test/results/clientpositive/llap/subquery_scalar.q.out index 9fa2bd8..e78aadd 100644 --- a/ql/src/test/results/clientpositive/llap/subquery_scalar.q.out +++ b/ql/src/test/results/clientpositive/llap/subquery_scalar.q.out @@ -148,10 +148,10 @@ STAGE PLANS: aggregations: avg(p_size) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: struct<count:bigint,sum:double,input:int>) Execution mode: llap LLAP IO: no inputs @@ -180,17 +180,17 @@ STAGE PLANS: 0 1 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col10 - Statistics: Num rows: 26 Data size: 3615 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 5383 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (UDFToDouble(_col5) > _col10) (type: boolean) - Statistics: Num rows: 8 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 8 Data size: 1656 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 8 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 8 Data size: 1656 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 8 Data size: 1112 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 8 Data size: 1656 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -225,10 +225,10 @@ STAGE PLANS: aggregations: avg(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: double) Stage: Stage-0 @@ -801,10 +801,10 @@ STAGE PLANS: aggregations: max(p_name) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string) Execution mode: llap LLAP IO: no inputs @@ -820,10 +820,10 @@ STAGE PLANS: 1 2 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 5565 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 8165 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 26 Data size: 5565 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 8165 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -858,15 +858,15 @@ STAGE PLANS: aggregations: max(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: _col0 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Select Operator - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Stage: Stage-0 Fetch Operator @@ -986,10 +986,10 @@ STAGE PLANS: aggregations: avg(p_size) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: struct<count:bigint,sum:double,input:int>) Execution mode: llap LLAP IO: no inputs @@ -1063,10 +1063,10 @@ STAGE PLANS: aggregations: avg(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: double) Reducer 2 Execution mode: llap @@ -1123,17 +1123,17 @@ STAGE PLANS: 0 1 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col10, _col12 - Statistics: Num rows: 26 Data size: 3979 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 5747 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: _col5 BETWEEN _col10 AND _col12 (type: boolean) - Statistics: Num rows: 2 Data size: 306 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 442 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 2 Data size: 306 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 442 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 2 Data size: 306 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 442 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -2501,10 +2501,10 @@ STAGE PLANS: aggregations: min(p_brand) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string) Execution mode: llap LLAP IO: no inputs @@ -2540,15 +2540,15 @@ STAGE PLANS: 1 2 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col10 - Statistics: Num rows: 26 Data size: 5565 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 8165 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (_col3 <> _col10) (type: boolean) - Statistics: Num rows: 26 Data size: 5565 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 8165 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col4 (type: string), _col5 (type: int) sort order: ++ Map-reduce partition columns: _col4 (type: string), _col5 (type: int) - Statistics: Num rows: 26 Data size: 5565 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 8165 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col6 (type: string), _col7 (type: double), _col8 (type: string) Reducer 3 Execution mode: llap @@ -2560,10 +2560,10 @@ STAGE PLANS: 0 _col4 (type: string), _col5 (type: int) 1 _col1 (type: string), _col0 (type: int) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 28 Data size: 6121 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 28 Data size: 8981 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 28 Data size: 6121 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 28 Data size: 8981 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -2598,10 +2598,10 @@ STAGE PLANS: aggregations: min(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string) Reducer 9 Execution mode: llap @@ -2924,10 +2924,10 @@ STAGE PLANS: aggregations: max(value) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string) Execution mode: llap LLAP IO: no inputs @@ -2938,10 +2938,10 @@ STAGE PLANS: aggregations: max(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string) Reducer 12 Execution mode: llap @@ -3020,25 +3020,25 @@ STAGE PLANS: 0 1 outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 500 Data size: 52312 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 500 Data size: 102312 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (_col1 <> _col3) (type: boolean) - Statistics: Num rows: 500 Data size: 52312 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 500 Data size: 102312 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string) outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 52312 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 500 Data size: 102312 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() keys: _col0 (type: string) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 52312 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 500 Data size: 102312 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 52312 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 500 Data size: 102312 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: bigint) Reducer 4 Execution mode: llap @@ -3048,10 +3048,10 @@ STAGE PLANS: keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 26156 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 250 Data size: 51156 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 250 Data size: 26156 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 250 Data size: 51156 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: bigint) Reducer 5 Execution mode: llap @@ -3063,10 +3063,10 @@ STAGE PLANS: 0 1 outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 28406 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 250 Data size: 53406 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 250 Data size: 28406 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 250 Data size: 53406 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: bigint) Reducer 6 Execution mode: llap @@ -3078,17 +3078,17 @@ STAGE PLANS: 0 1 outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 31250 Data size: 3914000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 31250 Data size: 7039000 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (_col1 > _col3) (type: boolean) - Statistics: Num rows: 10416 Data size: 1304583 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 10416 Data size: 2346183 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 - Statistics: Num rows: 10416 Data size: 1304583 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 10416 Data size: 2346183 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10416 Data size: 1304583 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 10416 Data size: 2346183 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -4055,10 +4055,10 @@ STAGE PLANS: aggregations: min(p_name) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string) Execution mode: llap LLAP IO: no inputs @@ -4117,17 +4117,17 @@ STAGE PLANS: 0 1 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col10 - Statistics: Num rows: 5 Data size: 3726 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 5 Data size: 4226 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (not (_col1 like _col10)) (type: boolean) - Statistics: Num rows: 3 Data size: 2235 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 2535 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 3 Data size: 2235 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 2535 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 3 Data size: 2235 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 2535 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) Reducer 4 Execution mode: llap @@ -4139,12 +4139,12 @@ STAGE PLANS: 0 1 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col11, _col12 - Statistics: Num rows: 3 Data size: 2286 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 2586 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col3 (type: string) sort order: + Map-reduce partition columns: _col3 (type: string) - Statistics: Num rows: 3 Data size: 2286 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 2586 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col11 (type: bigint), _col12 (type: bigint) Reducer 5 Execution mode: llap @@ -4201,10 +4201,10 @@ STAGE PLANS: aggregations: min(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string) Stage: Stage-0
