http://git-wip-us.apache.org/repos/asf/hive/blob/1321293e/ql/src/test/results/clientpositive/llap/column_table_stats.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/column_table_stats.q.out
b/ql/src/test/results/clientpositive/llap/column_table_stats.q.out
new file mode 100644
index 0000000..b736477
--- /dev/null
+++ b/ql/src/test/results/clientpositive/llap/column_table_stats.q.out
@@ -0,0 +1,1421 @@
+PREHOOK: query: DROP TABLE IF EXISTS s
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE IF EXISTS s
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE s (key STRING COMMENT 'default', value STRING
COMMENT 'default') STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@s
+POSTHOOK: query: CREATE TABLE s (key STRING COMMENT 'default', value STRING
COMMENT 'default') STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@s
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE s
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@s
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE s
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@s
+PREHOOK: query: desc formatted s
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@s
+POSTHOOK: query: desc formatted s
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@s
+# col_name data_type comment
+
+key string default
+value string default
+
+# Detailed Table Information
+Database: default
+#### A masked pattern was here ####
+Retention: 0
+#### A masked pattern was here ####
+Table Type: MANAGED_TABLE
+Table Parameters:
+ numFiles 1
+ numRows 0
+ rawDataSize 0
+ totalSize 5812
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: explain extended analyze table s compute statistics for columns
+PREHOOK: type: QUERY
+POSTHOOK: query: explain extended analyze table s compute statistics for
columns
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-0 is a root stage
+ Stage-2 depends on stages: Stage-0
+ Stage-3 depends on stages: Stage-2
+
+STAGE PLANS:
+ Stage: Stage-0
+ Tez
+#### A masked pattern was here ####
+ Edges:
+ Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: s
+ Statistics: Num rows: 29 Data size: 5812 Basic stats:
COMPLETE Column stats: NONE
+ Statistics Aggregation Key Prefix: default.s/
+ GatherStats: true
+ Select Operator
+ expressions: key (type: string), value (type: string)
+ outputColumnNames: key, value
+ Statistics: Num rows: 29 Data size: 5812 Basic stats:
COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: compute_stats(key, 16),
compute_stats(value, 16)
+ mode: hash
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 1 Data size: 984 Basic stats:
COMPLETE Column stats: NONE
+ Reduce Output Operator
+ null sort order:
+ sort order:
+ Statistics: Num rows: 1 Data size: 984 Basic stats:
COMPLETE Column stats: NONE
+ tag: -1
+ value expressions: _col0 (type:
struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>),
_col1 (type:
struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>)
+ auto parallelism: false
+ Execution mode: llap
+ LLAP IO: no inputs
+ Path -> Alias:
+#### A masked pattern was here ####
+ Path -> Partition:
+#### A masked pattern was here ####
+ Partition
+ base file name: s
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ column.name.delimiter ,
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.s
+ numFiles 1
+ numRows 0
+ rawDataSize 0
+ serialization.ddl struct s { string key, string value}
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ column.name.delimiter ,
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.s
+ numFiles 1
+ numRows 0
+ rawDataSize 0
+ serialization.ddl struct s { string key, string value}
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.s
+ name: default.s
+ Truncated Path -> Alias:
+ /s [s]
+ Reducer 2
+ Execution mode: llap
+ Needs Tagging: false
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: compute_stats(VALUE._col0),
compute_stats(VALUE._col1)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 1 Data size: 984 Basic stats: COMPLETE
Column stats: NONE
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+#### A masked pattern was here ####
+ NumFilesPerFileSink: 1
+ Statistics: Num rows: 1 Data size: 984 Basic stats: COMPLETE
Column stats: NONE
+#### A masked pattern was here ####
+ table:
+ input format:
org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ properties:
+ columns _col0,_col1
+ columns.types
struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>:struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>
+ escape.delim \
+ hive.serialization.extend.additional.nesting.levels
true
+ serialization.escape.crlf true
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
+
+ Stage: Stage-2
+ Stats-Aggr Operator
+ Stats Aggregation Key Prefix: default.s/
+
+ Stage: Stage-3
+ Column Stats Work
+ Column Stats Desc:
+ Columns: key, value
+ Column Types: string, string
+ Table: default.s
+ Is Table Level Stats: true
+
+PREHOOK: query: analyze table s compute statistics for columns
+PREHOOK: type: QUERY
+PREHOOK: Input: default@s
+PREHOOK: Output: default@s
+#### A masked pattern was here ####
+POSTHOOK: query: analyze table s compute statistics for columns
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@s
+POSTHOOK: Output: default@s
+#### A masked pattern was here ####
+PREHOOK: query: desc formatted s
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@s
+POSTHOOK: query: desc formatted s
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@s
+# col_name data_type comment
+
+key string default
+value string default
+
+# Detailed Table Information
+Database: default
+#### A masked pattern was here ####
+Retention: 0
+#### A masked pattern was here ####
+Table Type: MANAGED_TABLE
+Table Parameters:
+ COLUMN_STATS_ACCURATE
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+ numFiles 1
+ numRows 500
+ rawDataSize 5312
+ totalSize 5812
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: DROP TABLE IF EXISTS spart
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: DROP TABLE IF EXISTS spart
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE spart (key STRING COMMENT 'default', value STRING
COMMENT 'default')
+PARTITIONED BY (ds STRING, hr STRING)
+STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@spart
+POSTHOOK: query: CREATE TABLE spart (key STRING COMMENT 'default', value
STRING COMMENT 'default')
+PARTITIONED BY (ds STRING, hr STRING)
+STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@spart
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/kv1.txt"
+OVERWRITE INTO TABLE spart PARTITION (ds="2008-04-08", hr="11")
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@spart
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/kv1.txt"
+OVERWRITE INTO TABLE spart PARTITION (ds="2008-04-08", hr="11")
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@spart
+POSTHOOK: Output: default@spart@ds=2008-04-08/hr=11
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/kv1.txt"
+OVERWRITE INTO TABLE spart PARTITION (ds="2008-04-08", hr="12")
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@spart
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/kv1.txt"
+OVERWRITE INTO TABLE spart PARTITION (ds="2008-04-08", hr="12")
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@spart
+POSTHOOK: Output: default@spart@ds=2008-04-08/hr=12
+PREHOOK: query: desc formatted spart
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@spart
+POSTHOOK: query: desc formatted spart
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@spart
+# col_name data_type comment
+
+key string default
+value string default
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+hr string
+
+# Detailed Table Information
+Database: default
+#### A masked pattern was here ####
+Retention: 0
+#### A masked pattern was here ####
+Table Type: MANAGED_TABLE
+Table Parameters:
+ numFiles 2
+ numRows 0
+ rawDataSize 0
+ totalSize 11624
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: explain extended analyze table spart compute statistics for
columns
+PREHOOK: type: QUERY
+POSTHOOK: query: explain extended analyze table spart compute statistics for
columns
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-0 is a root stage
+ Stage-2 depends on stages: Stage-0
+ Stage-3 depends on stages: Stage-2
+
+STAGE PLANS:
+ Stage: Stage-0
+ Tez
+#### A masked pattern was here ####
+ Edges:
+ Reducer 2 <- Map 1 (SIMPLE_EDGE)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: spart
+ Statistics: Num rows: 58 Data size: 32968 Basic stats:
COMPLETE Column stats: PARTIAL
+ Statistics Aggregation Key Prefix: default.spart/
+ GatherStats: true
+ Select Operator
+ expressions: ds (type: string), hr (type: string), key
(type: string), value (type: string)
+ outputColumnNames: ds, hr, key, value
+ Statistics: Num rows: 58 Data size: 32968 Basic stats:
COMPLETE Column stats: PARTIAL
+ Group By Operator
+ aggregations: compute_stats(key, 16),
compute_stats(value, 16)
+ keys: ds (type: string), hr (type: string)
+ mode: hash
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 2 Data size: 2704 Basic stats:
COMPLETE Column stats: PARTIAL
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col1 (type:
string)
+ null sort order: aa
+ sort order: ++
+ Map-reduce partition columns: _col0 (type: string),
_col1 (type: string)
+ Statistics: Num rows: 2 Data size: 2704 Basic stats:
COMPLETE Column stats: PARTIAL
+ tag: -1
+ value expressions: _col2 (type:
struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>),
_col3 (type:
struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>)
+ auto parallelism: true
+ Execution mode: llap
+ LLAP IO: no inputs
+ Path -> Alias:
+#### A masked pattern was here ####
+ Path -> Partition:
+#### A masked pattern was here ####
+ Partition
+ base file name: hr=11
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ partition values:
+ ds 2008-04-08
+ hr 11
+ properties:
+ bucket_count -1
+ column.name.delimiter ,
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.spart
+ numFiles 1
+ numRows 0
+ partition_columns ds/hr
+ partition_columns.types string:string
+ rawDataSize 0
+ serialization.ddl struct spart { string key, string value}
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ column.name.delimiter ,
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.spart
+ partition_columns ds/hr
+ partition_columns.types string:string
+ serialization.ddl struct spart { string key, string
value}
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.spart
+ name: default.spart
+#### A masked pattern was here ####
+ Partition
+ base file name: hr=12
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ partition values:
+ ds 2008-04-08
+ hr 12
+ properties:
+ bucket_count -1
+ column.name.delimiter ,
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.spart
+ numFiles 1
+ numRows 0
+ partition_columns ds/hr
+ partition_columns.types string:string
+ rawDataSize 0
+ serialization.ddl struct spart { string key, string value}
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ column.name.delimiter ,
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.spart
+ partition_columns ds/hr
+ partition_columns.types string:string
+ serialization.ddl struct spart { string key, string
value}
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.spart
+ name: default.spart
+ Truncated Path -> Alias:
+ /spart/ds=2008-04-08/hr=11 [spart]
+ /spart/ds=2008-04-08/hr=12 [spart]
+ Reducer 2
+ Execution mode: llap
+ Needs Tagging: false
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: compute_stats(VALUE._col0),
compute_stats(VALUE._col1)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 2 Data size: 2656 Basic stats: COMPLETE
Column stats: PARTIAL
+ Select Operator
+ expressions: _col2 (type:
struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>),
_col3 (type:
struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>),
_col0 (type: string), _col1 (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 2 Data size: 2656 Basic stats:
COMPLETE Column stats: PARTIAL
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+#### A masked pattern was here ####
+ NumFilesPerFileSink: 1
+ Statistics: Num rows: 2 Data size: 2656 Basic stats:
COMPLETE Column stats: PARTIAL
+#### A masked pattern was here ####
+ table:
+ input format:
org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ properties:
+ columns _col0,_col1,_col2,_col3
+ columns.types
struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>:struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>:string:string
+ escape.delim \
+ hive.serialization.extend.additional.nesting.levels
true
+ serialization.escape.crlf true
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ serde:
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
+
+ Stage: Stage-2
+ Stats-Aggr Operator
+ Stats Aggregation Key Prefix: default.spart/
+
+ Stage: Stage-3
+ Column Stats Work
+ Column Stats Desc:
+ Columns: key, value
+ Column Types: string, string
+ Table: default.spart
+ Is Table Level Stats: false
+
+PREHOOK: query: analyze table spart compute statistics for columns
+PREHOOK: type: QUERY
+PREHOOK: Input: default@spart
+PREHOOK: Input: default@spart@ds=2008-04-08/hr=11
+PREHOOK: Input: default@spart@ds=2008-04-08/hr=12
+PREHOOK: Output: default@spart
+PREHOOK: Output: default@spart@ds=2008-04-08/hr=11
+PREHOOK: Output: default@spart@ds=2008-04-08/hr=12
+#### A masked pattern was here ####
+POSTHOOK: query: analyze table spart compute statistics for columns
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@spart
+POSTHOOK: Input: default@spart@ds=2008-04-08/hr=11
+POSTHOOK: Input: default@spart@ds=2008-04-08/hr=12
+POSTHOOK: Output: default@spart
+POSTHOOK: Output: default@spart@ds=2008-04-08/hr=11
+POSTHOOK: Output: default@spart@ds=2008-04-08/hr=12
+#### A masked pattern was here ####
+PREHOOK: query: desc formatted spart
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@spart
+POSTHOOK: query: desc formatted spart
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@spart
+# col_name data_type comment
+
+key string default
+value string default
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+hr string
+
+# Detailed Table Information
+Database: default
+#### A masked pattern was here ####
+Retention: 0
+#### A masked pattern was here ####
+Table Type: MANAGED_TABLE
+Table Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 2
+ numRows 1000
+ rawDataSize 10624
+ totalSize 11624
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: desc formatted spart PARTITION(ds='2008-04-08', hr=11)
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@spart
+POSTHOOK: query: desc formatted spart PARTITION(ds='2008-04-08', hr=11)
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@spart
+# col_name data_type comment
+
+key string default
+value string default
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+hr string
+
+# Detailed Partition Information
+Partition Value: [2008-04-08, 11]
+Database: default
+Table: spart
+#### A masked pattern was here ####
+Partition Parameters:
+ COLUMN_STATS_ACCURATE
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+ numFiles 1
+ numRows 500
+ rawDataSize 5312
+ totalSize 5812
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: desc formatted spart PARTITION(ds='2008-04-08', hr=12)
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@spart
+POSTHOOK: query: desc formatted spart PARTITION(ds='2008-04-08', hr=12)
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@spart
+# col_name data_type comment
+
+key string default
+value string default
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+hr string
+
+# Detailed Partition Information
+Partition Value: [2008-04-08, 12]
+Database: default
+Table: spart
+#### A masked pattern was here ####
+Partition Parameters:
+ COLUMN_STATS_ACCURATE
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+ numFiles 1
+ numRows 500
+ rawDataSize 5312
+ totalSize 5812
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: DROP TABLE IF EXISTS spart
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@spart
+PREHOOK: Output: default@spart
+POSTHOOK: query: DROP TABLE IF EXISTS spart
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@spart
+POSTHOOK: Output: default@spart
+PREHOOK: query: CREATE TABLE spart (key STRING COMMENT 'default', value STRING
COMMENT 'default')
+PARTITIONED BY (ds STRING, hr STRING)
+STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@spart
+POSTHOOK: query: CREATE TABLE spart (key STRING COMMENT 'default', value
STRING COMMENT 'default')
+PARTITIONED BY (ds STRING, hr STRING)
+STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@spart
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/kv1.txt"
+OVERWRITE INTO TABLE spart PARTITION (ds="2008-04-08", hr="11")
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@spart
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/kv1.txt"
+OVERWRITE INTO TABLE spart PARTITION (ds="2008-04-08", hr="11")
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@spart
+POSTHOOK: Output: default@spart@ds=2008-04-08/hr=11
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/kv1.txt"
+OVERWRITE INTO TABLE spart PARTITION (ds="2008-04-08", hr="12")
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@spart
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/kv1.txt"
+OVERWRITE INTO TABLE spart PARTITION (ds="2008-04-08", hr="12")
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@spart
+POSTHOOK: Output: default@spart@ds=2008-04-08/hr=12
+PREHOOK: query: desc formatted spart
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@spart
+POSTHOOK: query: desc formatted spart
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@spart
+# col_name data_type comment
+
+key string default
+value string default
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+hr string
+
+# Detailed Table Information
+Database: default
+#### A masked pattern was here ####
+Retention: 0
+#### A masked pattern was here ####
+Table Type: MANAGED_TABLE
+Table Parameters:
+ numFiles 2
+ numRows 0
+ rawDataSize 0
+ totalSize 11624
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: explain extended analyze table spart partition(ds,hr) compute
statistics for columns
+PREHOOK: type: QUERY
+POSTHOOK: query: explain extended analyze table spart partition(ds,hr) compute
statistics for columns
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-0 is a root stage
+ Stage-2 depends on stages: Stage-0
+ Stage-3 depends on stages: Stage-2
+
+STAGE PLANS:
+ Stage: Stage-0
+ Tez
+#### A masked pattern was here ####
+ Edges:
+ Reducer 2 <- Map 1 (SIMPLE_EDGE)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: spart
+ Statistics: Num rows: 58 Data size: 32968 Basic stats:
COMPLETE Column stats: PARTIAL
+ Statistics Aggregation Key Prefix: default.spart/
+ GatherStats: true
+ Select Operator
+ expressions: ds (type: string), hr (type: string), key
(type: string), value (type: string)
+ outputColumnNames: ds, hr, key, value
+ Statistics: Num rows: 58 Data size: 32968 Basic stats:
COMPLETE Column stats: PARTIAL
+ Group By Operator
+ aggregations: compute_stats(key, 16),
compute_stats(value, 16)
+ keys: ds (type: string), hr (type: string)
+ mode: hash
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 2 Data size: 2704 Basic stats:
COMPLETE Column stats: PARTIAL
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col1 (type:
string)
+ null sort order: aa
+ sort order: ++
+ Map-reduce partition columns: _col0 (type: string),
_col1 (type: string)
+ Statistics: Num rows: 2 Data size: 2704 Basic stats:
COMPLETE Column stats: PARTIAL
+ tag: -1
+ value expressions: _col2 (type:
struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>),
_col3 (type:
struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>)
+ auto parallelism: true
+ Execution mode: llap
+ LLAP IO: no inputs
+ Path -> Alias:
+#### A masked pattern was here ####
+ Path -> Partition:
+#### A masked pattern was here ####
+ Partition
+ base file name: hr=11
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ partition values:
+ ds 2008-04-08
+ hr 11
+ properties:
+ bucket_count -1
+ column.name.delimiter ,
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.spart
+ numFiles 1
+ numRows 0
+ partition_columns ds/hr
+ partition_columns.types string:string
+ rawDataSize 0
+ serialization.ddl struct spart { string key, string value}
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ column.name.delimiter ,
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.spart
+ partition_columns ds/hr
+ partition_columns.types string:string
+ serialization.ddl struct spart { string key, string
value}
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.spart
+ name: default.spart
+#### A masked pattern was here ####
+ Partition
+ base file name: hr=12
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ partition values:
+ ds 2008-04-08
+ hr 12
+ properties:
+ bucket_count -1
+ column.name.delimiter ,
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.spart
+ numFiles 1
+ numRows 0
+ partition_columns ds/hr
+ partition_columns.types string:string
+ rawDataSize 0
+ serialization.ddl struct spart { string key, string value}
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ column.name.delimiter ,
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.spart
+ partition_columns ds/hr
+ partition_columns.types string:string
+ serialization.ddl struct spart { string key, string
value}
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.spart
+ name: default.spart
+ Truncated Path -> Alias:
+ /spart/ds=2008-04-08/hr=11 [spart]
+ /spart/ds=2008-04-08/hr=12 [spart]
+ Reducer 2
+ Execution mode: llap
+ Needs Tagging: false
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: compute_stats(VALUE._col0),
compute_stats(VALUE._col1)
+ keys: KEY._col0 (type: string), KEY._col1 (type: string)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 2 Data size: 2656 Basic stats: COMPLETE
Column stats: PARTIAL
+ Select Operator
+ expressions: _col2 (type:
struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>),
_col3 (type:
struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>),
_col0 (type: string), _col1 (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 2 Data size: 2656 Basic stats:
COMPLETE Column stats: PARTIAL
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+#### A masked pattern was here ####
+ NumFilesPerFileSink: 1
+ Statistics: Num rows: 2 Data size: 2656 Basic stats:
COMPLETE Column stats: PARTIAL
+#### A masked pattern was here ####
+ table:
+ input format:
org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ properties:
+ columns _col0,_col1,_col2,_col3
+ columns.types
struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>:struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>:string:string
+ escape.delim \
+ hive.serialization.extend.additional.nesting.levels
true
+ serialization.escape.crlf true
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ serde:
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
+
+ Stage: Stage-2
+ Stats-Aggr Operator
+ Stats Aggregation Key Prefix: default.spart/
+
+ Stage: Stage-3
+ Column Stats Work
+ Column Stats Desc:
+ Columns: key, value
+ Column Types: string, string
+ Table: default.spart
+ Is Table Level Stats: false
+
+PREHOOK: query: analyze table spart partition(ds,hr) compute statistics for
columns
+PREHOOK: type: QUERY
+PREHOOK: Input: default@spart
+PREHOOK: Input: default@spart@ds=2008-04-08/hr=11
+PREHOOK: Input: default@spart@ds=2008-04-08/hr=12
+PREHOOK: Output: default@spart
+PREHOOK: Output: default@spart@ds=2008-04-08/hr=11
+PREHOOK: Output: default@spart@ds=2008-04-08/hr=12
+#### A masked pattern was here ####
+POSTHOOK: query: analyze table spart partition(ds,hr) compute statistics for
columns
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@spart
+POSTHOOK: Input: default@spart@ds=2008-04-08/hr=11
+POSTHOOK: Input: default@spart@ds=2008-04-08/hr=12
+POSTHOOK: Output: default@spart
+POSTHOOK: Output: default@spart@ds=2008-04-08/hr=11
+POSTHOOK: Output: default@spart@ds=2008-04-08/hr=12
+#### A masked pattern was here ####
+PREHOOK: query: desc formatted spart
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@spart
+POSTHOOK: query: desc formatted spart
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@spart
+# col_name data_type comment
+
+key string default
+value string default
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+hr string
+
+# Detailed Table Information
+Database: default
+#### A masked pattern was here ####
+Retention: 0
+#### A masked pattern was here ####
+Table Type: MANAGED_TABLE
+Table Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 2
+ numRows 1000
+ rawDataSize 10624
+ totalSize 11624
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: desc formatted spart PARTITION(ds='2008-04-08', hr=11)
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@spart
+POSTHOOK: query: desc formatted spart PARTITION(ds='2008-04-08', hr=11)
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@spart
+# col_name data_type comment
+
+key string default
+value string default
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+hr string
+
+# Detailed Partition Information
+Partition Value: [2008-04-08, 11]
+Database: default
+Table: spart
+#### A masked pattern was here ####
+Partition Parameters:
+ COLUMN_STATS_ACCURATE
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+ numFiles 1
+ numRows 500
+ rawDataSize 5312
+ totalSize 5812
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: desc formatted spart PARTITION(ds='2008-04-08', hr=12)
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@spart
+POSTHOOK: query: desc formatted spart PARTITION(ds='2008-04-08', hr=12)
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@spart
+# col_name data_type comment
+
+key string default
+value string default
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+hr string
+
+# Detailed Partition Information
+Partition Value: [2008-04-08, 12]
+Database: default
+Table: spart
+#### A masked pattern was here ####
+Partition Parameters:
+ COLUMN_STATS_ACCURATE
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+ numFiles 1
+ numRows 500
+ rawDataSize 5312
+ totalSize 5812
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: DROP TABLE IF EXISTS spart
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@spart
+PREHOOK: Output: default@spart
+POSTHOOK: query: DROP TABLE IF EXISTS spart
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@spart
+POSTHOOK: Output: default@spart
+PREHOOK: query: CREATE TABLE spart (key STRING COMMENT 'default', value STRING
COMMENT 'default')
+PARTITIONED BY (ds STRING, hr STRING)
+STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@spart
+POSTHOOK: query: CREATE TABLE spart (key STRING COMMENT 'default', value
STRING COMMENT 'default')
+PARTITIONED BY (ds STRING, hr STRING)
+STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@spart
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/kv1.txt"
+OVERWRITE INTO TABLE spart PARTITION (ds="2008-04-08", hr="11")
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@spart
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/kv1.txt"
+OVERWRITE INTO TABLE spart PARTITION (ds="2008-04-08", hr="11")
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@spart
+POSTHOOK: Output: default@spart@ds=2008-04-08/hr=11
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/kv1.txt"
+OVERWRITE INTO TABLE spart PARTITION (ds="2008-04-08", hr="12")
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@spart
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/kv1.txt"
+OVERWRITE INTO TABLE spart PARTITION (ds="2008-04-08", hr="12")
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@spart
+POSTHOOK: Output: default@spart@ds=2008-04-08/hr=12
+PREHOOK: query: desc formatted spart
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@spart
+POSTHOOK: query: desc formatted spart
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@spart
+# col_name data_type comment
+
+key string default
+value string default
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+hr string
+
+# Detailed Table Information
+Database: default
+#### A masked pattern was here ####
+Retention: 0
+#### A masked pattern was here ####
+Table Type: MANAGED_TABLE
+Table Parameters:
+ numFiles 2
+ numRows 0
+ rawDataSize 0
+ totalSize 11624
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: explain extended analyze table spart partition(hr="11")
compute statistics for columns
+PREHOOK: type: QUERY
+POSTHOOK: query: explain extended analyze table spart partition(hr="11")
compute statistics for columns
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+ Stage-0 is a root stage
+ Stage-2 depends on stages: Stage-0
+ Stage-3 depends on stages: Stage-2
+
+STAGE PLANS:
+ Stage: Stage-0
+ Tez
+#### A masked pattern was here ####
+ Edges:
+ Reducer 2 <- Map 1 (SIMPLE_EDGE)
+#### A masked pattern was here ####
+ Vertices:
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: spart
+ Statistics: Num rows: 29 Data size: 11148 Basic stats:
COMPLETE Column stats: PARTIAL
+ Statistics Aggregation Key Prefix: default.spart/
+ GatherStats: true
+ Select Operator
+ expressions: ds (type: string), key (type: string), value
(type: string)
+ outputColumnNames: ds, key, value
+ Statistics: Num rows: 29 Data size: 11148 Basic stats:
COMPLETE Column stats: PARTIAL
+ Group By Operator
+ aggregations: compute_stats(key, 16),
compute_stats(value, 16)
+ keys: ds (type: string), '11' (type: string)
+ mode: hash
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 1 Data size: 1254 Basic stats:
COMPLETE Column stats: PARTIAL
+ Reduce Output Operator
+ key expressions: _col0 (type: string), '11' (type:
string)
+ null sort order: aa
+ sort order: ++
+ Map-reduce partition columns: _col0 (type: string),
'11' (type: string)
+ Statistics: Num rows: 1 Data size: 1254 Basic stats:
COMPLETE Column stats: PARTIAL
+ tag: -1
+ value expressions: _col2 (type:
struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>),
_col3 (type:
struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:string,numbitvectors:int>)
+ auto parallelism: true
+ Execution mode: llap
+ LLAP IO: no inputs
+ Path -> Alias:
+#### A masked pattern was here ####
+ Path -> Partition:
+#### A masked pattern was here ####
+ Partition
+ base file name: hr=11
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ partition values:
+ ds 2008-04-08
+ hr 11
+ properties:
+ bucket_count -1
+ column.name.delimiter ,
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.spart
+ numFiles 1
+ numRows 0
+ partition_columns ds/hr
+ partition_columns.types string:string
+ rawDataSize 0
+ serialization.ddl struct spart { string key, string value}
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 5812
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ properties:
+ bucket_count -1
+ column.name.delimiter ,
+ columns key,value
+ columns.comments 'default','default'
+ columns.types string:string
+#### A masked pattern was here ####
+ name default.spart
+ partition_columns ds/hr
+ partition_columns.types string:string
+ serialization.ddl struct spart { string key, string
value}
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+#### A masked pattern was here ####
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.spart
+ name: default.spart
+ Truncated Path -> Alias:
+ /spart/ds=2008-04-08/hr=11 [spart]
+ Reducer 2
+ Execution mode: llap
+ Needs Tagging: false
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: compute_stats(VALUE._col0),
compute_stats(VALUE._col1)
+ keys: KEY._col0 (type: string), '11' (type: string)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 1 Data size: 1230 Basic stats: COMPLETE
Column stats: PARTIAL
+ Select Operator
+ expressions: _col2 (type:
struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>),
_col3 (type:
struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>),
_col0 (type: string), '11' (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 1 Data size: 1230 Basic stats:
COMPLETE Column stats: PARTIAL
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+#### A masked pattern was here ####
+ NumFilesPerFileSink: 1
+ Statistics: Num rows: 1 Data size: 1230 Basic stats:
COMPLETE Column stats: PARTIAL
+#### A masked pattern was here ####
+ table:
+ input format:
org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format:
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ properties:
+ columns _col0,_col1,_col2,_col3
+ columns.types
struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>:struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:string>:string:string
+ escape.delim \
+ hive.serialization.extend.additional.nesting.levels
true
+ serialization.escape.crlf true
+ serialization.format 1
+ serialization.lib
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ serde:
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
+
+ Stage: Stage-2
+ Stats-Aggr Operator
+ Stats Aggregation Key Prefix: default.spart/
+
+ Stage: Stage-3
+ Column Stats Work
+ Column Stats Desc:
+ Columns: key, value
+ Column Types: string, string
+ Table: default.spart
+ Is Table Level Stats: false
+
+PREHOOK: query: analyze table spart partition(hr="11") compute statistics for
columns
+PREHOOK: type: QUERY
+PREHOOK: Input: default@spart
+PREHOOK: Input: default@spart@ds=2008-04-08/hr=11
+PREHOOK: Output: default@spart
+PREHOOK: Output: default@spart@ds=2008-04-08/hr=11
+#### A masked pattern was here ####
+POSTHOOK: query: analyze table spart partition(hr="11") compute statistics for
columns
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@spart
+POSTHOOK: Input: default@spart@ds=2008-04-08/hr=11
+POSTHOOK: Output: default@spart
+POSTHOOK: Output: default@spart@ds=2008-04-08/hr=11
+#### A masked pattern was here ####
+PREHOOK: query: desc formatted spart
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@spart
+POSTHOOK: query: desc formatted spart
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@spart
+# col_name data_type comment
+
+key string default
+value string default
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+hr string
+
+# Detailed Table Information
+Database: default
+#### A masked pattern was here ####
+Retention: 0
+#### A masked pattern was here ####
+Table Type: MANAGED_TABLE
+Table Parameters:
+ numFiles 2
+ numRows 500
+ rawDataSize 5312
+ totalSize 11624
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: desc formatted spart PARTITION(ds='2008-04-08', hr=11)
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@spart
+POSTHOOK: query: desc formatted spart PARTITION(ds='2008-04-08', hr=11)
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@spart
+# col_name data_type comment
+
+key string default
+value string default
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+hr string
+
+# Detailed Partition Information
+Partition Value: [2008-04-08, 11]
+Database: default
+Table: spart
+#### A masked pattern was here ####
+Partition Parameters:
+ COLUMN_STATS_ACCURATE
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+ numFiles 1
+ numRows 500
+ rawDataSize 5312
+ totalSize 5812
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
+PREHOOK: query: desc formatted spart PARTITION(ds='2008-04-08', hr=12)
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@spart
+POSTHOOK: query: desc formatted spart PARTITION(ds='2008-04-08', hr=12)
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@spart
+# col_name data_type comment
+
+key string default
+value string default
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+hr string
+
+# Detailed Partition Information
+Partition Value: [2008-04-08, 12]
+Database: default
+Table: spart
+#### A masked pattern was here ####
+Partition Parameters:
+ numFiles 1
+ numRows 0
+ rawDataSize 0
+ totalSize 5812
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1