Repository: hive Updated Branches: refs/heads/branch-3 26c5749eb -> ba54eb5de
http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out_spark ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out_spark b/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out_spark index 63752f3..dbb2460 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out_spark +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out_spark @@ -78,12 +78,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 69 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 69 Data size: 27500 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 69 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 69 Data size: 27500 Basic stats: PARTIAL Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: int) @@ -162,12 +162,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 69 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 69 Data size: 27500 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 69 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 69 Data size: 27500 Basic stats: PARTIAL Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -178,17 +178,17 @@ STAGE PLANS: input vertices: 1 Map 3 Position of Big Table: 0 - Statistics: Num rows: 75 Data size: 30250 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 75 Data size: 30250 Basic stats: PARTIAL Column stats: NONE BucketMapJoin: true Select Operator expressions: _col0 (type: int), _col8 (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 75 Data size: 30250 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 75 Data size: 30250 Basic stats: PARTIAL Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string) null sort order: aa sort order: ++ - Statistics: Num rows: 75 Data size: 30250 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 75 Data size: 30250 Basic stats: PARTIAL Column stats: NONE tag: -1 TopN: 1 TopN Hash Memory Usage: 0.1 @@ -261,16 +261,16 @@ STAGE PLANS: Select Operator expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 75 Data size: 30250 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 75 Data size: 30250 Basic stats: PARTIAL Column stats: NONE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 403 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 403 Basic stats: PARTIAL Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 403 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 403 Basic stats: PARTIAL Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out index b22022c..ecd33c7 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out @@ -84,12 +84,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: int) @@ -167,12 +167,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -182,17 +182,17 @@ STAGE PLANS: input vertices: 1 Map 3 Position of Big Table: 0 - Statistics: Num rows: 104 Data size: 30250 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 104 Data size: 30250 Basic stats: PARTIAL Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE Reduce Output Operator null sort order: sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE tag: -1 value expressions: _col0 (type: bigint) auto parallelism: false @@ -264,13 +264,13 @@ STAGE PLANS: aggregations: count(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat @@ -345,12 +345,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: int) @@ -428,12 +428,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -443,17 +443,17 @@ STAGE PLANS: input vertices: 1 Map 3 Position of Big Table: 0 - Statistics: Num rows: 104 Data size: 30250 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 104 Data size: 30250 Basic stats: PARTIAL Column stats: NONE BucketMapJoin: true Group By Operator aggregations: count() mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE Reduce Output Operator null sort order: sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE tag: -1 value expressions: _col0 (type: bigint) auto parallelism: false @@ -525,13 +525,13 @@ STAGE PLANS: aggregations: count(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out index 7d455ba..d03b938 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out @@ -92,12 +92,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 145 Data size: 42000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 145 Data size: 42000 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 145 Data size: 42000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 145 Data size: 42000 Basic stats: PARTIAL Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: int) @@ -170,12 +170,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -185,16 +185,16 @@ STAGE PLANS: input vertices: 1 Map 3 Position of Big Table: 0 - Statistics: Num rows: 159 Data size: 46200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 159 Data size: 46200 Basic stats: PARTIAL Column stats: NONE Group By Operator aggregations: count() mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE Reduce Output Operator null sort order: sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE tag: -1 value expressions: _col0 (type: bigint) auto parallelism: false @@ -261,13 +261,13 @@ STAGE PLANS: aggregations: count(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat @@ -375,12 +375,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: int) @@ -453,12 +453,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 95 Data size: 27500 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 95 Data size: 27500 Basic stats: PARTIAL Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -468,16 +468,16 @@ STAGE PLANS: input vertices: 1 Map 3 Position of Big Table: 0 - Statistics: Num rows: 104 Data size: 30250 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 104 Data size: 30250 Basic stats: PARTIAL Column stats: NONE Group By Operator aggregations: count() mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE Reduce Output Operator null sort order: sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE tag: -1 value expressions: _col0 (type: bigint) auto parallelism: false @@ -544,13 +544,13 @@ STAGE PLANS: aggregations: count(VALUE._col0) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out index ece73fe..e53de75 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out @@ -90,12 +90,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 140 Data size: 42000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 140 Data size: 42000 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 140 Data size: 42000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 140 Data size: 42000 Basic stats: PARTIAL Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: int) @@ -183,17 +183,17 @@ STAGE PLANS: input vertices: 1 Map 2 Position of Big Table: 0 - Statistics: Num rows: 154 Data size: 46200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 154 Data size: 46200 Basic stats: PARTIAL Column stats: NONE Select Operator expressions: UDFToString(_col0) (type: string), _col1 (type: string), _col6 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 154 Data size: 46200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 154 Data size: 46200 Basic stats: PARTIAL Column stats: NONE File Output Operator compressed: false GlobalTableId: 1 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 154 Data size: 46200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 154 Data size: 46200 Basic stats: PARTIAL Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.TextInputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out index 350f860..c1fd0c1 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out @@ -99,12 +99,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 156 Data size: 61240 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 156 Data size: 61240 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: key is not null (type: boolean) - Statistics: Num rows: 156 Data size: 61240 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 156 Data size: 61240 Basic stats: PARTIAL Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: int) @@ -247,18 +247,18 @@ STAGE PLANS: input vertices: 1 Map 2 Position of Big Table: 0 - Statistics: Num rows: 171 Data size: 67364 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 171 Data size: 67364 Basic stats: PARTIAL Column stats: NONE BucketMapJoin: true Select Operator expressions: UDFToString(_col0) (type: string), _col1 (type: string), _col6 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 171 Data size: 67364 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 171 Data size: 67364 Basic stats: PARTIAL Column stats: NONE File Output Operator compressed: false GlobalTableId: 1 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 171 Data size: 67364 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 171 Data size: 67364 Basic stats: PARTIAL Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.TextInputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/smb_mapjoin_10.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/smb_mapjoin_10.q.out b/ql/src/test/results/clientpositive/spark/smb_mapjoin_10.q.out index 68d3952..c64dc41 100644 --- a/ql/src/test/results/clientpositive/spark/smb_mapjoin_10.q.out +++ b/ql/src/test/results/clientpositive/spark/smb_mapjoin_10.q.out @@ -81,10 +81,10 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 13 Data size: 4140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 13 Data size: 4140 Basic stats: PARTIAL Column stats: NONE Filter Operator predicate: (pageid is not null and postid is not null and type is not null and userid is not null) (type: boolean) - Statistics: Num rows: 13 Data size: 4140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 13 Data size: 4140 Basic stats: PARTIAL Column stats: NONE Sorted Merge Bucket Map Join Operator condition map: Inner Join 0 to 1 @@ -92,14 +92,14 @@ STAGE PLANS: 0 userid (type: int), pageid (type: int), postid (type: int), type (type: string) 1 userid (type: int), pageid (type: int), postid (type: int), type (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col8, _col9, _col10, _col11 - Statistics: Num rows: 14 Data size: 4554 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 14 Data size: 4554 Basic stats: PARTIAL Column stats: NONE Select Operator expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: string), '1' (type: string), _col8 (type: int), _col9 (type: int), _col10 (type: int), _col11 (type: string), '2' (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9 - Statistics: Num rows: 14 Data size: 4554 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 14 Data size: 4554 Basic stats: PARTIAL Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 14 Data size: 4554 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 14 Data size: 4554 Basic stats: PARTIAL Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/stats12.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/stats12.q.out b/ql/src/test/results/clientpositive/spark/stats12.q.out index c976da8..b5dac094b 100644 --- a/ql/src/test/results/clientpositive/spark/stats12.q.out +++ b/ql/src/test/results/clientpositive/spark/stats12.q.out @@ -52,7 +52,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart_n3 - Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL Column stats: NONE Statistics Aggregation Key Prefix: default.analyze_srcpart_n3/ GatherStats: true Path -> Alias: http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/stats13.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/stats13.q.out b/ql/src/test/results/clientpositive/spark/stats13.q.out index d59ca8b..074fd3f 100644 --- a/ql/src/test/results/clientpositive/spark/stats13.q.out +++ b/ql/src/test/results/clientpositive/spark/stats13.q.out @@ -52,7 +52,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart - Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL Column stats: NONE Statistics Aggregation Key Prefix: default.analyze_srcpart/ GatherStats: true Path -> Alias: http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/stats2.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/stats2.q.out b/ql/src/test/results/clientpositive/spark/stats2.q.out index c6283c4..04831f8 100644 --- a/ql/src/test/results/clientpositive/spark/stats2.q.out +++ b/ql/src/test/results/clientpositive/spark/stats2.q.out @@ -136,7 +136,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_t1 - Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL Column stats: NONE Stage: Stage-1 Stats Work http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/stats7.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/stats7.q.out b/ql/src/test/results/clientpositive/spark/stats7.q.out index f23f409..3ae274e 100644 --- a/ql/src/test/results/clientpositive/spark/stats7.q.out +++ b/ql/src/test/results/clientpositive/spark/stats7.q.out @@ -50,7 +50,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart_n4 - Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL Column stats: NONE Stage: Stage-1 Stats Work http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/stats8.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/stats8.q.out b/ql/src/test/results/clientpositive/spark/stats8.q.out index c19ff2e..767ff21 100644 --- a/ql/src/test/results/clientpositive/spark/stats8.q.out +++ b/ql/src/test/results/clientpositive/spark/stats8.q.out @@ -33,6 +33,41 @@ POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=11).key SIMPLE POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: describe formatted analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=11) +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@analyze_srcpart_n1 +POSTHOOK: query: describe formatted analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=11) +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@analyze_srcpart_n1 +# col_name data_type comment +key string default +value string default + +# Partition Information +# col_name data_type comment +ds string +hr string + +# Detailed Partition Information +Partition Value: [2008-04-08, 11] +Database: default +Table: analyze_srcpart_n1 +#### A masked pattern was here #### +Partition Parameters: + numFiles 1 + totalSize 5812 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +InputFormat: org.apache.hadoop.mapred.TextInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Storage Desc Params: + serialization.format 1 PREHOOK: query: explain analyze table analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=11) compute statistics PREHOOK: type: QUERY POSTHOOK: query: explain analyze table analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=11) compute statistics @@ -50,7 +85,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart_n1 - Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL Column stats: NONE Stage: Stage-1 Stats Work @@ -145,6 +180,41 @@ Bucket Columns: [] Sort Columns: [] Storage Desc Params: serialization.format 1 +PREHOOK: query: describe formatted analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=12) +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@analyze_srcpart_n1 +POSTHOOK: query: describe formatted analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=12) +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@analyze_srcpart_n1 +# col_name data_type comment +key string default +value string default + +# Partition Information +# col_name data_type comment +ds string +hr string + +# Detailed Partition Information +Partition Value: [2008-04-08, 12] +Database: default +Table: analyze_srcpart_n1 +#### A masked pattern was here #### +Partition Parameters: + numFiles 1 + totalSize 5812 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +InputFormat: org.apache.hadoop.mapred.TextInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Storage Desc Params: + serialization.format 1 PREHOOK: query: explain analyze table analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=12) compute statistics PREHOOK: type: QUERY POSTHOOK: query: explain analyze table analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=12) compute statistics @@ -162,7 +232,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart_n1 - Statistics: Num rows: 500 Data size: 5312 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 794 Data size: 179672 Basic stats: PARTIAL Column stats: NONE Stage: Stage-1 Stats Work @@ -235,7 +305,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart_n1 - Statistics: Num rows: 1000 Data size: 10624 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1196 Data size: 126864 Basic stats: PARTIAL Column stats: NONE Stage: Stage-1 Stats Work @@ -308,7 +378,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart_n1 - Statistics: Num rows: 1500 Data size: 15936 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1598 Data size: 74056 Basic stats: PARTIAL Column stats: NONE Stage: Stage-1 Stats Work http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/stats12.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/stats12.q.out b/ql/src/test/results/clientpositive/stats12.q.out index ec0143e..923fc78 100644 --- a/ql/src/test/results/clientpositive/stats12.q.out +++ b/ql/src/test/results/clientpositive/stats12.q.out @@ -49,7 +49,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart_n3 - Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL Column stats: NONE Statistics Aggregation Key Prefix: default.analyze_srcpart_n3/ GatherStats: true Path -> Alias: http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/stats13.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/stats13.q.out b/ql/src/test/results/clientpositive/stats13.q.out index 6dcb779..04a100c 100644 --- a/ql/src/test/results/clientpositive/stats13.q.out +++ b/ql/src/test/results/clientpositive/stats13.q.out @@ -49,7 +49,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart - Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL Column stats: NONE Statistics Aggregation Key Prefix: default.analyze_srcpart/ GatherStats: true Path -> Alias: http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/stats2.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/stats2.q.out b/ql/src/test/results/clientpositive/stats2.q.out index bcd5877..e6e91a6 100644 --- a/ql/src/test/results/clientpositive/stats2.q.out +++ b/ql/src/test/results/clientpositive/stats2.q.out @@ -130,7 +130,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_t1 - Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL Column stats: NONE Stage: Stage-1 Stats Work http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/stats7.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/stats7.q.out b/ql/src/test/results/clientpositive/stats7.q.out index 403c645..971c934 100644 --- a/ql/src/test/results/clientpositive/stats7.q.out +++ b/ql/src/test/results/clientpositive/stats7.q.out @@ -47,7 +47,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart_n4 - Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL Column stats: NONE Stage: Stage-1 Stats Work http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/stats8.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/stats8.q.out b/ql/src/test/results/clientpositive/stats8.q.out index e68b1d1..8d77e02 100644 --- a/ql/src/test/results/clientpositive/stats8.q.out +++ b/ql/src/test/results/clientpositive/stats8.q.out @@ -33,6 +33,41 @@ POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=11).key SIMPLE POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: describe formatted analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=11) +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@analyze_srcpart_n1 +POSTHOOK: query: describe formatted analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=11) +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@analyze_srcpart_n1 +# col_name data_type comment +key string default +value string default + +# Partition Information +# col_name data_type comment +ds string +hr string + +# Detailed Partition Information +Partition Value: [2008-04-08, 11] +Database: default +Table: analyze_srcpart_n1 +#### A masked pattern was here #### +Partition Parameters: + numFiles 1 + totalSize 5812 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +InputFormat: org.apache.hadoop.mapred.TextInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Storage Desc Params: + serialization.format 1 PREHOOK: query: explain analyze table analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=11) compute statistics PREHOOK: type: QUERY POSTHOOK: query: explain analyze table analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=11) compute statistics @@ -47,7 +82,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart_n1 - Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL Column stats: NONE Stage: Stage-1 Stats Work @@ -142,6 +177,41 @@ Bucket Columns: [] Sort Columns: [] Storage Desc Params: serialization.format 1 +PREHOOK: query: describe formatted analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=12) +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@analyze_srcpart_n1 +POSTHOOK: query: describe formatted analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=12) +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@analyze_srcpart_n1 +# col_name data_type comment +key string default +value string default + +# Partition Information +# col_name data_type comment +ds string +hr string + +# Detailed Partition Information +Partition Value: [2008-04-08, 12] +Database: default +Table: analyze_srcpart_n1 +#### A masked pattern was here #### +Partition Parameters: + numFiles 1 + totalSize 5812 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +InputFormat: org.apache.hadoop.mapred.TextInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Storage Desc Params: + serialization.format 1 PREHOOK: query: explain analyze table analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=12) compute statistics PREHOOK: type: QUERY POSTHOOK: query: explain analyze table analyze_srcpart_n1 PARTITION(ds='2008-04-08',hr=12) compute statistics @@ -156,7 +226,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart_n1 - Statistics: Num rows: 500 Data size: 5312 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 794 Data size: 179672 Basic stats: PARTIAL Column stats: NONE Stage: Stage-1 Stats Work @@ -226,7 +296,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart_n1 - Statistics: Num rows: 1000 Data size: 10624 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1196 Data size: 126864 Basic stats: PARTIAL Column stats: NONE Stage: Stage-1 Stats Work @@ -296,7 +366,7 @@ STAGE PLANS: Map Operator Tree: TableScan alias: analyze_srcpart_n1 - Statistics: Num rows: 1500 Data size: 15936 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1598 Data size: 74056 Basic stats: PARTIAL Column stats: NONE Stage: Stage-1 Stats Work