Repository: hive
Updated Branches:
  refs/heads/branch-3 26c5749eb -> ba54eb5de


http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out_spark
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out_spark 
b/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out_spark
index 63752f3..dbb2460 100644
--- a/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out_spark
+++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out_spark
@@ -78,12 +78,12 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: b
-                  Statistics: Num rows: 69 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 69 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                   GatherStats: false
                   Filter Operator
                     isSamplingPred: false
                     predicate: key is not null (type: boolean)
-                    Statistics: Num rows: 69 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 69 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                     Spark HashTable Sink Operator
                       keys:
                         0 key (type: int)
@@ -162,12 +162,12 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: a
-                  Statistics: Num rows: 69 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 69 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                   GatherStats: false
                   Filter Operator
                     isSamplingPred: false
                     predicate: key is not null (type: boolean)
-                    Statistics: Num rows: 69 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 69 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                     Map Join Operator
                       condition map:
                            Inner Join 0 to 1
@@ -178,17 +178,17 @@ STAGE PLANS:
                       input vertices:
                         1 Map 3
                       Position of Big Table: 0
-                      Statistics: Num rows: 75 Data size: 30250 Basic stats: 
COMPLETE Column stats: NONE
+                      Statistics: Num rows: 75 Data size: 30250 Basic stats: 
PARTIAL Column stats: NONE
                       BucketMapJoin: true
                       Select Operator
                         expressions: _col0 (type: int), _col8 (type: string)
                         outputColumnNames: _col0, _col1
-                        Statistics: Num rows: 75 Data size: 30250 Basic stats: 
COMPLETE Column stats: NONE
+                        Statistics: Num rows: 75 Data size: 30250 Basic stats: 
PARTIAL Column stats: NONE
                         Reduce Output Operator
                           key expressions: _col0 (type: int), _col1 (type: 
string)
                           null sort order: aa
                           sort order: ++
-                          Statistics: Num rows: 75 Data size: 30250 Basic 
stats: COMPLETE Column stats: NONE
+                          Statistics: Num rows: 75 Data size: 30250 Basic 
stats: PARTIAL Column stats: NONE
                           tag: -1
                           TopN: 1
                           TopN Hash Memory Usage: 0.1
@@ -261,16 +261,16 @@ STAGE PLANS:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: int), 
KEY.reducesinkkey1 (type: string)
                 outputColumnNames: _col0, _col1
-                Statistics: Num rows: 75 Data size: 30250 Basic stats: 
COMPLETE Column stats: NONE
+                Statistics: Num rows: 75 Data size: 30250 Basic stats: PARTIAL 
Column stats: NONE
                 Limit
                   Number of rows: 1
-                  Statistics: Num rows: 1 Data size: 403 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 403 Basic stats: PARTIAL 
Column stats: NONE
                   File Output Operator
                     compressed: false
                     GlobalTableId: 0
 #### A masked pattern was here ####
                     NumFilesPerFileSink: 1
-                    Statistics: Num rows: 1 Data size: 403 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 403 Basic stats: 
PARTIAL Column stats: NONE
 #### A masked pattern was here ####
                     table:
                         input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out 
b/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out
index b22022c..ecd33c7 100644
--- a/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out
@@ -84,12 +84,12 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: b
-                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                   GatherStats: false
                   Filter Operator
                     isSamplingPred: false
                     predicate: key is not null (type: boolean)
-                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                     Spark HashTable Sink Operator
                       keys:
                         0 key (type: int)
@@ -167,12 +167,12 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: a
-                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                   GatherStats: false
                   Filter Operator
                     isSamplingPred: false
                     predicate: key is not null (type: boolean)
-                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                     Map Join Operator
                       condition map:
                            Inner Join 0 to 1
@@ -182,17 +182,17 @@ STAGE PLANS:
                       input vertices:
                         1 Map 3
                       Position of Big Table: 0
-                      Statistics: Num rows: 104 Data size: 30250 Basic stats: 
COMPLETE Column stats: NONE
+                      Statistics: Num rows: 104 Data size: 30250 Basic stats: 
PARTIAL Column stats: NONE
                       BucketMapJoin: true
                       Group By Operator
                         aggregations: count()
                         mode: hash
                         outputColumnNames: _col0
-                        Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 8 Basic stats: 
PARTIAL Column stats: NONE
                         Reduce Output Operator
                           null sort order: 
                           sort order: 
-                          Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
+                          Statistics: Num rows: 1 Data size: 8 Basic stats: 
PARTIAL Column stats: NONE
                           tag: -1
                           value expressions: _col0 (type: bigint)
                           auto parallelism: false
@@ -264,13 +264,13 @@ STAGE PLANS:
                 aggregations: count(VALUE._col0)
                 mode: mergepartial
                 outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL 
Column stats: NONE
                 File Output Operator
                   compressed: false
                   GlobalTableId: 0
 #### A masked pattern was here ####
                   NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL 
Column stats: NONE
 #### A masked pattern was here ####
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -345,12 +345,12 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: b
-                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                   GatherStats: false
                   Filter Operator
                     isSamplingPred: false
                     predicate: key is not null (type: boolean)
-                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                     Spark HashTable Sink Operator
                       keys:
                         0 key (type: int)
@@ -428,12 +428,12 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: a
-                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                   GatherStats: false
                   Filter Operator
                     isSamplingPred: false
                     predicate: key is not null (type: boolean)
-                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                     Map Join Operator
                       condition map:
                            Inner Join 0 to 1
@@ -443,17 +443,17 @@ STAGE PLANS:
                       input vertices:
                         1 Map 3
                       Position of Big Table: 0
-                      Statistics: Num rows: 104 Data size: 30250 Basic stats: 
COMPLETE Column stats: NONE
+                      Statistics: Num rows: 104 Data size: 30250 Basic stats: 
PARTIAL Column stats: NONE
                       BucketMapJoin: true
                       Group By Operator
                         aggregations: count()
                         mode: hash
                         outputColumnNames: _col0
-                        Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 8 Basic stats: 
PARTIAL Column stats: NONE
                         Reduce Output Operator
                           null sort order: 
                           sort order: 
-                          Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
+                          Statistics: Num rows: 1 Data size: 8 Basic stats: 
PARTIAL Column stats: NONE
                           tag: -1
                           value expressions: _col0 (type: bigint)
                           auto parallelism: false
@@ -525,13 +525,13 @@ STAGE PLANS:
                 aggregations: count(VALUE._col0)
                 mode: mergepartial
                 outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL 
Column stats: NONE
                 File Output Operator
                   compressed: false
                   GlobalTableId: 0
 #### A masked pattern was here ####
                   NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL 
Column stats: NONE
 #### A masked pattern was here ####
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out 
b/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out
index 7d455ba..d03b938 100644
--- a/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out
@@ -92,12 +92,12 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: b
-                  Statistics: Num rows: 145 Data size: 42000 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 145 Data size: 42000 Basic stats: 
PARTIAL Column stats: NONE
                   GatherStats: false
                   Filter Operator
                     isSamplingPred: false
                     predicate: key is not null (type: boolean)
-                    Statistics: Num rows: 145 Data size: 42000 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 145 Data size: 42000 Basic stats: 
PARTIAL Column stats: NONE
                     Spark HashTable Sink Operator
                       keys:
                         0 key (type: int)
@@ -170,12 +170,12 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: a
-                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                   GatherStats: false
                   Filter Operator
                     isSamplingPred: false
                     predicate: key is not null (type: boolean)
-                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                     Map Join Operator
                       condition map:
                            Inner Join 0 to 1
@@ -185,16 +185,16 @@ STAGE PLANS:
                       input vertices:
                         1 Map 3
                       Position of Big Table: 0
-                      Statistics: Num rows: 159 Data size: 46200 Basic stats: 
COMPLETE Column stats: NONE
+                      Statistics: Num rows: 159 Data size: 46200 Basic stats: 
PARTIAL Column stats: NONE
                       Group By Operator
                         aggregations: count()
                         mode: hash
                         outputColumnNames: _col0
-                        Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 8 Basic stats: 
PARTIAL Column stats: NONE
                         Reduce Output Operator
                           null sort order: 
                           sort order: 
-                          Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
+                          Statistics: Num rows: 1 Data size: 8 Basic stats: 
PARTIAL Column stats: NONE
                           tag: -1
                           value expressions: _col0 (type: bigint)
                           auto parallelism: false
@@ -261,13 +261,13 @@ STAGE PLANS:
                 aggregations: count(VALUE._col0)
                 mode: mergepartial
                 outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL 
Column stats: NONE
                 File Output Operator
                   compressed: false
                   GlobalTableId: 0
 #### A masked pattern was here ####
                   NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL 
Column stats: NONE
 #### A masked pattern was here ####
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -375,12 +375,12 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: b
-                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                   GatherStats: false
                   Filter Operator
                     isSamplingPred: false
                     predicate: key is not null (type: boolean)
-                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                     Spark HashTable Sink Operator
                       keys:
                         0 key (type: int)
@@ -453,12 +453,12 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: a
-                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                   GatherStats: false
                   Filter Operator
                     isSamplingPred: false
                     predicate: key is not null (type: boolean)
-                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 95 Data size: 27500 Basic stats: 
PARTIAL Column stats: NONE
                     Map Join Operator
                       condition map:
                            Inner Join 0 to 1
@@ -468,16 +468,16 @@ STAGE PLANS:
                       input vertices:
                         1 Map 3
                       Position of Big Table: 0
-                      Statistics: Num rows: 104 Data size: 30250 Basic stats: 
COMPLETE Column stats: NONE
+                      Statistics: Num rows: 104 Data size: 30250 Basic stats: 
PARTIAL Column stats: NONE
                       Group By Operator
                         aggregations: count()
                         mode: hash
                         outputColumnNames: _col0
-                        Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 8 Basic stats: 
PARTIAL Column stats: NONE
                         Reduce Output Operator
                           null sort order: 
                           sort order: 
-                          Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
+                          Statistics: Num rows: 1 Data size: 8 Basic stats: 
PARTIAL Column stats: NONE
                           tag: -1
                           value expressions: _col0 (type: bigint)
                           auto parallelism: false
@@ -544,13 +544,13 @@ STAGE PLANS:
                 aggregations: count(VALUE._col0)
                 mode: mergepartial
                 outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL 
Column stats: NONE
                 File Output Operator
                   compressed: false
                   GlobalTableId: 0
 #### A masked pattern was here ####
                   NumFilesPerFileSink: 1
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL 
Column stats: NONE
 #### A masked pattern was here ####
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out 
b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out
index ece73fe..e53de75 100644
--- a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out
@@ -90,12 +90,12 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: b
-                  Statistics: Num rows: 140 Data size: 42000 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 140 Data size: 42000 Basic stats: 
PARTIAL Column stats: NONE
                   GatherStats: false
                   Filter Operator
                     isSamplingPred: false
                     predicate: key is not null (type: boolean)
-                    Statistics: Num rows: 140 Data size: 42000 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 140 Data size: 42000 Basic stats: 
PARTIAL Column stats: NONE
                     Spark HashTable Sink Operator
                       keys:
                         0 key (type: int)
@@ -183,17 +183,17 @@ STAGE PLANS:
                       input vertices:
                         1 Map 2
                       Position of Big Table: 0
-                      Statistics: Num rows: 154 Data size: 46200 Basic stats: 
COMPLETE Column stats: NONE
+                      Statistics: Num rows: 154 Data size: 46200 Basic stats: 
PARTIAL Column stats: NONE
                       Select Operator
                         expressions: UDFToString(_col0) (type: string), _col1 
(type: string), _col6 (type: string)
                         outputColumnNames: _col0, _col1, _col2
-                        Statistics: Num rows: 154 Data size: 46200 Basic 
stats: COMPLETE Column stats: NONE
+                        Statistics: Num rows: 154 Data size: 46200 Basic 
stats: PARTIAL Column stats: NONE
                         File Output Operator
                           compressed: false
                           GlobalTableId: 1
 #### A masked pattern was here ####
                           NumFilesPerFileSink: 1
-                          Statistics: Num rows: 154 Data size: 46200 Basic 
stats: COMPLETE Column stats: NONE
+                          Statistics: Num rows: 154 Data size: 46200 Basic 
stats: PARTIAL Column stats: NONE
 #### A masked pattern was here ####
                           table:
                               input format: 
org.apache.hadoop.mapred.TextInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out 
b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out
index 350f860..c1fd0c1 100644
--- a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out
@@ -99,12 +99,12 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: b
-                  Statistics: Num rows: 156 Data size: 61240 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 156 Data size: 61240 Basic stats: 
PARTIAL Column stats: NONE
                   GatherStats: false
                   Filter Operator
                     isSamplingPred: false
                     predicate: key is not null (type: boolean)
-                    Statistics: Num rows: 156 Data size: 61240 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 156 Data size: 61240 Basic stats: 
PARTIAL Column stats: NONE
                     Spark HashTable Sink Operator
                       keys:
                         0 key (type: int)
@@ -247,18 +247,18 @@ STAGE PLANS:
                       input vertices:
                         1 Map 2
                       Position of Big Table: 0
-                      Statistics: Num rows: 171 Data size: 67364 Basic stats: 
COMPLETE Column stats: NONE
+                      Statistics: Num rows: 171 Data size: 67364 Basic stats: 
PARTIAL Column stats: NONE
                       BucketMapJoin: true
                       Select Operator
                         expressions: UDFToString(_col0) (type: string), _col1 
(type: string), _col6 (type: string)
                         outputColumnNames: _col0, _col1, _col2
-                        Statistics: Num rows: 171 Data size: 67364 Basic 
stats: COMPLETE Column stats: NONE
+                        Statistics: Num rows: 171 Data size: 67364 Basic 
stats: PARTIAL Column stats: NONE
                         File Output Operator
                           compressed: false
                           GlobalTableId: 1
 #### A masked pattern was here ####
                           NumFilesPerFileSink: 1
-                          Statistics: Num rows: 171 Data size: 67364 Basic 
stats: COMPLETE Column stats: NONE
+                          Statistics: Num rows: 171 Data size: 67364 Basic 
stats: PARTIAL Column stats: NONE
 #### A masked pattern was here ####
                           table:
                               input format: 
org.apache.hadoop.mapred.TextInputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/smb_mapjoin_10.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/smb_mapjoin_10.q.out 
b/ql/src/test/results/clientpositive/spark/smb_mapjoin_10.q.out
index 68d3952..c64dc41 100644
--- a/ql/src/test/results/clientpositive/spark/smb_mapjoin_10.q.out
+++ b/ql/src/test/results/clientpositive/spark/smb_mapjoin_10.q.out
@@ -81,10 +81,10 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: b
-                  Statistics: Num rows: 13 Data size: 4140 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 13 Data size: 4140 Basic stats: 
PARTIAL Column stats: NONE
                   Filter Operator
                     predicate: (pageid is not null and postid is not null and 
type is not null and userid is not null) (type: boolean)
-                    Statistics: Num rows: 13 Data size: 4140 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 13 Data size: 4140 Basic stats: 
PARTIAL Column stats: NONE
                     Sorted Merge Bucket Map Join Operator
                       condition map:
                            Inner Join 0 to 1
@@ -92,14 +92,14 @@ STAGE PLANS:
                         0 userid (type: int), pageid (type: int), postid 
(type: int), type (type: string)
                         1 userid (type: int), pageid (type: int), postid 
(type: int), type (type: string)
                       outputColumnNames: _col0, _col1, _col2, _col3, _col8, 
_col9, _col10, _col11
-                      Statistics: Num rows: 14 Data size: 4554 Basic stats: 
COMPLETE Column stats: NONE
+                      Statistics: Num rows: 14 Data size: 4554 Basic stats: 
PARTIAL Column stats: NONE
                       Select Operator
                         expressions: _col0 (type: int), _col1 (type: int), 
_col2 (type: int), _col3 (type: string), '1' (type: string), _col8 (type: int), 
_col9 (type: int), _col10 (type: int), _col11 (type: string), '2' (type: string)
                         outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7, _col8, _col9
-                        Statistics: Num rows: 14 Data size: 4554 Basic stats: 
COMPLETE Column stats: NONE
+                        Statistics: Num rows: 14 Data size: 4554 Basic stats: 
PARTIAL Column stats: NONE
                         File Output Operator
                           compressed: false
-                          Statistics: Num rows: 14 Data size: 4554 Basic 
stats: COMPLETE Column stats: NONE
+                          Statistics: Num rows: 14 Data size: 4554 Basic 
stats: PARTIAL Column stats: NONE
                           table:
                               input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                               output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/stats12.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/stats12.q.out 
b/ql/src/test/results/clientpositive/spark/stats12.q.out
index c976da8..b5dac094b 100644
--- a/ql/src/test/results/clientpositive/spark/stats12.q.out
+++ b/ql/src/test/results/clientpositive/spark/stats12.q.out
@@ -52,7 +52,7 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: analyze_srcpart_n3
-                  Statistics: Num rows: 392 Data size: 232480 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 392 Data size: 232480 Basic stats: 
PARTIAL Column stats: NONE
                   Statistics Aggregation Key Prefix: 
default.analyze_srcpart_n3/
                   GatherStats: true
             Path -> Alias:

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/stats13.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/stats13.q.out 
b/ql/src/test/results/clientpositive/spark/stats13.q.out
index d59ca8b..074fd3f 100644
--- a/ql/src/test/results/clientpositive/spark/stats13.q.out
+++ b/ql/src/test/results/clientpositive/spark/stats13.q.out
@@ -52,7 +52,7 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: analyze_srcpart
-                  Statistics: Num rows: 392 Data size: 232480 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 392 Data size: 232480 Basic stats: 
PARTIAL Column stats: NONE
                   Statistics Aggregation Key Prefix: default.analyze_srcpart/
                   GatherStats: true
             Path -> Alias:

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/stats2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/stats2.q.out 
b/ql/src/test/results/clientpositive/spark/stats2.q.out
index c6283c4..04831f8 100644
--- a/ql/src/test/results/clientpositive/spark/stats2.q.out
+++ b/ql/src/test/results/clientpositive/spark/stats2.q.out
@@ -136,7 +136,7 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: analyze_t1
-                  Statistics: Num rows: 392 Data size: 232480 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 392 Data size: 232480 Basic stats: 
PARTIAL Column stats: NONE
 
   Stage: Stage-1
     Stats Work

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/stats7.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/stats7.q.out 
b/ql/src/test/results/clientpositive/spark/stats7.q.out
index f23f409..3ae274e 100644
--- a/ql/src/test/results/clientpositive/spark/stats7.q.out
+++ b/ql/src/test/results/clientpositive/spark/stats7.q.out
@@ -50,7 +50,7 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: analyze_srcpart_n4
-                  Statistics: Num rows: 392 Data size: 232480 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 392 Data size: 232480 Basic stats: 
PARTIAL Column stats: NONE
 
   Stage: Stage-1
     Stats Work

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/spark/stats8.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/stats8.q.out 
b/ql/src/test/results/clientpositive/spark/stats8.q.out
index c19ff2e..767ff21 100644
--- a/ql/src/test/results/clientpositive/spark/stats8.q.out
+++ b/ql/src/test/results/clientpositive/spark/stats8.q.out
@@ -33,6 +33,41 @@ POSTHOOK: Lineage: analyze_srcpart_n1 
PARTITION(ds=2008-04-09,hr=11).key SIMPLE
 POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=11).value 
SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), 
]
 POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=12).key 
SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=12).value 
SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), 
]
+PREHOOK: query: describe formatted analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=11)
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@analyze_srcpart_n1
+POSTHOOK: query: describe formatted analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=11)
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@analyze_srcpart_n1
+# col_name             data_type               comment             
+key                    string                  default             
+value                  string                  default             
+                
+# Partition Information                 
+# col_name             data_type               comment             
+ds                     string                                      
+hr                     string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [2008-04-08, 11]         
+Database:              default                  
+Table:                 analyze_srcpart_n1       
+#### A masked pattern was here ####
+Partition Parameters:           
+       numFiles                1                   
+       totalSize               5812                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
 PREHOOK: query: explain analyze table analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=11) compute statistics
 PREHOOK: type: QUERY
 POSTHOOK: query: explain analyze table analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=11) compute statistics
@@ -50,7 +85,7 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: analyze_srcpart_n1
-                  Statistics: Num rows: 392 Data size: 232480 Basic stats: 
COMPLETE Column stats: NONE
+                  Statistics: Num rows: 392 Data size: 232480 Basic stats: 
PARTIAL Column stats: NONE
 
   Stage: Stage-1
     Stats Work
@@ -145,6 +180,41 @@ Bucket Columns:            []
 Sort Columns:          []                       
 Storage Desc Params:            
        serialization.format    1                   
+PREHOOK: query: describe formatted analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=12)
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@analyze_srcpart_n1
+POSTHOOK: query: describe formatted analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=12)
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@analyze_srcpart_n1
+# col_name             data_type               comment             
+key                    string                  default             
+value                  string                  default             
+                
+# Partition Information                 
+# col_name             data_type               comment             
+ds                     string                                      
+hr                     string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [2008-04-08, 12]         
+Database:              default                  
+Table:                 analyze_srcpart_n1       
+#### A masked pattern was here ####
+Partition Parameters:           
+       numFiles                1                   
+       totalSize               5812                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
 PREHOOK: query: explain analyze table analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=12) compute statistics
 PREHOOK: type: QUERY
 POSTHOOK: query: explain analyze table analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=12) compute statistics
@@ -162,7 +232,7 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: analyze_srcpart_n1
-                  Statistics: Num rows: 500 Data size: 5312 Basic stats: 
PARTIAL Column stats: NONE
+                  Statistics: Num rows: 794 Data size: 179672 Basic stats: 
PARTIAL Column stats: NONE
 
   Stage: Stage-1
     Stats Work
@@ -235,7 +305,7 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: analyze_srcpart_n1
-                  Statistics: Num rows: 1000 Data size: 10624 Basic stats: 
PARTIAL Column stats: NONE
+                  Statistics: Num rows: 1196 Data size: 126864 Basic stats: 
PARTIAL Column stats: NONE
 
   Stage: Stage-1
     Stats Work
@@ -308,7 +378,7 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: analyze_srcpart_n1
-                  Statistics: Num rows: 1500 Data size: 15936 Basic stats: 
PARTIAL Column stats: NONE
+                  Statistics: Num rows: 1598 Data size: 74056 Basic stats: 
PARTIAL Column stats: NONE
 
   Stage: Stage-1
     Stats Work

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/stats12.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/stats12.q.out 
b/ql/src/test/results/clientpositive/stats12.q.out
index ec0143e..923fc78 100644
--- a/ql/src/test/results/clientpositive/stats12.q.out
+++ b/ql/src/test/results/clientpositive/stats12.q.out
@@ -49,7 +49,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: analyze_srcpart_n3
-            Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL 
Column stats: NONE
             Statistics Aggregation Key Prefix: default.analyze_srcpart_n3/
             GatherStats: true
       Path -> Alias:

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/stats13.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/stats13.q.out 
b/ql/src/test/results/clientpositive/stats13.q.out
index 6dcb779..04a100c 100644
--- a/ql/src/test/results/clientpositive/stats13.q.out
+++ b/ql/src/test/results/clientpositive/stats13.q.out
@@ -49,7 +49,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: analyze_srcpart
-            Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL 
Column stats: NONE
             Statistics Aggregation Key Prefix: default.analyze_srcpart/
             GatherStats: true
       Path -> Alias:

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/stats2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/stats2.q.out 
b/ql/src/test/results/clientpositive/stats2.q.out
index bcd5877..e6e91a6 100644
--- a/ql/src/test/results/clientpositive/stats2.q.out
+++ b/ql/src/test/results/clientpositive/stats2.q.out
@@ -130,7 +130,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: analyze_t1
-            Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL 
Column stats: NONE
 
   Stage: Stage-1
     Stats Work

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/stats7.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/stats7.q.out 
b/ql/src/test/results/clientpositive/stats7.q.out
index 403c645..971c934 100644
--- a/ql/src/test/results/clientpositive/stats7.q.out
+++ b/ql/src/test/results/clientpositive/stats7.q.out
@@ -47,7 +47,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: analyze_srcpart_n4
-            Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL 
Column stats: NONE
 
   Stage: Stage-1
     Stats Work

http://git-wip-us.apache.org/repos/asf/hive/blob/ba54eb5d/ql/src/test/results/clientpositive/stats8.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/stats8.q.out 
b/ql/src/test/results/clientpositive/stats8.q.out
index e68b1d1..8d77e02 100644
--- a/ql/src/test/results/clientpositive/stats8.q.out
+++ b/ql/src/test/results/clientpositive/stats8.q.out
@@ -33,6 +33,41 @@ POSTHOOK: Lineage: analyze_srcpart_n1 
PARTITION(ds=2008-04-09,hr=11).key SIMPLE
 POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=11).value 
SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), 
]
 POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=12).key 
SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: analyze_srcpart_n1 PARTITION(ds=2008-04-09,hr=12).value 
SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), 
]
+PREHOOK: query: describe formatted analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=11)
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@analyze_srcpart_n1
+POSTHOOK: query: describe formatted analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=11)
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@analyze_srcpart_n1
+# col_name             data_type               comment             
+key                    string                  default             
+value                  string                  default             
+                
+# Partition Information                 
+# col_name             data_type               comment             
+ds                     string                                      
+hr                     string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [2008-04-08, 11]         
+Database:              default                  
+Table:                 analyze_srcpart_n1       
+#### A masked pattern was here ####
+Partition Parameters:           
+       numFiles                1                   
+       totalSize               5812                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
 PREHOOK: query: explain analyze table analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=11) compute statistics
 PREHOOK: type: QUERY
 POSTHOOK: query: explain analyze table analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=11) compute statistics
@@ -47,7 +82,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: analyze_srcpart_n1
-            Statistics: Num rows: 392 Data size: 232480 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 392 Data size: 232480 Basic stats: PARTIAL 
Column stats: NONE
 
   Stage: Stage-1
     Stats Work
@@ -142,6 +177,41 @@ Bucket Columns:            []
 Sort Columns:          []                       
 Storage Desc Params:            
        serialization.format    1                   
+PREHOOK: query: describe formatted analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=12)
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@analyze_srcpart_n1
+POSTHOOK: query: describe formatted analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=12)
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@analyze_srcpart_n1
+# col_name             data_type               comment             
+key                    string                  default             
+value                  string                  default             
+                
+# Partition Information                 
+# col_name             data_type               comment             
+ds                     string                                      
+hr                     string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [2008-04-08, 12]         
+Database:              default                  
+Table:                 analyze_srcpart_n1       
+#### A masked pattern was here ####
+Partition Parameters:           
+       numFiles                1                   
+       totalSize               5812                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
 PREHOOK: query: explain analyze table analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=12) compute statistics
 PREHOOK: type: QUERY
 POSTHOOK: query: explain analyze table analyze_srcpart_n1 
PARTITION(ds='2008-04-08',hr=12) compute statistics
@@ -156,7 +226,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: analyze_srcpart_n1
-            Statistics: Num rows: 500 Data size: 5312 Basic stats: PARTIAL 
Column stats: NONE
+            Statistics: Num rows: 794 Data size: 179672 Basic stats: PARTIAL 
Column stats: NONE
 
   Stage: Stage-1
     Stats Work
@@ -226,7 +296,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: analyze_srcpart_n1
-            Statistics: Num rows: 1000 Data size: 10624 Basic stats: PARTIAL 
Column stats: NONE
+            Statistics: Num rows: 1196 Data size: 126864 Basic stats: PARTIAL 
Column stats: NONE
 
   Stage: Stage-1
     Stats Work
@@ -296,7 +366,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: analyze_srcpart_n1
-            Statistics: Num rows: 1500 Data size: 15936 Basic stats: PARTIAL 
Column stats: NONE
+            Statistics: Num rows: 1598 Data size: 74056 Basic stats: PARTIAL 
Column stats: NONE
 
   Stage: Stage-1
     Stats Work

Reply via email to