http://git-wip-us.apache.org/repos/asf/hive/blob/6a99cf53/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out index 475b709..2fa0214 100644 --- a/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out @@ -160,12 +160,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: r - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: string), value (type: string) @@ -240,12 +240,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: l - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -256,18 +256,18 @@ STAGE PLANS: input vertices: 1 Map 2 Position of Big Table: 0 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Select Operator expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat @@ -369,12 +369,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: r - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: string), value (type: string) @@ -449,12 +449,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: l - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -465,18 +465,18 @@ STAGE PLANS: input vertices: 1 Map 2 Position of Big Table: 0 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE BucketMapJoin: true Select Operator expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat @@ -578,12 +578,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: r - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 (key + key) (type: double) @@ -653,12 +653,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: l - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key + key) is not null (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -669,17 +669,17 @@ STAGE PLANS: input vertices: 1 Map 2 Position of Big Table: 0 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat @@ -776,12 +776,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: r - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: string), value (type: string) @@ -851,12 +851,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: l - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -867,17 +867,17 @@ STAGE PLANS: input vertices: 1 Map 2 Position of Big Table: 0 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat @@ -974,12 +974,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: r - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: string), value (type: string) @@ -1049,12 +1049,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: l - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -1065,17 +1065,17 @@ STAGE PLANS: input vertices: 1 Map 2 Position of Big Table: 0 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat @@ -1172,12 +1172,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: r - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: string), value (type: string) @@ -1247,12 +1247,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: l - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -1263,17 +1263,17 @@ STAGE PLANS: input vertices: 1 Map 2 Position of Big Table: 0 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat @@ -1370,12 +1370,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: r - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: string), value (type: string) @@ -1445,12 +1445,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: l - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -1461,17 +1461,17 @@ STAGE PLANS: input vertices: 1 Map 2 Position of Big Table: 0 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat @@ -1568,12 +1568,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: r - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: string), value (type: string) @@ -1643,12 +1643,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: l - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -1659,17 +1659,17 @@ STAGE PLANS: input vertices: 1 Map 2 Position of Big Table: 0 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat @@ -1766,12 +1766,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: r - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 key (type: string), value (type: string) @@ -1841,12 +1841,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: l - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 4200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4200 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -1857,17 +1857,17 @@ STAGE PLANS: input vertices: 1 Map 2 Position of Big Table: 0 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 23 Data size: 4620 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 4620 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/6a99cf53/ql/src/test/results/clientpositive/spark/count.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/count.q.out b/ql/src/test/results/clientpositive/spark/count.q.out index 8df0d68..bb29363 100644 --- a/ql/src/test/results/clientpositive/spark/count.q.out +++ b/ql/src/test/results/clientpositive/spark/count.q.out @@ -48,22 +48,22 @@ STAGE PLANS: Map Operator Tree: TableScan alias: abcd - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: a (type: int), b (type: int), c (type: int), d (type: int) outputColumnNames: a, b, c, d - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT b), count(DISTINCT c), sum(d) keys: a (type: int), b (type: int), c (type: int) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) sort order: +++ Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE value expressions: _col5 (type: bigint) Reducer 2 Reduce Operator Tree: @@ -72,10 +72,10 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -118,21 +118,21 @@ STAGE PLANS: Map Operator Tree: TableScan alias: abcd - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: a (type: int), b (type: int), c (type: int), d (type: int) outputColumnNames: a, b, c, d - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(), count(a), count(b), count(c), count(d), count(DISTINCT a), count(DISTINCT b), count(DISTINCT c), count(DISTINCT d), count(DISTINCT a, b), count(DISTINCT b, c), count(DISTINCT c, d), count(DISTINCT a, d), count(DISTINCT a, c), count(DISTINCT b, d), count(DISTINCT a, b, c), count(DISTINCT b, c, d), count(DISTINCT a, c, d), count(DISTINCT a, b, d), count(DISTINCT a, b, c, d) keys: a (type: int), b (type: int), c (type: int), d (type: int) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23 - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: int) sort order: ++++ - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE value expressions: _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint) Reducer 2 Reduce Operator Tree: @@ -187,16 +187,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: abcd - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: a (type: int), b (type: int), c (type: int), d (type: int) outputColumnNames: a, b, c, d - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: a (type: int), b (type: int), c (type: int) sort order: +++ Map-reduce partition columns: a (type: int) - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE value expressions: d (type: int) Reducer 2 Reduce Operator Tree: @@ -205,10 +205,10 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: complete outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -251,15 +251,15 @@ STAGE PLANS: Map Operator Tree: TableScan alias: abcd - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: a (type: int), b (type: int), c (type: int), d (type: int) outputColumnNames: a, b, c, d - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: a (type: int), b (type: int), c (type: int), d (type: int) sort order: ++++ - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Group By Operator @@ -313,21 +313,21 @@ STAGE PLANS: Map Operator Tree: TableScan alias: abcd - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: a (type: int), b (type: int), c (type: int), d (type: int) outputColumnNames: a, b, c, d - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(), count(a), count(b), count(c), count(d), count(DISTINCT a), count(DISTINCT b), count(DISTINCT c), count(DISTINCT d), count(DISTINCT a, b), count(DISTINCT b, c), count(DISTINCT c, d), count(DISTINCT a, d), count(DISTINCT a, c), count(DISTINCT b, d), count(DISTINCT a, b, c), count(DISTINCT b, c, d), count(DISTINCT a, c, d), count(DISTINCT a, b, d), count(DISTINCT a, b, c, d) keys: a (type: int), b (type: int), c (type: int), d (type: int) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23 - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: int) sort order: ++++ - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE value expressions: _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint) Reducer 2 Reduce Operator Tree: @@ -383,28 +383,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: abcd - Statistics: Num rows: 9 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: a (type: int), b (type: int) outputColumnNames: a, b - Statistics: Num rows: 9 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: a (type: int), b (type: int) sort order: ++ Map-reduce partition columns: a (type: int), b (type: int) - Statistics: Num rows: 9 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Group By Operator keys: KEY._col0 (type: int), KEY._col1 (type: int) mode: complete outputColumnNames: a, b - Statistics: Num rows: 4 Data size: 34 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: a (type: int) sort order: + Map-reduce partition columns: a (type: int) - Statistics: Num rows: 4 Data size: 34 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE value expressions: b (type: int) Reducer 3 Reduce Operator Tree: @@ -413,14 +413,14 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: complete outputColumnNames: a, $f1 - Statistics: Num rows: 2 Data size: 17 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: $f1 (type: bigint) outputColumnNames: _o__c0 - Statistics: Num rows: 2 Data size: 17 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 2 Data size: 17 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -464,28 +464,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: abcd - Statistics: Num rows: 19 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: b (type: int) outputColumnNames: b - Statistics: Num rows: 19 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: b (type: int) sort order: + Map-reduce partition columns: b (type: int) - Statistics: Num rows: 19 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Group By Operator keys: KEY._col0 (type: int) mode: complete outputColumnNames: b - Statistics: Num rows: 9 Data size: 36 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: b (type: int) sort order: + Map-reduce partition columns: b (type: int) - Statistics: Num rows: 9 Data size: 36 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reducer 3 Reduce Operator Tree: Group By Operator @@ -493,14 +493,14 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: complete outputColumnNames: b, $f1 - Statistics: Num rows: 4 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: $f1 (type: bigint) outputColumnNames: _o__c0 - Statistics: Num rows: 4 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 4 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -544,28 +544,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: abcd - Statistics: Num rows: 9 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: b (type: int), c (type: int) outputColumnNames: b, c - Statistics: Num rows: 9 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: b (type: int), c (type: int) sort order: ++ Map-reduce partition columns: b (type: int), c (type: int) - Statistics: Num rows: 9 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Group By Operator keys: KEY._col0 (type: int), KEY._col1 (type: int) mode: complete outputColumnNames: b, c - Statistics: Num rows: 4 Data size: 34 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: c (type: int) sort order: + Map-reduce partition columns: c (type: int) - Statistics: Num rows: 4 Data size: 34 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE value expressions: b (type: int) Reducer 3 Reduce Operator Tree: @@ -574,14 +574,14 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: complete outputColumnNames: c, $f1 - Statistics: Num rows: 2 Data size: 17 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: $f1 (type: bigint) outputColumnNames: _o__c0 - Statistics: Num rows: 2 Data size: 17 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 2 Data size: 17 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -627,16 +627,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: abcd - Statistics: Num rows: 6 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: b (type: int), c (type: int), d (type: int) outputColumnNames: b, c, d - Statistics: Num rows: 6 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: d (type: int), c (type: int) sort order: ++ Map-reduce partition columns: d (type: int) - Statistics: Num rows: 6 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE value expressions: b (type: int) Reducer 2 Reduce Operator Tree: @@ -645,14 +645,14 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: complete outputColumnNames: d, $f1, $f2 - Statistics: Num rows: 3 Data size: 39 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: $f1 (type: bigint), $f2 (type: bigint) outputColumnNames: _o__c0, _o__c1 - Statistics: Num rows: 3 Data size: 39 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 3 Data size: 39 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -698,16 +698,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: abcd - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: a (type: int), b (type: int), c (type: int), d (type: int), (d + d) (type: int), (d * 3) (type: int) outputColumnNames: $f0, $f1, $f2, $f3, $f4, $f5 - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: $f0 (type: int), $f1 (type: int), $f2 (type: int) sort order: +++ Map-reduce partition columns: $f0 (type: int) - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE value expressions: $f3 (type: int), $f4 (type: int), $f5 (type: int) Reducer 2 Reduce Operator Tree: @@ -716,10 +716,10 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: complete outputColumnNames: $f0, $f1, $f2, $f3, $f4, $f5, $f6, $f7, $f8, $f9, $f10 - Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -762,16 +762,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: abcd - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: a (type: int), b (type: int), c (type: int), d (type: int) outputColumnNames: a, b, c, d - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: a (type: int), b (type: int), c (type: int) sort order: +++ Map-reduce partition columns: a (type: int) - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE value expressions: d (type: int) Reducer 2 Reduce Operator Tree: @@ -780,10 +780,10 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: complete outputColumnNames: a, $f1, $f2, $f3, $f4 - Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 2 Data size: 39 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -826,15 +826,15 @@ STAGE PLANS: Map Operator Tree: TableScan alias: abcd - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: a (type: int), b (type: int), c (type: int), d (type: int) outputColumnNames: a, b, c, d - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: a (type: int), b (type: int), c (type: int), d (type: int) sort order: ++++ - Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Group By Operator http://git-wip-us.apache.org/repos/asf/hive/blob/6a99cf53/ql/src/test/results/clientpositive/spark/gen_udf_example_add10.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/gen_udf_example_add10.q.out b/ql/src/test/results/clientpositive/spark/gen_udf_example_add10.q.out index 6abdddc..88eaf86 100644 --- a/ql/src/test/results/clientpositive/spark/gen_udf_example_add10.q.out +++ b/ql/src/test/results/clientpositive/spark/gen_udf_example_add10.q.out @@ -39,28 +39,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: t1 - Statistics: Num rows: 2 Data size: 30 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: add10(x) (type: int), add10(y) (type: double) outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 30 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: double) sort order: -+ - Statistics: Num rows: 2 Data size: 30 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE TopN Hash Memory Usage: 0.1 Reducer 2 Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: double) outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 30 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 10 - Statistics: Num rows: 2 Data size: 30 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 2 Data size: 30 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/6a99cf53/ql/src/test/results/clientpositive/spark/groupby10.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/groupby10.q.out b/ql/src/test/results/clientpositive/spark/groupby10.q.out index b572995..b7d19fe 100644 --- a/ql/src/test/results/clientpositive/spark/groupby10.q.out +++ b/ql/src/test/results/clientpositive/spark/groupby10.q.out @@ -61,16 +61,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: input - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: int), value (type: string) outputColumnNames: key, value - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: key (type: int), substr(value, 5) (type: string) sort order: ++ Map-reduce partition columns: key (type: int) - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Group By Operator @@ -78,12 +78,12 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: partial1 outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: bigint), _col2 (type: bigint) Reducer 3 Reduce Operator Tree: @@ -92,14 +92,14 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), UDFToInteger(_col1) (type: int), UDFToInteger(_col2) (type: int) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -112,12 +112,12 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: partial1 outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: double), _col2 (type: double) Reducer 5 Reduce Operator Tree: @@ -126,14 +126,14 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), UDFToInteger(_col1) (type: int), UDFToInteger(_col2) (type: int) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -273,16 +273,16 @@ STAGE PLANS: Map Operator Tree: TableScan alias: input - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: int), value (type: string) outputColumnNames: key, value - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: key (type: int), substr(value, 5) (type: string) sort order: ++ Map-reduce partition columns: key (type: int) - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Group By Operator @@ -290,12 +290,12 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: partial1 outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: bigint), _col2 (type: bigint) Reducer 3 Reduce Operator Tree: @@ -304,14 +304,14 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), UDFToInteger(_col1) (type: int), UDFToInteger(_col2) (type: int) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -324,12 +324,12 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: partial1 outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: double), _col2 (type: double) Reducer 5 Reduce Operator Tree: @@ -338,14 +338,14 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), UDFToInteger(_col1) (type: int), UDFToInteger(_col2) (type: int) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -482,33 +482,33 @@ STAGE PLANS: Map Operator Tree: TableScan alias: input - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: int), value (type: string) outputColumnNames: key, value - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: key (type: int), substr(value, 5) (type: string) sort order: ++ Map-reduce partition columns: key (type: int) - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Forward - Statistics: Num rows: 2 Data size: 280 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(DISTINCT KEY._col1:0._col0), count(DISTINCT KEY._col1:1._col0) keys: KEY._col0 (type: int) mode: complete outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), UDFToInteger(_col1) (type: int), UDFToInteger(_col2) (type: int) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -519,14 +519,14 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: complete outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), UDFToInteger(_col1) (type: int), UDFToInteger(_col2) (type: int) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 140 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
