Repository: hive Updated Branches: refs/heads/spark c9073aadc -> fd1192914
http://git-wip-us.apache.org/repos/asf/hive/blob/fd119291/ql/src/test/results/clientpositive/spark/vector_outer_join5.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/vector_outer_join5.q.out b/ql/src/test/results/clientpositive/spark/vector_outer_join5.q.out index 2c7cd5b..2b13dc6 100644 --- a/ql/src/test/results/clientpositive/spark/vector_outer_join5.q.out +++ b/ql/src/test/results/clientpositive/spark/vector_outer_join5.q.out @@ -90,18 +90,18 @@ STAGE PLANS: Map Operator Tree: TableScan alias: st - Statistics: Num rows: 100 Data size: 372 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 380 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ctinyint (type: tinyint) outputColumnNames: _col0 - Statistics: Num rows: 100 Data size: 372 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 380 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 _col0 (type: tinyint) 1 _col0 (type: tinyint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Stage: Stage-1 Spark @@ -113,11 +113,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: s - Statistics: Num rows: 6058 Data size: 2018 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ctinyint (type: tinyint) outputColumnNames: _col0 - Statistics: Num rows: 6058 Data size: 2018 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2027 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Left Outer Join0 to 1 @@ -126,7 +126,7 @@ STAGE PLANS: 1 _col0 (type: tinyint) input vertices: 1 Map 3 - Statistics: Num rows: 6663 Data size: 2219 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6663 Data size: 2229 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -136,10 +136,11 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Reducer 2 + Execution mode: vectorized Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) @@ -153,7 +154,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized Stage: Stage-0 Fetch Operator @@ -208,11 +208,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: sm - Statistics: Num rows: 100 Data size: 372 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 380 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ctinyint (type: tinyint) outputColumnNames: _col0 - Statistics: Num rows: 100 Data size: 372 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 380 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator filter predicates: 0 {(_col1 = 2)} @@ -220,9 +220,9 @@ STAGE PLANS: keys: 0 _col0 (type: tinyint) 1 _col0 (type: tinyint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Stage: Stage-1 Spark @@ -234,11 +234,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: s - Statistics: Num rows: 6058 Data size: 2018 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ctinyint (type: tinyint), cmodint (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 6058 Data size: 2018 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2027 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Left Outer Join0 to 1 @@ -250,7 +250,7 @@ STAGE PLANS: 1 _col0 (type: tinyint) input vertices: 1 Map 3 - Statistics: Num rows: 6663 Data size: 2219 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6663 Data size: 2229 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -260,10 +260,11 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Reducer 2 + Execution mode: vectorized Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) @@ -277,7 +278,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized Stage: Stage-0 Fetch Operator @@ -332,11 +332,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: sm - Statistics: Num rows: 100 Data size: 372 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 380 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ctinyint (type: tinyint) outputColumnNames: _col0 - Statistics: Num rows: 100 Data size: 372 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 380 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator filter predicates: 0 {((UDFToInteger(_col0) pmod 4) = _col1)} @@ -344,9 +344,9 @@ STAGE PLANS: keys: 0 _col0 (type: tinyint) 1 _col0 (type: tinyint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Stage: Stage-1 Spark @@ -358,11 +358,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: s - Statistics: Num rows: 6058 Data size: 2018 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ctinyint (type: tinyint), cmodint (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 6058 Data size: 2018 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2027 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Left Outer Join0 to 1 @@ -374,7 +374,7 @@ STAGE PLANS: 1 _col0 (type: tinyint) input vertices: 1 Map 3 - Statistics: Num rows: 6663 Data size: 2219 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6663 Data size: 2229 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -384,10 +384,11 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Reducer 2 + Execution mode: vectorized Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) @@ -401,7 +402,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized Stage: Stage-0 Fetch Operator @@ -456,11 +456,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: sm - Statistics: Num rows: 100 Data size: 372 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 380 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ctinyint (type: tinyint) outputColumnNames: _col0 - Statistics: Num rows: 100 Data size: 372 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 380 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator filter predicates: 0 {(_col0 < 100)} @@ -468,9 +468,9 @@ STAGE PLANS: keys: 0 _col0 (type: tinyint) 1 _col0 (type: tinyint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Stage: Stage-1 Spark @@ -482,11 +482,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: s - Statistics: Num rows: 6058 Data size: 2018 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ctinyint (type: tinyint) outputColumnNames: _col0 - Statistics: Num rows: 6058 Data size: 2018 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2027 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Left Outer Join0 to 1 @@ -498,7 +498,7 @@ STAGE PLANS: 1 _col0 (type: tinyint) input vertices: 1 Map 3 - Statistics: Num rows: 6663 Data size: 2219 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6663 Data size: 2229 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -508,10 +508,11 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Reducer 2 + Execution mode: vectorized Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) @@ -525,7 +526,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized Stage: Stage-0 Fetch Operator @@ -584,34 +584,34 @@ STAGE PLANS: Map Operator Tree: TableScan alias: sm - Statistics: Num rows: 100 Data size: 372 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 380 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cbigint (type: bigint) outputColumnNames: _col0 - Statistics: Num rows: 100 Data size: 372 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 380 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 UDFToLong(_col1) (type: bigint) 1 (_col0 pmod UDFToLong(8)) (type: bigint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Map 4 Map Operator Tree: TableScan alias: s - Statistics: Num rows: 6058 Data size: 2018 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ctinyint (type: tinyint) outputColumnNames: _col0 - Statistics: Num rows: 6058 Data size: 2018 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2027 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 _col0 (type: tinyint) 1 _col0 (type: tinyint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Stage: Stage-1 Spark @@ -623,11 +623,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: s - Statistics: Num rows: 6058 Data size: 2018 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ctinyint (type: tinyint), cmodint (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 6058 Data size: 2018 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2027 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Left Outer Join0 to 1 @@ -637,7 +637,7 @@ STAGE PLANS: outputColumnNames: _col0 input vertices: 1 Map 3 - Statistics: Num rows: 6663 Data size: 2219 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6663 Data size: 2229 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Left Outer Join0 to 1 @@ -646,7 +646,7 @@ STAGE PLANS: 1 _col0 (type: tinyint) input vertices: 1 Map 4 - Statistics: Num rows: 7329 Data size: 2440 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 7329 Data size: 2451 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -656,10 +656,11 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Reducer 2 + Execution mode: vectorized Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) @@ -673,7 +674,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized Stage: Stage-0 Fetch Operator @@ -790,18 +790,18 @@ STAGE PLANS: Map Operator Tree: TableScan alias: st - Statistics: Num rows: 100 Data size: 352 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 363 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cmodtinyint (type: int) outputColumnNames: _col0 - Statistics: Num rows: 100 Data size: 352 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 363 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 _col0 (type: int) 1 _col0 (type: int) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Stage: Stage-1 Spark @@ -813,11 +813,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: s - Statistics: Num rows: 6058 Data size: 2785 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2793 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cmodtinyint (type: int) outputColumnNames: _col0 - Statistics: Num rows: 6058 Data size: 2785 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2793 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Left Outer Join0 to 1 @@ -826,7 +826,7 @@ STAGE PLANS: 1 _col0 (type: int) input vertices: 1 Map 3 - Statistics: Num rows: 6663 Data size: 3063 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6663 Data size: 3072 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -836,10 +836,11 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Reducer 2 + Execution mode: vectorized Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) @@ -853,7 +854,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized Stage: Stage-0 Fetch Operator @@ -908,11 +908,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: sm - Statistics: Num rows: 100 Data size: 352 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 363 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cmodtinyint (type: int) outputColumnNames: _col0 - Statistics: Num rows: 100 Data size: 352 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 363 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator filter predicates: 0 {(_col1 = 2)} @@ -920,9 +920,9 @@ STAGE PLANS: keys: 0 _col0 (type: int) 1 _col0 (type: int) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Stage: Stage-1 Spark @@ -934,11 +934,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: s - Statistics: Num rows: 6058 Data size: 2785 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2793 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cmodtinyint (type: int), cmodint (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 6058 Data size: 2785 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2793 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Left Outer Join0 to 1 @@ -950,7 +950,7 @@ STAGE PLANS: 1 _col0 (type: int) input vertices: 1 Map 3 - Statistics: Num rows: 6663 Data size: 3063 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6663 Data size: 3072 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -960,10 +960,11 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Reducer 2 + Execution mode: vectorized Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) @@ -977,7 +978,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized Stage: Stage-0 Fetch Operator @@ -1032,11 +1032,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: sm - Statistics: Num rows: 100 Data size: 352 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 363 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cmodtinyint (type: int) outputColumnNames: _col0 - Statistics: Num rows: 100 Data size: 352 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 363 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator filter predicates: 0 {((_col0 pmod 4) = _col1)} @@ -1044,9 +1044,9 @@ STAGE PLANS: keys: 0 _col0 (type: int) 1 _col0 (type: int) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Stage: Stage-1 Spark @@ -1058,11 +1058,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: s - Statistics: Num rows: 6058 Data size: 2785 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2793 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cmodtinyint (type: int), cmodint (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 6058 Data size: 2785 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2793 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Left Outer Join0 to 1 @@ -1074,7 +1074,7 @@ STAGE PLANS: 1 _col0 (type: int) input vertices: 1 Map 3 - Statistics: Num rows: 6663 Data size: 3063 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6663 Data size: 3072 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -1084,10 +1084,11 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Reducer 2 + Execution mode: vectorized Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) @@ -1101,7 +1102,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized Stage: Stage-0 Fetch Operator @@ -1156,11 +1156,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: sm - Statistics: Num rows: 100 Data size: 352 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 363 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cmodtinyint (type: int) outputColumnNames: _col0 - Statistics: Num rows: 100 Data size: 352 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 363 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator filter predicates: 0 {(_col0 < 3)} @@ -1168,9 +1168,9 @@ STAGE PLANS: keys: 0 _col0 (type: int) 1 _col0 (type: int) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Stage: Stage-1 Spark @@ -1182,11 +1182,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: s - Statistics: Num rows: 6058 Data size: 2785 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2793 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cmodtinyint (type: int) outputColumnNames: _col0 - Statistics: Num rows: 6058 Data size: 2785 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2793 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Left Outer Join0 to 1 @@ -1198,7 +1198,7 @@ STAGE PLANS: 1 _col0 (type: int) input vertices: 1 Map 3 - Statistics: Num rows: 6663 Data size: 3063 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6663 Data size: 3072 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -1208,10 +1208,11 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Reducer 2 + Execution mode: vectorized Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) @@ -1225,7 +1226,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized Stage: Stage-0 Fetch Operator @@ -1284,34 +1284,34 @@ STAGE PLANS: Map Operator Tree: TableScan alias: sm - Statistics: Num rows: 100 Data size: 352 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 363 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cbigint (type: bigint) outputColumnNames: _col0 - Statistics: Num rows: 100 Data size: 352 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 363 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 UDFToLong(_col1) (type: bigint) 1 (_col0 pmod UDFToLong(8)) (type: bigint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Map 4 Map Operator Tree: TableScan alias: s - Statistics: Num rows: 6058 Data size: 2785 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2793 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cmodtinyint (type: int) outputColumnNames: _col0 - Statistics: Num rows: 6058 Data size: 2785 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2793 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: 0 _col0 (type: int) 1 _col0 (type: int) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Stage: Stage-1 Spark @@ -1323,11 +1323,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: s - Statistics: Num rows: 6058 Data size: 2785 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2793 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cmodtinyint (type: int), cmodint (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 6058 Data size: 2785 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6058 Data size: 2793 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Left Outer Join0 to 1 @@ -1337,7 +1337,7 @@ STAGE PLANS: outputColumnNames: _col0 input vertices: 1 Map 3 - Statistics: Num rows: 6663 Data size: 3063 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6663 Data size: 3072 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Left Outer Join0 to 1 @@ -1346,7 +1346,7 @@ STAGE PLANS: 1 _col0 (type: int) input vertices: 1 Map 4 - Statistics: Num rows: 7329 Data size: 3369 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 7329 Data size: 3379 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -1356,10 +1356,11 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) + Execution mode: vectorized Local Work: Map Reduce Local Work - Execution mode: vectorized Reducer 2 + Execution mode: vectorized Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) @@ -1373,7 +1374,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized Stage: Stage-0 Fetch Operator
