Author: xuefu
Date: Tue Mar 10 22:19:16 2015
New Revision: 1665728
URL: http://svn.apache.org/r1665728
Log:
PIG-4456: Sort the leaves by SparkOperator.operatorKey in
SparkLauncher#sparkOperToRDD (Liyun via Xuefu)
Modified:
pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/SparkLauncher.java
Modified:
pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/SparkLauncher.java
URL:
http://svn.apache.org/viewvc/pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/SparkLauncher.java?rev=1665728&r1=1665727&r2=1665728&view=diff
==============================================================================
---
pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/SparkLauncher.java
(original)
+++
pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/SparkLauncher.java
Tue Mar 10 22:19:16 2015
@@ -25,6 +25,7 @@ import java.nio.file.Paths;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.Collections;
+import java.util.Comparator;
import java.util.HashMap;
import java.util.HashSet;
import java.util.LinkedList;
@@ -479,6 +480,7 @@ public class SparkLauncher extends Launc
Set<Integer> seenJobIDs = new HashSet<Integer>();
if (sparkPlan != null) {
List<SparkOperator> leaves = sparkPlan.getLeaves();
+ Collections.sort(leaves);
Map<OperatorKey, RDD<Tuple>> sparkOpToRdds = new
HashMap();
for (SparkOperator leaf : leaves) {
Map<OperatorKey, RDD<Tuple>> physicalOpToRdds =
new HashMap();