Author: xuefu
Date: Tue Mar 10 22:19:16 2015
New Revision: 1665728

URL: http://svn.apache.org/r1665728
Log:
PIG-4456: Sort the leaves by SparkOperator.operatorKey in 
SparkLauncher#sparkOperToRDD (Liyun via Xuefu)

Modified:
    
pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/SparkLauncher.java

Modified: 
pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/SparkLauncher.java
URL: 
http://svn.apache.org/viewvc/pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/SparkLauncher.java?rev=1665728&r1=1665727&r2=1665728&view=diff
==============================================================================
--- 
pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/SparkLauncher.java
 (original)
+++ 
pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/SparkLauncher.java
 Tue Mar 10 22:19:16 2015
@@ -25,6 +25,7 @@ import java.nio.file.Paths;
 import java.util.ArrayList;
 import java.util.Arrays;
 import java.util.Collections;
+import java.util.Comparator;
 import java.util.HashMap;
 import java.util.HashSet;
 import java.util.LinkedList;
@@ -479,6 +480,7 @@ public class SparkLauncher extends Launc
                Set<Integer> seenJobIDs = new HashSet<Integer>();
                if (sparkPlan != null) {
                        List<SparkOperator> leaves = sparkPlan.getLeaves();
+                       Collections.sort(leaves);
                        Map<OperatorKey, RDD<Tuple>> sparkOpToRdds = new 
HashMap();
                        for (SparkOperator leaf : leaves) {
                                Map<OperatorKey, RDD<Tuple>> physicalOpToRdds = 
new HashMap();


Reply via email to