Github user skonto commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22004#discussion_r207754816
  
    --- Diff: 
core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala ---
    @@ -2369,39 +2369,12 @@ class DAGSchedulerSuite extends SparkFunSuite with 
LocalSparkContext with TimeLi
         assert(scheduler.getShuffleDependencies(rddE) === Set(shuffleDepA, 
shuffleDepC))
       }
     
    -  test("SPARK-17644: After one stage is aborted for too many failed 
attempts, subsequent stages" +
    +  test("SPARK-17644: After one stage is aborted for too many failed 
attempts, subsequent stages " +
         "still behave correctly on fetch failures") {
    -    // Runs a job that always encounters a fetch failure, so should 
eventually be aborted
    --- End diff --
    
    Yeah, I was also wondering if I had to implement this as well, but I feel 
people need to move to 2.12 with a different mindset as things have changed. 
Not sure if it is possible as well, so asked @LRytz in jira.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to