Github user kayousterhout commented on a diff in the pull request:

    https://github.com/apache/spark/pull/14079#discussion_r86607623
  
    --- Diff: 
core/src/test/scala/org/apache/spark/scheduler/BlacklistTrackerSuite.scala ---
    @@ -17,10 +17,299 @@
     
     package org.apache.spark.scheduler
     
    -import org.apache.spark.{SparkConf, SparkFunSuite}
    +import org.mockito.Mockito.when
    +import org.scalatest.BeforeAndAfterEach
    +import org.scalatest.mock.MockitoSugar
    +
    +import org.apache.spark._
     import org.apache.spark.internal.config
    +import org.apache.spark.util.ManualClock
    +
    +class BlacklistTrackerSuite extends SparkFunSuite with BeforeAndAfterEach 
with MockitoSugar
    +    with LocalSparkContext {
    +
    +  private val clock = new ManualClock(0)
    +
    +  private var blacklist: BlacklistTracker = _
    +  private var scheduler: TaskSchedulerImpl = _
    +  private var conf: SparkConf = _
    +
    +  override def afterEach(): Unit = {
    +    if (blacklist != null) {
    +      blacklist = null
    +    }
    +    if (scheduler != null) {
    +      scheduler.stop()
    +      scheduler = null
    +    }
    +    super.afterEach()
    +  }
    +
    +  val allExecutorAndHostIds = (('A' to 'Z').map("host" + _) ++ (1 to 
100).map{_.toString}).toSet
    +
    +  /**
    +   * Its easier to write our tests as if we could directly look at the 
sets of nodes & executors in
    +   * the blacklist.  However the api doesn't expose a set (for 
thread-safety), so this is a simple
    +   * way to test something similar, since we know the universe of values 
that might appear in these
    +   * sets.
    +   */
    +  def assertEquivalentToSet(f: String => Boolean, expected: Set[String]): 
Unit = {
    +    allExecutorAndHostIds.foreach { opt =>
    +      val actual = f(opt)
    +      val exp = expected.contains(opt)
    +      assert(actual === exp, raw"""for string "$opt" """)
    +    }
    +  }
    +
    +  def mockTaskSchedWithConf(conf: SparkConf): TaskSchedulerImpl = {
    +    sc = new SparkContext(conf)
    +    val scheduler = mock[TaskSchedulerImpl]
    +    when(scheduler.sc).thenReturn(sc)
    +    
when(scheduler.mapOutputTracker).thenReturn(SparkEnv.get.mapOutputTracker)
    +    scheduler
    +  }
     
    -class BlacklistTrackerSuite extends SparkFunSuite {
    +  def createTaskSetBlacklist(stageId: Int = 0): TaskSetBlacklist = {
    +    new TaskSetBlacklist(conf, stageId, clock)
    +  }
    +
    +  def configureBlacklistAndScheduler(confs: (String, String)*): Unit = {
    --- End diff --
    
    should this just be in a beforeEach()? I know it's not needed for the old 
tests but seems like it doesn't do any harm?
    
    Also eliminate confs, since it doesn't seem to ever be used?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to