Github user tnachen commented on a diff in the pull request:

    https://github.com/apache/spark/pull/21150#discussion_r227028091
  
    --- Diff: 
resource-managers/mesos/src/main/scala/org/apache/spark/scheduler/cluster/mesos/MesosClusterScheduler.scala
 ---
    @@ -728,6 +729,28 @@ private[spark] class MesosClusterScheduler(
           state == MesosTaskState.TASK_LOST
       }
     
    +  /**
    +   * Check if the driver has exceed the number of retries.
    +   * When "spark.mesos.driver.supervise.maxRetries" is not set,
    +   * the default behavior is to retry indefinitely
    +   *
    +   * @param retryState Retry state of the driver
    +   * @param conf Spark Context to check if it contains 
"spark.mesos.driver.supervise.maxRetries"
    +   * @return true if driver has reached retry limit
    +   *         false if driver can be retried
    +   */
    +  private[scheduler] def hasDriverExceededRetries(retryState: 
Option[MesosClusterRetryState],
    --- End diff --
    
    Please fix the param style:
    hasDriverExceededRetries(
         retryState: Option[MesosClusterRetryState],
         conf.....) 


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to