Github user vanzin commented on a diff in the pull request:
    --- Diff: 
core/src/main/scala/org/apache/spark/ExecutorAllocationClient.scala ---
    @@ -55,18 +55,18 @@ private[spark] trait ExecutorAllocationClient {
        * Request that the cluster manager kill the specified executors.
    -   * When asking the executor to be replaced, the executor loss is 
considered a failure, and
    -   * killed tasks that are running on the executor will count towards the 
failure limits. If no
    -   * replacement is being requested, then the tasks will not count towards 
the limit.
    -   *
        * @param executorIds identifiers of executors to kill
    -   * @param replace whether to replace the killed executors with new ones, 
default false
    +   * @param adjustTargetNumExecutors whether the target number of 
executors will be adjusted down
    +   *                                 after these executors have been killed
    +   * @param countFailures if there are tasks running on the executors when 
they are killed, whether
    +   *                      those failures be counted to task failure limits?
    --- End diff --
    nit: "whether to count those failures toward task failure limits"


To unsubscribe, e-mail:
For additional commands, e-mail:

Reply via email to