[spark] branch branch-3.0 updated: [SPARK-31619][CORE] Rename config "spark.dynamicAllocation.shuffleTimeout" to "spark.dynamicAllocation.shuffleTracking.timeout"

2020-04-30 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 78df2ca  [SPARK-31619][CORE] Rename config 
"spark.dynamicAllocation.shuffleTimeout" to 
"spark.dynamicAllocation.shuffleTracking.timeout"
78df2ca is described below

commit 78df2caec8c94c31e5c9ddc30ed8acb424084181
Author: Xingbo Jiang 
AuthorDate: Fri May 1 11:46:17 2020 +0900

[SPARK-31619][CORE] Rename config "spark.dynamicAllocation.shuffleTimeout" 
to "spark.dynamicAllocation.shuffleTracking.timeout"

### What changes were proposed in this pull request?
The "spark.dynamicAllocation.shuffleTimeout" configuration only takes 
effect if "spark.dynamicAllocation.shuffleTracking.enabled" is true, so we 
should re-namespace that configuration so that it's nested under the 
"shuffleTracking" one.

### How was this patch tested?
Covered by current existing test cases.

Closes #28426 from jiangxb1987/confName.

Authored-by: Xingbo Jiang 
Signed-off-by: HyukjinKwon 
(cherry picked from commit b7cde42b04b21c9bfee6535199cf385855c15853)
Signed-off-by: HyukjinKwon 
---
 .../scala/org/apache/spark/ExecutorAllocationManager.scala   |  2 +-
 .../scala/org/apache/spark/internal/config/package.scala |  6 +++---
 .../apache/spark/scheduler/dynalloc/ExecutorMonitor.scala|  4 ++--
 .../spark/scheduler/dynalloc/ExecutorMonitorSuite.scala  | 12 ++--
 docs/configuration.md|  2 +-
 5 files changed, 13 insertions(+), 13 deletions(-)

diff --git 
a/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala 
b/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala
index 677386c..74103dc 100644
--- a/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala
+++ b/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala
@@ -196,7 +196,7 @@ private[spark] class ExecutorAllocationManager(
 s"s${DYN_ALLOCATION_SUSTAINED_SCHEDULER_BACKLOG_TIMEOUT.key} must be > 
0!")
 }
 if (!conf.get(config.SHUFFLE_SERVICE_ENABLED)) {
-  if (conf.get(config.DYN_ALLOCATION_SHUFFLE_TRACKING)) {
+  if (conf.get(config.DYN_ALLOCATION_SHUFFLE_TRACKING_ENABLED)) {
 logWarning("Dynamic allocation without a shuffle service is an 
experimental feature.")
   } else if (!testing) {
 throw new SparkException("Dynamic allocation of executors requires the 
external " +
diff --git a/core/src/main/scala/org/apache/spark/internal/config/package.scala 
b/core/src/main/scala/org/apache/spark/internal/config/package.scala
index f70ee2e..bfdd0602 100644
--- a/core/src/main/scala/org/apache/spark/internal/config/package.scala
+++ b/core/src/main/scala/org/apache/spark/internal/config/package.scala
@@ -520,14 +520,14 @@ package object config {
   .checkValue(_ >= 0L, "Timeout must be >= 0.")
   .createWithDefault(60)
 
-  private[spark] val DYN_ALLOCATION_SHUFFLE_TRACKING =
+  private[spark] val DYN_ALLOCATION_SHUFFLE_TRACKING_ENABLED =
 ConfigBuilder("spark.dynamicAllocation.shuffleTracking.enabled")
   .version("3.0.0")
   .booleanConf
   .createWithDefault(false)
 
-  private[spark] val DYN_ALLOCATION_SHUFFLE_TIMEOUT =
-ConfigBuilder("spark.dynamicAllocation.shuffleTimeout")
+  private[spark] val DYN_ALLOCATION_SHUFFLE_TRACKING_TIMEOUT =
+ConfigBuilder("spark.dynamicAllocation.shuffleTracking.timeout")
   .version("3.0.0")
   .timeConf(TimeUnit.MILLISECONDS)
   .checkValue(_ >= 0L, "Timeout must be >= 0.")
diff --git 
a/core/src/main/scala/org/apache/spark/scheduler/dynalloc/ExecutorMonitor.scala 
b/core/src/main/scala/org/apache/spark/scheduler/dynalloc/ExecutorMonitor.scala
index a24f190..3664d3f 100644
--- 
a/core/src/main/scala/org/apache/spark/scheduler/dynalloc/ExecutorMonitor.scala
+++ 
b/core/src/main/scala/org/apache/spark/scheduler/dynalloc/ExecutorMonitor.scala
@@ -45,12 +45,12 @@ private[spark] class ExecutorMonitor(
   private val storageTimeoutNs = TimeUnit.SECONDS.toNanos(
 conf.get(DYN_ALLOCATION_CACHED_EXECUTOR_IDLE_TIMEOUT))
   private val shuffleTimeoutNs = TimeUnit.MILLISECONDS.toNanos(
-conf.get(DYN_ALLOCATION_SHUFFLE_TIMEOUT))
+conf.get(DYN_ALLOCATION_SHUFFLE_TRACKING_TIMEOUT))
 
   private val fetchFromShuffleSvcEnabled = conf.get(SHUFFLE_SERVICE_ENABLED) &&
 conf.get(SHUFFLE_SERVICE_FETCH_RDD_ENABLED)
   private val shuffleTrackingEnabled = !conf.get(SHUFFLE_SERVICE_ENABLED) &&
-conf.get(DYN_ALLOCATION_SHUFFLE_TRACKING)
+conf.get(DYN_ALLOCATION_SHUFFLE_TRACKING_ENABLED)
 
   private val executors = new ConcurrentHashMap[String, Tracker]()
   private val execResourceProfileCount = new ConcurrentHashMap[Int, Int]()
diff --git 
a/core/src/test/scala/org/apache/spark/scheduler/dynalloc/ExecutorMonitorSuite.scala
 

[spark] branch branch-3.0 updated: [SPARK-31619][CORE] Rename config "spark.dynamicAllocation.shuffleTimeout" to "spark.dynamicAllocation.shuffleTracking.timeout"

2020-04-30 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 78df2ca  [SPARK-31619][CORE] Rename config 
"spark.dynamicAllocation.shuffleTimeout" to 
"spark.dynamicAllocation.shuffleTracking.timeout"
78df2ca is described below

commit 78df2caec8c94c31e5c9ddc30ed8acb424084181
Author: Xingbo Jiang 
AuthorDate: Fri May 1 11:46:17 2020 +0900

[SPARK-31619][CORE] Rename config "spark.dynamicAllocation.shuffleTimeout" 
to "spark.dynamicAllocation.shuffleTracking.timeout"

### What changes were proposed in this pull request?
The "spark.dynamicAllocation.shuffleTimeout" configuration only takes 
effect if "spark.dynamicAllocation.shuffleTracking.enabled" is true, so we 
should re-namespace that configuration so that it's nested under the 
"shuffleTracking" one.

### How was this patch tested?
Covered by current existing test cases.

Closes #28426 from jiangxb1987/confName.

Authored-by: Xingbo Jiang 
Signed-off-by: HyukjinKwon 
(cherry picked from commit b7cde42b04b21c9bfee6535199cf385855c15853)
Signed-off-by: HyukjinKwon 
---
 .../scala/org/apache/spark/ExecutorAllocationManager.scala   |  2 +-
 .../scala/org/apache/spark/internal/config/package.scala |  6 +++---
 .../apache/spark/scheduler/dynalloc/ExecutorMonitor.scala|  4 ++--
 .../spark/scheduler/dynalloc/ExecutorMonitorSuite.scala  | 12 ++--
 docs/configuration.md|  2 +-
 5 files changed, 13 insertions(+), 13 deletions(-)

diff --git 
a/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala 
b/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala
index 677386c..74103dc 100644
--- a/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala
+++ b/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala
@@ -196,7 +196,7 @@ private[spark] class ExecutorAllocationManager(
 s"s${DYN_ALLOCATION_SUSTAINED_SCHEDULER_BACKLOG_TIMEOUT.key} must be > 
0!")
 }
 if (!conf.get(config.SHUFFLE_SERVICE_ENABLED)) {
-  if (conf.get(config.DYN_ALLOCATION_SHUFFLE_TRACKING)) {
+  if (conf.get(config.DYN_ALLOCATION_SHUFFLE_TRACKING_ENABLED)) {
 logWarning("Dynamic allocation without a shuffle service is an 
experimental feature.")
   } else if (!testing) {
 throw new SparkException("Dynamic allocation of executors requires the 
external " +
diff --git a/core/src/main/scala/org/apache/spark/internal/config/package.scala 
b/core/src/main/scala/org/apache/spark/internal/config/package.scala
index f70ee2e..bfdd0602 100644
--- a/core/src/main/scala/org/apache/spark/internal/config/package.scala
+++ b/core/src/main/scala/org/apache/spark/internal/config/package.scala
@@ -520,14 +520,14 @@ package object config {
   .checkValue(_ >= 0L, "Timeout must be >= 0.")
   .createWithDefault(60)
 
-  private[spark] val DYN_ALLOCATION_SHUFFLE_TRACKING =
+  private[spark] val DYN_ALLOCATION_SHUFFLE_TRACKING_ENABLED =
 ConfigBuilder("spark.dynamicAllocation.shuffleTracking.enabled")
   .version("3.0.0")
   .booleanConf
   .createWithDefault(false)
 
-  private[spark] val DYN_ALLOCATION_SHUFFLE_TIMEOUT =
-ConfigBuilder("spark.dynamicAllocation.shuffleTimeout")
+  private[spark] val DYN_ALLOCATION_SHUFFLE_TRACKING_TIMEOUT =
+ConfigBuilder("spark.dynamicAllocation.shuffleTracking.timeout")
   .version("3.0.0")
   .timeConf(TimeUnit.MILLISECONDS)
   .checkValue(_ >= 0L, "Timeout must be >= 0.")
diff --git 
a/core/src/main/scala/org/apache/spark/scheduler/dynalloc/ExecutorMonitor.scala 
b/core/src/main/scala/org/apache/spark/scheduler/dynalloc/ExecutorMonitor.scala
index a24f190..3664d3f 100644
--- 
a/core/src/main/scala/org/apache/spark/scheduler/dynalloc/ExecutorMonitor.scala
+++ 
b/core/src/main/scala/org/apache/spark/scheduler/dynalloc/ExecutorMonitor.scala
@@ -45,12 +45,12 @@ private[spark] class ExecutorMonitor(
   private val storageTimeoutNs = TimeUnit.SECONDS.toNanos(
 conf.get(DYN_ALLOCATION_CACHED_EXECUTOR_IDLE_TIMEOUT))
   private val shuffleTimeoutNs = TimeUnit.MILLISECONDS.toNanos(
-conf.get(DYN_ALLOCATION_SHUFFLE_TIMEOUT))
+conf.get(DYN_ALLOCATION_SHUFFLE_TRACKING_TIMEOUT))
 
   private val fetchFromShuffleSvcEnabled = conf.get(SHUFFLE_SERVICE_ENABLED) &&
 conf.get(SHUFFLE_SERVICE_FETCH_RDD_ENABLED)
   private val shuffleTrackingEnabled = !conf.get(SHUFFLE_SERVICE_ENABLED) &&
-conf.get(DYN_ALLOCATION_SHUFFLE_TRACKING)
+conf.get(DYN_ALLOCATION_SHUFFLE_TRACKING_ENABLED)
 
   private val executors = new ConcurrentHashMap[String, Tracker]()
   private val execResourceProfileCount = new ConcurrentHashMap[Int, Int]()
diff --git 
a/core/src/test/scala/org/apache/spark/scheduler/dynalloc/ExecutorMonitorSuite.scala