beliefer commented on a change in pull request #27852:
[SPARK-31002][CORE][DOC][FOLLOWUP] Add version information to the configuration
of Core
URL: https://github.com/apache/spark/pull/27852#discussion_r389486734
##########
File path: core/src/main/scala/org/apache/spark/internal/config/package.scala
##########
@@ -461,97 +461,133 @@ package object config {
.intConf
.createWithDefault(5)
- private[spark] val IS_PYTHON_APP =
ConfigBuilder("spark.yarn.isPython").internal()
- .booleanConf.createWithDefault(false)
+ private[spark] val IS_PYTHON_APP =
+ ConfigBuilder("spark.yarn.isPython")
+ .internal()
+ .version("1.5.0")
+ .booleanConf
+ .createWithDefault(false)
- private[spark] val CPUS_PER_TASK =
ConfigBuilder("spark.task.cpus").intConf.createWithDefault(1)
+ private[spark] val CPUS_PER_TASK =
+
ConfigBuilder("spark.task.cpus").version("0.5.0").intConf.createWithDefault(1)
private[spark] val DYN_ALLOCATION_ENABLED =
-
ConfigBuilder("spark.dynamicAllocation.enabled").booleanConf.createWithDefault(false)
+ ConfigBuilder("spark.dynamicAllocation.enabled")
+ .version("1.2.0")
+ .booleanConf
+ .createWithDefault(false)
private[spark] val DYN_ALLOCATION_TESTING =
-
ConfigBuilder("spark.dynamicAllocation.testing").booleanConf.createWithDefault(false)
+ ConfigBuilder("spark.dynamicAllocation.testing")
+ .version("1.2.0")
+ .booleanConf
+ .createWithDefault(false)
private[spark] val DYN_ALLOCATION_MIN_EXECUTORS =
-
ConfigBuilder("spark.dynamicAllocation.minExecutors").intConf.createWithDefault(0)
+ ConfigBuilder("spark.dynamicAllocation.minExecutors")
+ .version("1.2.0")
+ .intConf
+ .createWithDefault(0)
private[spark] val DYN_ALLOCATION_INITIAL_EXECUTORS =
ConfigBuilder("spark.dynamicAllocation.initialExecutors")
+ .version("1.3.0")
.fallbackConf(DYN_ALLOCATION_MIN_EXECUTORS)
private[spark] val DYN_ALLOCATION_MAX_EXECUTORS =
-
ConfigBuilder("spark.dynamicAllocation.maxExecutors").intConf.createWithDefault(Int.MaxValue)
+ ConfigBuilder("spark.dynamicAllocation.maxExecutors")
+ .version("1.2.0")
+ .intConf
+ .createWithDefault(Int.MaxValue)
private[spark] val DYN_ALLOCATION_EXECUTOR_ALLOCATION_RATIO =
ConfigBuilder("spark.dynamicAllocation.executorAllocationRatio")
- .doubleConf.createWithDefault(1.0)
+ .version("2.4.0")
+ .doubleConf
+ .createWithDefault(1.0)
private[spark] val DYN_ALLOCATION_CACHED_EXECUTOR_IDLE_TIMEOUT =
ConfigBuilder("spark.dynamicAllocation.cachedExecutorIdleTimeout")
+ .version("1.4.0")
.timeConf(TimeUnit.SECONDS)
.checkValue(_ >= 0L, "Timeout must be >= 0.")
.createWithDefault(Integer.MAX_VALUE)
private[spark] val DYN_ALLOCATION_EXECUTOR_IDLE_TIMEOUT =
ConfigBuilder("spark.dynamicAllocation.executorIdleTimeout")
+ .version("1.2.0")
Review comment:
SPARK-3795, commit ID:
8d59b37b02eb36f37bcefafb952519d7dca744ad#diff-364713d7776956cb8b0a771e9b62f82d
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]
With regards,
Apache Git Services
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]