This is an automated email from the ASF dual-hosted git repository.
ruifengz pushed a commit to branch branch-4.1
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/branch-4.1 by this push:
new c76797925eae [SPARK-54782][SQL] Correct the config versions
c76797925eae is described below
commit c76797925eaec6e1904e3d44a42b50a838f1b6fe
Author: Ruifeng Zheng <[email protected]>
AuthorDate: Mon Dec 22 07:51:48 2025 +0800
[SPARK-54782][SQL] Correct the config versions
### What changes were proposed in this pull request?
Correct the config versions
### Why are the changes needed?
the config versions are incorrect
### Does this PR introduce _any_ user-facing change?
doc-only change
### How was this patch tested?
manually check
### Was this patch authored or co-authored using generative AI tooling?
no
Closes #53545 from zhengruifeng/correct_version_info.
Authored-by: Ruifeng Zheng <[email protected]>
Signed-off-by: Ruifeng Zheng <[email protected]>
(cherry picked from commit d0cbad56a10502a1c931d5967beeae2369f6fa15)
Signed-off-by: Ruifeng Zheng <[email protected]>
---
.../org/apache/spark/sql/internal/SQLConf.scala | 64 ++++++++++++----------
1 file changed, 36 insertions(+), 28 deletions(-)
diff --git
a/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
b/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
index 36febe67f409..994291d3db5a 100644
--- a/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
+++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
@@ -243,40 +243,43 @@ object SQLConf {
val PREFER_COLUMN_OVER_LCA_IN_ARRAY_INDEX =
buildConf("spark.sql.analyzer.preferColumnOverLcaInArrayIndex")
- .internal()
- .doc(
- "When true, prefer the column from the underlying relation over the
lateral column alias " +
- "reference with the same name (see SPARK-53734)."
- )
- .booleanConf
- .createWithDefault(true)
+ .internal()
+ .version("4.1.0")
+ .doc(
+ "When true, prefer the column from the underlying relation over the
lateral column alias " +
+ "reference with the same name (see SPARK-53734).")
+ .booleanConf
+ .createWithDefault(true)
val DONT_DEDUPLICATE_EXPRESSION_IF_EXPR_ID_IN_OUTPUT =
buildConf("spark.sql.analyzer.dontDeduplicateExpressionIfExprIdInOutput")
- .internal()
- .doc(
- "DeduplicateRelations shouldn't remap expressions to new ExprIds if old
ExprId still " +
- "exists in output.")
- .booleanConf
- .createWithDefault(true)
+ .internal()
+ .version("4.1.0")
+ .doc(
+ "DeduplicateRelations shouldn't remap expressions to new ExprIds if
old ExprId still " +
+ "exists in output.")
+ .booleanConf
+ .createWithDefault(true)
val UNION_IS_RESOLVED_WHEN_DUPLICATES_PER_CHILD_RESOLVED =
buildConf("spark.sql.analyzer.unionIsResolvedWhenDuplicatesPerChildResolved")
- .internal()
- .doc(
- "When true, union should only be resolved once there are no duplicate
attributes in " +
- "each branch.")
- .booleanConf
- .createWithDefault(true)
+ .internal()
+ .version("4.1.0")
+ .doc(
+ "When true, union should only be resolved once there are no duplicate
attributes in " +
+ "each branch.")
+ .booleanConf
+ .createWithDefault(true)
val ONLY_NECESSARY_AND_UNIQUE_METADATA_COLUMNS =
buildConf("spark.sql.analyzer.uniqueNecessaryMetadataColumns")
- .internal()
- .doc(
- "When this conf is enabled, AddMetadataColumns rule should only add
necessary metadata " +
- "columns and only if those columns are not already present in the
project list.")
- .booleanConf
- .createWithDefault(true)
+ .internal()
+ .version("4.1.0")
+ .doc(
+ "When this conf is enabled, AddMetadataColumns rule should only add
necessary metadata " +
+ "columns and only if those columns are not already present in the
project list.")
+ .booleanConf
+ .createWithDefault(true)
val BLOCK_CREATE_TEMP_TABLE_USING_PROVIDER =
buildConf("spark.sql.legacy.blockCreateTempTableUsingProvider")
@@ -324,7 +327,7 @@ object SQLConf {
"(AliasResolution.resolve, FunctionResolution.resolveFunction, etc)." +
"This feature is currently under development."
)
- .version("4.0.0")
+ .version("4.1.0")
.booleanConf
.createWithDefault(false)
@@ -1057,7 +1060,7 @@ object SQLConf {
"An object with an explicitly set collation will not inherit the
collation from the " +
"schema."
)
- .version("4.0.0")
+ .version("4.1.0")
.booleanConf
.createWithDefault(false)
@@ -1911,6 +1914,7 @@ object SQLConf {
val DATA_SOURCE_V2_JOIN_PUSHDOWN =
buildConf("spark.sql.optimizer.datasourceV2JoinPushdown")
.internal()
+ .version("4.1.0")
.doc("When this config is set to true, join is tried to be pushed down" +
"for DSv2 data sources in V2ScanRelationPushdown optimization rule.")
.booleanConf
@@ -1919,6 +1923,7 @@ object SQLConf {
val DATA_SOURCE_V2_EXPR_FOLDING =
buildConf("spark.sql.optimizer.datasourceV2ExprFolding")
.internal()
+ .version("4.1.0")
.doc("When this config is set to true, do safe constant folding for the
" +
"expressions before translation and pushdown.")
.booleanConf
@@ -2549,6 +2554,7 @@ object SQLConf {
val STATE_STORE_MAINTENANCE_SHUTDOWN_TIMEOUT =
buildConf("spark.sql.streaming.stateStore.maintenanceShutdownTimeout")
.internal()
+ .version("4.1.0")
.doc("Timeout in seconds for maintenance pool operations to complete on
shutdown")
.timeConf(TimeUnit.SECONDS)
.createWithDefault(300L)
@@ -2556,6 +2562,7 @@ object SQLConf {
val STATE_STORE_MAINTENANCE_PROCESSING_TIMEOUT =
buildConf("spark.sql.streaming.stateStore.maintenanceProcessingTimeout")
.internal()
+ .version("4.1.0")
.doc("Timeout in seconds to wait for maintenance to process this
partition.")
.timeConf(TimeUnit.SECONDS)
.createWithDefault(30L)
@@ -2674,6 +2681,7 @@ object SQLConf {
"Note: For structured streaming, this configuration cannot be changed
between query " +
"restarts from the same checkpoint location.")
.internal()
+ .version("4.1.0")
.intConf
.checkValue(_ > 0,
"The value of spark.sql.streaming.internal.stateStore.partitions must
be a positive " +
@@ -3483,7 +3491,7 @@ object SQLConf {
.doc("When true, if a microbatch is retried, if a file already exists
but its checksum " +
"file does not exist, the file checksum will not be created. This is
useful for " +
"compatibility with files created before file checksums were enabled.")
- .version("4.2.0")
+ .version("4.1.0")
.booleanConf
.createWithDefault(true)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]