This is an automated email from the ASF dual-hosted git repository.
gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push:
new d2aee0a95b5 [MINOR][SQL] Fix typo and whitespaces in SQLConf
d2aee0a95b5 is described below
commit d2aee0a95b5c1664efbda54fad83e422ff81da1a
Author: AndreyBozhko <[email protected]>
AuthorDate: Mon Feb 20 09:27:39 2023 +0900
[MINOR][SQL] Fix typo and whitespaces in SQLConf
### What changes were proposed in this pull request?
- Fix `SIMPLIFIEID` -> `SIMPLIFIED`
- Fix indentation and whitespaces around a few `val` definitions
### Why are the changes needed?
Fix typo and code formatting
### Does this PR introduce _any_ user-facing change?
No
### How was this patch tested?
Existing tests pass
Closes #40086 from AndreyBozhko/sqlconf-typo.
Authored-by: AndreyBozhko <[email protected]>
Signed-off-by: Hyukjin Kwon <[email protected]>
---
.../main/scala/org/apache/spark/sql/internal/SQLConf.scala | 12 ++++++------
1 file changed, 6 insertions(+), 6 deletions(-)
diff --git
a/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
b/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
index e764e0510d9..32042ffd862 100644
--- a/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
+++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
@@ -823,7 +823,7 @@ object SQLConf {
.booleanConf
.createWithDefault(true)
- val ADAPTIVE_REBALANCE_PARTITIONS_SMALL_PARTITION_FACTOR =
+ val ADAPTIVE_REBALANCE_PARTITIONS_SMALL_PARTITION_FACTOR =
buildConf("spark.sql.adaptive.rebalancePartitionsSmallPartitionFactor")
.doc(s"A partition will be merged during splitting if its size is small
than this factor " +
s"multiply ${ADVISORY_PARTITION_SIZE_IN_BYTES.key}.")
@@ -1111,7 +1111,7 @@ object SQLConf {
.intConf
.createWithDefault(4096)
- val PARQUET_FIELD_ID_WRITE_ENABLED =
+ val PARQUET_FIELD_ID_WRITE_ENABLED =
buildConf("spark.sql.parquet.fieldId.write.enabled")
.doc("Field ID is a native field of the Parquet schema spec. When
enabled, " +
"Parquet writers will populate the field Id " +
@@ -2798,7 +2798,7 @@ object SQLConf {
.version("3.0.0")
.fallbackConf(BUFFER_SIZE)
- val PYSPARK_SIMPLIFIEID_TRACEBACK =
+ val PYSPARK_SIMPLIFIED_TRACEBACK =
buildConf("spark.sql.execution.pyspark.udf.simplifiedTraceback.enabled")
.doc(
"When true, the traceback from Python UDFs is simplified. It hides " +
@@ -3770,7 +3770,7 @@ object SQLConf {
.booleanConf
.createWithDefault(false)
- val LEGACY_INTEGER_GROUPING_ID =
+ val LEGACY_INTEGER_GROUPING_ID =
buildConf("spark.sql.legacy.integerGroupingId")
.internal()
.doc("When true, grouping_id() returns int values instead of long
values.")
@@ -3778,7 +3778,7 @@ object SQLConf {
.booleanConf
.createWithDefault(false)
- val LEGACY_GROUPING_ID_WITH_APPENDED_USER_GROUPBY =
+ val LEGACY_GROUPING_ID_WITH_APPENDED_USER_GROUPBY =
buildConf("spark.sql.legacy.groupingIdWithAppendedUserGroupBy")
.internal()
.doc("When true, grouping_id() returns values based on grouping set
columns plus " +
@@ -4772,7 +4772,7 @@ class SQLConf extends Serializable with Logging {
def pandasUDFBufferSize: Int = getConf(PANDAS_UDF_BUFFER_SIZE)
- def pysparkSimplifiedTraceback: Boolean =
getConf(PYSPARK_SIMPLIFIEID_TRACEBACK)
+ def pysparkSimplifiedTraceback: Boolean =
getConf(PYSPARK_SIMPLIFIED_TRACEBACK)
def pandasGroupedMapAssignColumnsByName: Boolean =
getConf(SQLConf.PANDAS_GROUPED_MAP_ASSIGN_COLUMNS_BY_NAME)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]