This is an automated email from the ASF dual-hosted git repository. maxgekk pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push: new 6b92cda04e6 [SPARK-41581][SQL] Update `_LEGACY_ERROR_TEMP_1230` as `INTERNAL_ERROR` 6b92cda04e6 is described below commit 6b92cda04e618f82711587d027fa20601e094418 Author: itholic <haejoon....@databricks.com> AuthorDate: Mon Jan 9 10:41:49 2023 +0300 [SPARK-41581][SQL] Update `_LEGACY_ERROR_TEMP_1230` as `INTERNAL_ERROR` ### What changes were proposed in this pull request? This PR proposes to update `_LEGACY_ERROR_TEMP_1230`, as `INTERNAL_ERROR`. ### Why are the changes needed? We should assign proper name to _LEGACY_ERROR_TEMP_* ### Does this PR introduce _any_ user-facing change? No ### How was this patch tested? `./build/sbt "sql/testOnly org.apache.spark.sql.SQLQueryTestSuite*` Closes #39282 from itholic/LEGACY_1230. Authored-by: itholic <haejoon....@databricks.com> Signed-off-by: Max Gekk <max.g...@gmail.com> --- core/src/main/resources/error/error-classes.json | 5 ----- .../apache/spark/sql/errors/QueryCompilationErrors.scala | 10 ++++------ .../scala/org/apache/spark/sql/types/DecimalSuite.scala | 13 ++++++++----- 3 files changed, 12 insertions(+), 16 deletions(-) diff --git a/core/src/main/resources/error/error-classes.json b/core/src/main/resources/error/error-classes.json index 5409507c3c8..a3acb940585 100644 --- a/core/src/main/resources/error/error-classes.json +++ b/core/src/main/resources/error/error-classes.json @@ -2944,11 +2944,6 @@ "<decimalType> can only support precision up to <precision>." ] }, - "_LEGACY_ERROR_TEMP_1230" : { - "message" : [ - "Negative scale is not allowed: <scale>. You can use <config>=true to enable legacy mode to allow it." - ] - }, "_LEGACY_ERROR_TEMP_1231" : { "message" : [ "<key> is not a valid partition column in table <tblName>." diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryCompilationErrors.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryCompilationErrors.scala index 2ced0b8ac7a..25005a1f609 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryCompilationErrors.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryCompilationErrors.scala @@ -21,7 +21,7 @@ import scala.collection.mutable import org.apache.hadoop.fs.Path -import org.apache.spark.{SparkThrowable, SparkThrowableHelper} +import org.apache.spark.{SparkException, SparkThrowable, SparkThrowableHelper} import org.apache.spark.sql.AnalysisException import org.apache.spark.sql.catalyst.{FunctionIdentifier, QualifiedTableName, TableIdentifier} import org.apache.spark.sql.catalyst.analysis.{CannotReplaceMissingTableException, FunctionAlreadyExistsException, NamespaceAlreadyExistsException, NoSuchFunctionException, NoSuchNamespaceException, NoSuchPartitionException, NoSuchTableException, ResolvedTable, Star, TableAlreadyExistsException, UnresolvedRegex} @@ -2242,11 +2242,9 @@ private[sql] object QueryCompilationErrors extends QueryErrorsBase { } def negativeScaleNotAllowedError(scale: Int): Throwable = { - new AnalysisException( - errorClass = "_LEGACY_ERROR_TEMP_1230", - messageParameters = Map( - "scale" -> scale.toString, - "config" -> LEGACY_ALLOW_NEGATIVE_SCALE_OF_DECIMAL_ENABLED.key)) + SparkException.internalError(s"Negative scale is not allowed: ${scale.toString}." + + s" Set the config ${toSQLConf(LEGACY_ALLOW_NEGATIVE_SCALE_OF_DECIMAL_ENABLED.key)}" + + " to \"true\" to allow it.") } def invalidPartitionColumnKeyInTableError(key: String, tblName: String): Throwable = { diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DecimalSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DecimalSuite.scala index 73944d9dff9..465c25118fa 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DecimalSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DecimalSuite.scala @@ -19,8 +19,7 @@ package org.apache.spark.sql.types import org.scalatest.PrivateMethodTester -import org.apache.spark.{SparkArithmeticException, SparkFunSuite, SparkNumberFormatException} -import org.apache.spark.sql.AnalysisException +import org.apache.spark.{SparkArithmeticException, SparkException, SparkFunSuite, SparkNumberFormatException} import org.apache.spark.sql.catalyst.plans.SQLHelper import org.apache.spark.sql.internal.SQLConf import org.apache.spark.sql.types.Decimal._ @@ -111,9 +110,13 @@ class DecimalSuite extends SparkFunSuite with PrivateMethodTester with SQLHelper test("SPARK-30252: Negative scale is not allowed by default") { def checkNegativeScaleDecimal(d: => Decimal): Unit = { - intercept[AnalysisException](d) - .getMessage - .contains("Negative scale is not allowed under ansi mode") + checkError( + exception = intercept[SparkException] (d), + errorClass = "INTERNAL_ERROR", + parameters = Map("message" -> ("Negative scale is not allowed: -3. " + + "Set the config \"spark.sql.legacy.allowNegativeScaleOfDecimal\" " + + "to \"true\" to allow it.")) + ) } checkNegativeScaleDecimal(Decimal(BigDecimal("98765"), 5, -3)) checkNegativeScaleDecimal(Decimal(BigDecimal("98765").underlying(), 5, -3)) --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org