This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push: new 06997d6eb73 [SPARK-40152][SQL][TESTS] Move tests from SplitPart to elementAt 06997d6eb73 is described below commit 06997d6eb73f271aede5b159d86d1db80a73b89f Author: Yuming Wang <yumw...@ebay.com> AuthorDate: Wed Aug 24 13:33:26 2022 +0900 [SPARK-40152][SQL][TESTS] Move tests from SplitPart to elementAt ### What changes were proposed in this pull request? Move tests from SplitPart to elementAt in CollectionExpressionsSuite. ### Why are the changes needed? Simplify test. ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? N/A. Closes #37637 from wangyum/SPARK-40152-3. Authored-by: Yuming Wang <yumw...@ebay.com> Signed-off-by: Hyukjin Kwon <gurwls...@apache.org> --- .../expressions/CollectionExpressionsSuite.scala | 38 ++++++++++------------ 1 file changed, 18 insertions(+), 20 deletions(-) diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/CollectionExpressionsSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/CollectionExpressionsSuite.scala index 94cf0a74467..229e698fb2e 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/CollectionExpressionsSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/CollectionExpressionsSuite.scala @@ -1535,6 +1535,24 @@ class CollectionExpressionsSuite extends SparkFunSuite with ExpressionEvalHelper } checkEvaluation(ElementAt(mb0, Literal(Array[Byte](2, 1), BinaryType)), "2") checkEvaluation(ElementAt(mb0, Literal(Array[Byte](3, 4))), null) + + // test defaultValueOutOfBound + val delimiter = Literal.create(".", StringType) + val str = StringSplitSQL(Literal.create("11.12.13", StringType), delimiter) + val outOfBoundValue = Some(Literal.create("", StringType)) + + checkEvaluation(ElementAt(str, Literal(3), outOfBoundValue), UTF8String.fromString("13")) + checkEvaluation(ElementAt(str, Literal(1), outOfBoundValue), UTF8String.fromString("11")) + checkEvaluation(ElementAt(str, Literal(10), outOfBoundValue), UTF8String.fromString("")) + checkEvaluation(ElementAt(str, Literal(-10), outOfBoundValue), UTF8String.fromString("")) + + checkEvaluation(ElementAt(StringSplitSQL(Literal.create(null, StringType), delimiter), + Literal(1), outOfBoundValue), null) + checkEvaluation(ElementAt(StringSplitSQL(Literal.create("11.12.13", StringType), + Literal.create(null, StringType)), Literal(1), outOfBoundValue), null) + + checkExceptionInExpression[Exception]( + ElementAt(str, Literal(0), outOfBoundValue), "The index 0 is invalid") } test("correctly handles ElementAt nullability for arrays") { @@ -2522,24 +2540,4 @@ class CollectionExpressionsSuite extends SparkFunSuite with ExpressionEvalHelper Date.valueOf("2017-02-12"))) } } - - test("SplitPart") { - val delimiter = Literal.create(".", StringType) - val str = StringSplitSQL(Literal.create("11.12.13", StringType), delimiter) - val outOfBoundValue = Some(Literal.create("", StringType)) - - checkEvaluation(ElementAt(str, Literal(3), outOfBoundValue), UTF8String.fromString("13")) - checkEvaluation(ElementAt(str, Literal(1), outOfBoundValue), UTF8String.fromString("11")) - checkEvaluation(ElementAt(str, Literal(10), outOfBoundValue), UTF8String.fromString("")) - checkEvaluation(ElementAt(str, Literal(-10), outOfBoundValue), UTF8String.fromString("")) - - checkEvaluation(ElementAt(StringSplitSQL(Literal.create(null, StringType), delimiter), - Literal(1), outOfBoundValue), null) - checkEvaluation(ElementAt(StringSplitSQL(Literal.create("11.12.13", StringType), - Literal.create(null, StringType)), Literal(1), outOfBoundValue), null) - - intercept[Exception] { - checkEvaluation(ElementAt(str, Literal(0), outOfBoundValue), null) - }.getMessage.contains("The index 0 is invalid") - } } --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org