(spark) branch master updated (dfb35bed522c -> c9435b8b4864)
This is an automated email from the ASF dual-hosted git repository. wenchen pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git from dfb35bed522c [SPARK-47146][CORE] Possible thread leak when doing sort merge join add c9435b8b4864 [SPARK-47176][SQL][FOLLOW-UP] resolveExpressions should have three versions which is the same as resolveOperators No new revisions were added by this update. Summary of changes: .../catalyst/plans/logical/AnalysisHelper.scala| 36 ++ 1 file changed, 36 insertions(+) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
(spark) branch master updated: [SPARK-47146][CORE] Possible thread leak when doing sort merge join
This is an automated email from the ASF dual-hosted git repository. mridulm80 pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/master by this push: new dfb35bed522c [SPARK-47146][CORE] Possible thread leak when doing sort merge join dfb35bed522c is described below commit dfb35bed522ca706f8fc18e37c05c1766c8d8a18 Author: JacobZheng0927 AuthorDate: Mon Mar 4 23:17:32 2024 -0600 [SPARK-47146][CORE] Possible thread leak when doing sort merge join ### What changes were proposed in this pull request? Add TaskCompletionListener to close inputStream to avoid thread leakage caused by unclosed ReadAheadInputStream. ### Why are the changes needed? SPARK-40849 modified the implementation of `newDaemonSingleThreadExecutor` to use `newFixedThreadPool` instead of `newSingleThreadExecutor` .The difference is that `newSingleThreadExecutor` uses the `FinalizableDelegatedExecutorService`, which provides a `finalize` method that automatically closes the thread pool. In some cases, sort merge join execution uses `ReadAheadSteam` and does not close it, so this change caused a thread leak. Since Finalization is deprecated and subject to re [...] ### Does this PR introduce _any_ user-facing change? No ### How was this patch tested? Unit test ### Was this patch authored or co-authored using generative AI tooling? No Closes #45327 from JacobZheng0927/SPARK-47146. Authored-by: JacobZheng0927 Signed-off-by: Mridul Muralidharan gmail.com> --- .../unsafe/sort/UnsafeSorterSpillReader.java | 12 .../scala/org/apache/spark/sql/JoinSuite.scala | 33 +- 2 files changed, 44 insertions(+), 1 deletion(-) diff --git a/core/src/main/java/org/apache/spark/util/collection/unsafe/sort/UnsafeSorterSpillReader.java b/core/src/main/java/org/apache/spark/util/collection/unsafe/sort/UnsafeSorterSpillReader.java index db79efd00853..8bd44c8c52c1 100644 --- a/core/src/main/java/org/apache/spark/util/collection/unsafe/sort/UnsafeSorterSpillReader.java +++ b/core/src/main/java/org/apache/spark/util/collection/unsafe/sort/UnsafeSorterSpillReader.java @@ -28,6 +28,8 @@ import org.apache.spark.io.ReadAheadInputStream; import org.apache.spark.serializer.SerializerManager; import org.apache.spark.storage.BlockId; import org.apache.spark.unsafe.Platform; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import java.io.*; @@ -36,6 +38,7 @@ import java.io.*; * of the file format). */ public final class UnsafeSorterSpillReader extends UnsafeSorterIterator implements Closeable { + private static final Logger logger = LoggerFactory.getLogger(ReadAheadInputStream.class); public static final int MAX_BUFFER_SIZE_BYTES = 16777216; // 16 mb private InputStream in; @@ -82,6 +85,15 @@ public final class UnsafeSorterSpillReader extends UnsafeSorterIterator implemen Closeables.close(bs, /* swallowIOException = */ true); throw e; } +if (taskContext != null) { + taskContext.addTaskCompletionListener(context -> { +try { + close(); +} catch (IOException e) { + logger.info("error while closing UnsafeSorterSpillReader", e); +} + }); +} } @Override diff --git a/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala index f31f60e8df56..be6862f5b96b 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala @@ -23,6 +23,7 @@ import scala.collection.mutable.ListBuffer import scala.jdk.CollectionConverters._ import org.apache.spark.TestUtils.{assertNotSpilled, assertSpilled} +import org.apache.spark.internal.config.SHUFFLE_SPILL_NUM_ELEMENTS_FORCE_SPILL_THRESHOLD import org.apache.spark.sql.catalyst.TableIdentifier import org.apache.spark.sql.catalyst.analysis.UnresolvedRelation import org.apache.spark.sql.catalyst.expressions.{Ascending, GenericRow, SortOrder} @@ -34,7 +35,7 @@ import org.apache.spark.sql.execution.exchange.{ShuffleExchangeExec, ShuffleExch import org.apache.spark.sql.execution.joins._ import org.apache.spark.sql.execution.python.BatchEvalPythonExec import org.apache.spark.sql.internal.SQLConf -import org.apache.spark.sql.test.SharedSparkSession +import org.apache.spark.sql.test.{SharedSparkSession, TestSparkSession} import org.apache.spark.sql.types.StructType import org.apache.spark.tags.SlowSQLTest @@ -1737,3 +1738,33 @@ class JoinSuite extends QueryTest with SharedSparkSession with AdaptiveSparkPlan } } } + +class ThreadLeakInSortMergeJoinSuite + extends QueryTest +with SharedSparkSession +with AdaptiveSparkPlanHelper { + + setupTestData() + override protected def createSparkSession: TestSparkSession = { +
(spark) branch branch-3.5 updated: [SPARK-47177][SQL] Cached SQL plan do not display final AQE plan in explain string
This is an automated email from the ASF dual-hosted git repository. ulyssesyou pushed a commit to branch branch-3.5 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.5 by this push: new 14762b372dd6 [SPARK-47177][SQL] Cached SQL plan do not display final AQE plan in explain string 14762b372dd6 is described below commit 14762b372dd623179aa2c985c44cd49048660dda Author: ulysses-you AuthorDate: Tue Mar 5 10:13:00 2024 +0800 [SPARK-47177][SQL] Cached SQL plan do not display final AQE plan in explain string ### What changes were proposed in this pull request? This pr adds lock for ExplainUtils.processPlan to avoid tag race condition. ### Why are the changes needed? To fix the issue [SPARK-47177](https://issues.apache.org/jira/browse/SPARK-47177) ### Does this PR introduce _any_ user-facing change? yes, affect plan explain ### How was this patch tested? add test ### Was this patch authored or co-authored using generative AI tooling? no Closes #45282 from ulysses-you/SPARK-47177. Authored-by: ulysses-you Signed-off-by: youxiduo (cherry picked from commit 6e62a5690b810edb99e4fc6ad39afbd4d49ef85e) Signed-off-by: youxiduo --- .../apache/spark/sql/catalyst/trees/TreeNode.scala | 7 ++-- .../apache/spark/sql/execution/ExplainUtils.scala | 6 +++- .../sql/execution/columnar/InMemoryRelation.scala | 12 +-- .../execution/columnar/InMemoryRelationSuite.scala | 41 +++--- 4 files changed, 38 insertions(+), 28 deletions(-) diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/trees/TreeNode.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/trees/TreeNode.scala index 9e605a45414b..82228a5b2aaf 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/trees/TreeNode.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/trees/TreeNode.scala @@ -1030,10 +1030,11 @@ abstract class TreeNode[BaseType <: TreeNode[BaseType]] append(str) append("\n") -if (innerChildren.nonEmpty) { +val innerChildrenLocal = innerChildren +if (innerChildrenLocal.nonEmpty) { lastChildren.add(children.isEmpty) lastChildren.add(false) - innerChildren.init.foreach(_.generateTreeString( + innerChildrenLocal.init.foreach(_.generateTreeString( depth + 2, lastChildren, append, verbose, addSuffix = addSuffix, maxFields = maxFields, printNodeId = printNodeId, indent = indent)) lastChildren.remove(lastChildren.size() - 1) @@ -1041,7 +1042,7 @@ abstract class TreeNode[BaseType <: TreeNode[BaseType]] lastChildren.add(children.isEmpty) lastChildren.add(true) - innerChildren.last.generateTreeString( + innerChildrenLocal.last.generateTreeString( depth + 2, lastChildren, append, verbose, addSuffix = addSuffix, maxFields = maxFields, printNodeId = printNodeId, indent = indent) lastChildren.remove(lastChildren.size() - 1) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/ExplainUtils.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/ExplainUtils.scala index 3da3e646f36b..11f6ae0e47ee 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/ExplainUtils.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/ExplainUtils.scala @@ -75,8 +75,12 @@ object ExplainUtils extends AdaptiveSparkPlanHelper { * Given a input physical plan, performs the following tasks. * 1. Generates the explain output for the input plan excluding the subquery plans. * 2. Generates the explain output for each subquery referenced in the plan. + * + * Note that, ideally this is a no-op as different explain actions operate on different plan, + * instances but cached plan is an exception. The `InMemoryRelation#innerChildren` use a shared + * plan instance across multi-queries. Add lock for this method to avoid tag race condition. */ - def processPlan[T <: QueryPlan[T]](plan: T, append: String => Unit): Unit = { + def processPlan[T <: QueryPlan[T]](plan: T, append: String => Unit): Unit = synchronized { try { // Initialize a reference-unique set of Operators to avoid accdiental overwrites and to allow // intentional overwriting of IDs generated in previous AQE iteration diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/columnar/InMemoryRelation.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/columnar/InMemoryRelation.scala index 5bab8e53eb16..f750a4503be1 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/columnar/InMemoryRelation.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/columnar/InMemoryRelation.scala @@ -392,17 +392,7 @@ case class InMemoryRelation( @volatile var statsOfPlanToCache:
(spark) branch master updated (73aa14466683 -> 6e62a5690b81)
This is an automated email from the ASF dual-hosted git repository. ulyssesyou pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git from 73aa14466683 [SPARK-47252][DOCS] Clarify that pivot may trigger an eager computation add 6e62a5690b81 [SPARK-47177][SQL] Cached SQL plan do not display final AQE plan in explain string No new revisions were added by this update. Summary of changes: .../apache/spark/sql/catalyst/trees/TreeNode.scala | 7 ++-- .../apache/spark/sql/execution/ExplainUtils.scala | 6 +++- .../sql/execution/columnar/InMemoryRelation.scala | 12 +-- .../execution/columnar/InMemoryRelationSuite.scala | 41 +++--- 4 files changed, 38 insertions(+), 28 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
(spark) branch master updated (35bced42474e -> 73aa14466683)
This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git from 35bced42474e [SPARK-47242][BUILD] Bump ap-loader 3.0(v8) to support for async-profiler 3.0 add 73aa14466683 [SPARK-47252][DOCS] Clarify that pivot may trigger an eager computation No new revisions were added by this update. Summary of changes: .../apache/spark/sql/RelationalGroupedDataset.scala| 18 +- python/pyspark/sql/group.py| 10 +- .../apache/spark/sql/RelationalGroupedDataset.scala| 16 3 files changed, 22 insertions(+), 22 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
(spark) branch master updated (db0e5c7bc464 -> 35bced42474e)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git from db0e5c7bc464 [SPARK-47269][BUILD] Upgrade jetty to 11.0.20 add 35bced42474e [SPARK-47242][BUILD] Bump ap-loader 3.0(v8) to support for async-profiler 3.0 No new revisions were added by this update. Summary of changes: connector/profiler/pom.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
(spark) branch master updated (6b5917beff30 -> db0e5c7bc464)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git from 6b5917beff30 [SPARK-46961][SS] Using ProcessorContext to store and retrieve handle add db0e5c7bc464 [SPARK-47269][BUILD] Upgrade jetty to 11.0.20 No new revisions were added by this update. Summary of changes: dev/deps/spark-deps-hadoop-3-hive-2.3 | 4 ++-- pom.xml | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
(spark) branch master updated: [SPARK-46961][SS] Using ProcessorContext to store and retrieve handle
This is an automated email from the ASF dual-hosted git repository. maxgekk pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/master by this push: new 6b5917beff30 [SPARK-46961][SS] Using ProcessorContext to store and retrieve handle 6b5917beff30 is described below commit 6b5917beff30c813a362584a135a587001df1390 Author: Eric Marnadi AuthorDate: Mon Mar 4 21:20:23 2024 +0300 [SPARK-46961][SS] Using ProcessorContext to store and retrieve handle ### What changes were proposed in this pull request? Setting the processorHandle as a part of the statefulProcessor, so that the user doesn't have to explicitly keep track of it, and can instead simply call `getStatefulProcessorHandle` ### Why are the changes needed? This enhances the usability of the State API ### Does this PR introduce _any_ user-facing change? Yes, this is an API change. This enhances usability of the StatefulProcessorHandle and the TransformWithState operator. ### How was this patch tested? Existing unit tests are sufficient ### Was this patch authored or co-authored using generative AI tooling? No Closes #45359 from ericm-db/handle-context. Authored-by: Eric Marnadi Signed-off-by: Max Gekk --- .../src/main/resources/error/error-classes.json| 7 +++ docs/sql-error-conditions.md | 7 +++ .../apache/spark/sql/errors/ExecutionErrors.scala | 6 +++ .../spark/sql/streaming/StatefulProcessor.scala| 38 --- .../streaming/TransformWithStateExec.scala | 4 +- .../streaming/TransformWithListStateSuite.scala| 14 ++ .../sql/streaming/TransformWithStateSuite.scala| 54 ++ 7 files changed, 84 insertions(+), 46 deletions(-) diff --git a/common/utils/src/main/resources/error/error-classes.json b/common/utils/src/main/resources/error/error-classes.json index 6ccd841ccd0f..7cf3e9c533ca 100644 --- a/common/utils/src/main/resources/error/error-classes.json +++ b/common/utils/src/main/resources/error/error-classes.json @@ -3337,6 +3337,13 @@ ], "sqlState" : "42802" }, + "STATE_STORE_HANDLE_NOT_INITIALIZED" : { +"message" : [ + "The handle has not been initialized for this StatefulProcessor.", + "Please only use the StatefulProcessor within the transformWithState operator." +], +"sqlState" : "42802" + }, "STATE_STORE_MULTIPLE_VALUES_PER_KEY" : { "message" : [ "Store does not support multiple values per key" diff --git a/docs/sql-error-conditions.md b/docs/sql-error-conditions.md index f026c456eb2d..7be01f8cb513 100644 --- a/docs/sql-error-conditions.md +++ b/docs/sql-error-conditions.md @@ -2091,6 +2091,13 @@ Star (*) is not allowed in a select list when GROUP BY an ordinal position is us Failed to remove default column family with reserved name=``. +### STATE_STORE_HANDLE_NOT_INITIALIZED + +[SQLSTATE: 42802](sql-error-conditions-sqlstates.html#class-42-syntax-error-or-access-rule-violation) + +The handle has not been initialized for this StatefulProcessor. +Please only use the StatefulProcessor within the transformWithState operator. + ### STATE_STORE_MULTIPLE_VALUES_PER_KEY [SQLSTATE: 42802](sql-error-conditions-sqlstates.html#class-42-syntax-error-or-access-rule-violation) diff --git a/sql/api/src/main/scala/org/apache/spark/sql/errors/ExecutionErrors.scala b/sql/api/src/main/scala/org/apache/spark/sql/errors/ExecutionErrors.scala index b74a67b49bda..7910c386fcf1 100644 --- a/sql/api/src/main/scala/org/apache/spark/sql/errors/ExecutionErrors.scala +++ b/sql/api/src/main/scala/org/apache/spark/sql/errors/ExecutionErrors.scala @@ -53,6 +53,12 @@ private[sql] trait ExecutionErrors extends DataTypeErrorsBase { e) } + def stateStoreHandleNotInitialized(): SparkRuntimeException = { +new SparkRuntimeException( + errorClass = "STATE_STORE_HANDLE_NOT_INITIALIZED", + messageParameters = Map.empty) + } + def failToRecognizePatternAfterUpgradeError( pattern: String, e: Throwable): SparkUpgradeException = { new SparkUpgradeException( diff --git a/sql/api/src/main/scala/org/apache/spark/sql/streaming/StatefulProcessor.scala b/sql/api/src/main/scala/org/apache/spark/sql/streaming/StatefulProcessor.scala index 76794136dd49..42a9430bf39d 100644 --- a/sql/api/src/main/scala/org/apache/spark/sql/streaming/StatefulProcessor.scala +++ b/sql/api/src/main/scala/org/apache/spark/sql/streaming/StatefulProcessor.scala @@ -20,6 +20,7 @@ package org.apache.spark.sql.streaming import java.io.Serializable import org.apache.spark.annotation.{Evolving, Experimental} +import org.apache.spark.sql.errors.ExecutionErrors /** * Represents the arbitrary stateful logic that needs to be provided by the user to perform @@ -29,17 +30,18 @@ import
(spark) branch master updated: [SPARK-47245][SQL] Improve error code for INVALID_PARTITION_COLUMN_DATA_TYPE
This is an automated email from the ASF dual-hosted git repository. maxgekk pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/master by this push: new 0b2eca28c307 [SPARK-47245][SQL] Improve error code for INVALID_PARTITION_COLUMN_DATA_TYPE 0b2eca28c307 is described below commit 0b2eca28c307b4feb4edd1f53c9aecc64523b7eb Author: Stefan Kandic AuthorDate: Mon Mar 4 19:58:52 2024 +0300 [SPARK-47245][SQL] Improve error code for INVALID_PARTITION_COLUMN_DATA_TYPE ### What changes were proposed in this pull request? Improving the error code for error class `INVALID_PARTITION_COLUMN_DATA_TYPE`. ### Why are the changes needed? `0A000` means a feature is not supported, It implies that in some future it may be and the user hit a limit rather than just writing something inherently wrong. ### Does this PR introduce _any_ user-facing change? Yes, new sql error code ### How was this patch tested? UTs ### Was this patch authored or co-authored using generative AI tooling? No Closes #45355 from stefankandic/SPARK-47245-improveErrorCode. Authored-by: Stefan Kandic Signed-off-by: Max Gekk --- common/utils/src/main/resources/error/error-classes.json | 2 +- docs/sql-error-conditions.md | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/common/utils/src/main/resources/error/error-classes.json b/common/utils/src/main/resources/error/error-classes.json index 493635d1f8d3..6ccd841ccd0f 100644 --- a/common/utils/src/main/resources/error/error-classes.json +++ b/common/utils/src/main/resources/error/error-classes.json @@ -2325,7 +2325,7 @@ "message" : [ "Cannot use for partition column." ], -"sqlState" : "42601" +"sqlState" : "0A000" }, "INVALID_PARTITION_OPERATION" : { "message" : [ diff --git a/docs/sql-error-conditions.md b/docs/sql-error-conditions.md index 510f56f413c6..f026c456eb2d 100644 --- a/docs/sql-error-conditions.md +++ b/docs/sql-error-conditions.md @@ -1320,7 +1320,7 @@ For more details see [INVALID_PARAMETER_VALUE](sql-error-conditions-invalid-para ### INVALID_PARTITION_COLUMN_DATA_TYPE -[SQLSTATE: 42601](sql-error-conditions-sqlstates.html#class-42-syntax-error-or-access-rule-violation) +[SQLSTATE: 0A000](sql-error-conditions-sqlstates.html#class-0A-feature-not-supported) Cannot use `` for partition column. - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
(spark) branch master updated (7e7f3ff9e281 -> ca7c60b49988)
This is an automated email from the ASF dual-hosted git repository. wenchen pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git from 7e7f3ff9e281 [SPARK-47070] Fix invalid aggregation after subquery rewrite add ca7c60b49988 [SPARK-47268][SQL][COLLATIONS] Support for repartition with collations No new revisions were added by this update. Summary of changes: .../spark/sql/catalyst/expressions/hash.scala | 43 -- .../org/apache/spark/sql/CollationSuite.scala | 21 ++- 2 files changed, 52 insertions(+), 12 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
(spark) branch master updated (479954cf73a5 -> 7e7f3ff9e281)
This is an automated email from the ASF dual-hosted git repository. wenchen pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git from 479954cf73a5 [SPARK-47131][SQL][COLLATION] String function support: contains, startswith, endswith add 7e7f3ff9e281 [SPARK-47070] Fix invalid aggregation after subquery rewrite No new revisions were added by this update. Summary of changes: .../spark/sql/catalyst/optimizer/subquery.scala| 65 +- .../exists-subquery/exists-aggregate.sql.out | 131 .../subquery/in-subquery/in-group-by.sql.out | 237 + .../subquery/exists-subquery/exists-aggregate.sql | 39 +++- .../inputs/subquery/in-subquery/in-group-by.sql| 71 ++ .../exists-subquery/exists-aggregate.sql.out | 83 .../subquery/in-subquery/in-group-by.sql.out | 117 ++ 7 files changed, 736 insertions(+), 7 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
(spark) branch master updated (b711efd6671a -> 479954cf73a5)
This is an automated email from the ASF dual-hosted git repository. wenchen pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git from b711efd6671a [SPARK-43258][SQL] Assign names to error _LEGACY_ERROR_TEMP_202[3,5] add 479954cf73a5 [SPARK-47131][SQL][COLLATION] String function support: contains, startswith, endswith No new revisions were added by this update. Summary of changes: .../spark/sql/catalyst/util/CollationFactory.java | 3 + .../org/apache/spark/unsafe/types/UTF8String.java | 44 + .../src/main/resources/error/error-classes.json| 18 ++ ...ror-conditions-datatype-mismatch-error-class.md | 4 + ...onditions-unsupported-collation-error-class.md} | 14 +- docs/sql-error-conditions.md | 8 + .../org/apache/spark/sql/types/StringType.scala| 3 +- .../catalyst/expressions/stringExpressions.scala | 89 - .../sql/catalyst/types/PhysicalDataType.scala | 4 +- .../spark/sql/execution/columnar/ColumnType.scala | 3 +- .../org/apache/spark/sql/CollationSuite.scala | 216 - 11 files changed, 379 insertions(+), 27 deletions(-) copy docs/{sql-error-conditions-unsupported-save-mode-error-class.md => sql-error-conditions-unsupported-collation-error-class.md} (82%) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
(spark) branch master updated (a1e57b32e495 -> b711efd6671a)
This is an automated email from the ASF dual-hosted git repository. maxgekk pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git from a1e57b32e495 [SPARK-47266][CONNECT] Make `ProtoUtils.abbreviate` return the same type as the input add b711efd6671a [SPARK-43258][SQL] Assign names to error _LEGACY_ERROR_TEMP_202[3,5] No new revisions were added by this update. Summary of changes: .../src/main/resources/error/error-classes.json| 22 +- docs/sql-error-conditions.md | 12 ++ .../sql/catalyst/encoders/ExpressionEncoder.scala | 2 +- .../sql/catalyst/expressions/Expression.scala | 12 +++--- .../sql/catalyst/expressions/arithmetic.scala | 4 +- .../expressions/higherOrderFunctions.scala | 2 +- .../spark/sql/errors/QueryExecutionErrors.scala| 6 +-- .../sql/errors/QueryExecutionErrorsSuite.scala | 50 -- 8 files changed, 84 insertions(+), 26 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
(spark) branch master updated (69743ad98f34 -> a1e57b32e495)
This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git from 69743ad98f34 [SPARK-46094][CORE][FOLLOW-UP] Mark `writing` as volatile at ExecutorJVMProfiler add a1e57b32e495 [SPARK-47266][CONNECT] Make `ProtoUtils.abbreviate` return the same type as the input No new revisions were added by this update. Summary of changes: .../spark/sql/connect/common/ProtoUtils.scala | 8 ++-- .../sql/connect/messages/AbbreviateSuite.scala | 43 +- 2 files changed, 21 insertions(+), 30 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org