This is an automated email from the ASF dual-hosted git repository.
csy pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/auron.git
The following commit(s) were added to refs/heads/master by this push:
new 85352137 [AURON #1751] rename scalarTypeSupported -> isTypeSupported
(#1753)
85352137 is described below
commit 8535213789ac681c77f1d76fbf0d117a8f9e5ef7
Author: James Xu <[email protected]>
AuthorDate: Mon Dec 15 12:51:32 2025 +0800
[AURON #1751] rename scalarTypeSupported -> isTypeSupported (#1753)
# Which issue does this PR close?
Closes #1751
# Rationale for this change
# What changes are included in this PR?
# Are there any user-facing changes?
No.
# How was this patch tested?
Existing UTs.
---
.../scala/org/apache/spark/sql/auron/AuronConverters.scala | 4 ++--
.../scala/org/apache/spark/sql/auron/NativeConverters.scala | 11 +++++++----
2 files changed, 9 insertions(+), 6 deletions(-)
diff --git
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala
index 5d348528..4e39198d 100644
---
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala
+++
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala
@@ -28,7 +28,7 @@ import org.apache.spark.Partition
import org.apache.spark.broadcast.Broadcast
import org.apache.spark.internal.{config, Logging}
import
org.apache.spark.sql.auron.AuronConvertStrategy.{childOrderingRequiredTag,
convertibleTag, convertStrategyTag, convertToNonNativeTag, isNeverConvert,
joinSmallerSideTag, neverConvertReasonTag}
-import org.apache.spark.sql.auron.NativeConverters.{existTimestampType,
roundRobinTypeSupported, scalarTypeSupported, StubExpr}
+import org.apache.spark.sql.auron.NativeConverters.{existTimestampType,
isTypeSupported, roundRobinTypeSupported, StubExpr}
import org.apache.spark.sql.auron.util.AuronLogUtils.logDebugPlanConversion
import org.apache.spark.sql.catalyst.expressions.AggregateWindowFunction
import org.apache.spark.sql.catalyst.expressions.Alias
@@ -403,7 +403,7 @@ object AuronConverters extends Logging {
outputPartitioning match {
case partitioning: RangePartitioning =>
val unsupportedOrderType = partitioning.ordering
- .find(e => !scalarTypeSupported(e.dataType))
+ .find(e => !isTypeSupported(e.dataType))
assert(
unsupportedOrderType.isEmpty,
s"Unsupported order type in range partitioning:
${unsupportedOrderType.get}")
diff --git
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
index 153e2b8e..4838804a 100644
---
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
+++
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
@@ -102,7 +102,10 @@ object NativeConverters extends Logging {
def castTrimStringEnabled: Boolean =
AuronConverters.getBooleanConf("spark.auron.cast.trimString", defaultValue
= true)
- def scalarTypeSupported(dataType: DataType): Boolean = {
+ /**
+ * Is the data type(scalar or complex) supported by Auron.
+ */
+ def isTypeSupported(dataType: DataType): Boolean = {
dataType match {
case NullType | BooleanType | ByteType | ShortType | IntegerType |
LongType | FloatType |
DoubleType | StringType | DateType | TimestampType =>
@@ -111,11 +114,11 @@ object NativeConverters extends Logging {
case _: DecimalType =>
// auron only supports 64-bit decimal
false
- case at: ArrayType => scalarTypeSupported(at.elementType)
+ case at: ArrayType => isTypeSupported(at.elementType)
case m: MapType =>
- scalarTypeSupported(m.keyType) && scalarTypeSupported(m.valueType)
+ isTypeSupported(m.keyType) && isTypeSupported(m.valueType)
case s: StructType =>
- s.fields.forall(e => scalarTypeSupported(e.dataType))
+ s.fields.forall(e => isTypeSupported(e.dataType))
case _ => false
}
}