This is an automated email from the ASF dual-hosted git repository.

csy pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/auron.git


The following commit(s) were added to refs/heads/master by this push:
     new 85352137 [AURON #1751] rename scalarTypeSupported -> isTypeSupported 
(#1753)
85352137 is described below

commit 8535213789ac681c77f1d76fbf0d117a8f9e5ef7
Author: James Xu <[email protected]>
AuthorDate: Mon Dec 15 12:51:32 2025 +0800

    [AURON #1751] rename scalarTypeSupported -> isTypeSupported (#1753)
    
    # Which issue does this PR close?
    
    Closes #1751
    
    # Rationale for this change
    
    # What changes are included in this PR?
    
    # Are there any user-facing changes?
    
    No.
    
    # How was this patch tested?
    
    Existing UTs.
---
 .../scala/org/apache/spark/sql/auron/AuronConverters.scala    |  4 ++--
 .../scala/org/apache/spark/sql/auron/NativeConverters.scala   | 11 +++++++----
 2 files changed, 9 insertions(+), 6 deletions(-)

diff --git 
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala
 
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala
index 5d348528..4e39198d 100644
--- 
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala
+++ 
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala
@@ -28,7 +28,7 @@ import org.apache.spark.Partition
 import org.apache.spark.broadcast.Broadcast
 import org.apache.spark.internal.{config, Logging}
 import 
org.apache.spark.sql.auron.AuronConvertStrategy.{childOrderingRequiredTag, 
convertibleTag, convertStrategyTag, convertToNonNativeTag, isNeverConvert, 
joinSmallerSideTag, neverConvertReasonTag}
-import org.apache.spark.sql.auron.NativeConverters.{existTimestampType, 
roundRobinTypeSupported, scalarTypeSupported, StubExpr}
+import org.apache.spark.sql.auron.NativeConverters.{existTimestampType, 
isTypeSupported, roundRobinTypeSupported, StubExpr}
 import org.apache.spark.sql.auron.util.AuronLogUtils.logDebugPlanConversion
 import org.apache.spark.sql.catalyst.expressions.AggregateWindowFunction
 import org.apache.spark.sql.catalyst.expressions.Alias
@@ -403,7 +403,7 @@ object AuronConverters extends Logging {
     outputPartitioning match {
       case partitioning: RangePartitioning =>
         val unsupportedOrderType = partitioning.ordering
-          .find(e => !scalarTypeSupported(e.dataType))
+          .find(e => !isTypeSupported(e.dataType))
         assert(
           unsupportedOrderType.isEmpty,
           s"Unsupported order type in range partitioning: 
${unsupportedOrderType.get}")
diff --git 
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
 
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
index 153e2b8e..4838804a 100644
--- 
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
+++ 
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
@@ -102,7 +102,10 @@ object NativeConverters extends Logging {
   def castTrimStringEnabled: Boolean =
     AuronConverters.getBooleanConf("spark.auron.cast.trimString", defaultValue 
= true)
 
-  def scalarTypeSupported(dataType: DataType): Boolean = {
+  /**
+   * Is the data type(scalar or complex) supported by Auron.
+   */
+  def isTypeSupported(dataType: DataType): Boolean = {
     dataType match {
       case NullType | BooleanType | ByteType | ShortType | IntegerType | 
LongType | FloatType |
           DoubleType | StringType | DateType | TimestampType =>
@@ -111,11 +114,11 @@ object NativeConverters extends Logging {
       case _: DecimalType =>
         // auron only supports 64-bit decimal
         false
-      case at: ArrayType => scalarTypeSupported(at.elementType)
+      case at: ArrayType => isTypeSupported(at.elementType)
       case m: MapType =>
-        scalarTypeSupported(m.keyType) && scalarTypeSupported(m.valueType)
+        isTypeSupported(m.keyType) && isTypeSupported(m.valueType)
       case s: StructType =>
-        s.fields.forall(e => scalarTypeSupported(e.dataType))
+        s.fields.forall(e => isTypeSupported(e.dataType))
       case _ => false
     }
   }

Reply via email to