This is an automated email from the ASF dual-hosted git repository.

csy pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/auron.git


The following commit(s) were added to refs/heads/master by this push:
     new 4964053c [AURON #1870][BUILD] Add `-Xfatal-warnings` to 
scala-maven-plugin and fix related compilation failures (#1872)
4964053c is described below

commit 4964053c12888848956dcf098d2cca09ad827c84
Author: yew1eb <[email protected]>
AuthorDate: Wed Jan 28 19:40:57 2026 +0800

    [AURON #1870][BUILD] Add `-Xfatal-warnings` to scala-maven-plugin and fix 
related compilation failures (#1872)
    
    # Which issue does this PR close?
    
    Closes #1870
    
    # Rationale for this change
    
    # What changes are included in this PR?
    
    # Are there any user-facing changes?
    
    # How was this patch tested?
---
 .../org/apache/spark/sql/SparkQueryTestsBase.scala      |  2 +-
 .../org/apache/spark/sql/AuronDataFrameSuite.scala      |  8 ++++----
 .../sql/execution/ui/AuronSQLAppStatusListener.scala    |  3 +++
 pom.xml                                                 | 17 +++++++++++++++++
 .../spark/sql/auron/InterceptedValidateSparkPlan.scala  |  3 +++
 .../scala/org/apache/spark/sql/auron/ShimsImpl.scala    |  5 +++++
 .../auron/shuffle/AuronBlockStoreShuffleReader.scala    |  3 +++
 .../auron/shuffle/AuronRssShuffleManagerBase.scala      |  3 +++
 .../auron/plan/NativeShuffledHashJoinExecProvider.scala |  4 ++++
 .../scala/org/apache/auron/AuronFunctionSuite.scala     |  4 ----
 .../test/scala/org/apache/auron/AuronQuerySuite.scala   | 15 ---------------
 .../apache/spark/sql/auron/AuronCallNativeWrapper.scala |  3 +++
 .../org/apache/spark/sql/auron/NativeConverters.scala   |  4 ++--
 .../scala/org/apache/spark/sql/auron/NativeHelper.scala |  2 +-
 .../main/scala/org/apache/spark/sql/auron/Shims.scala   |  3 +++
 .../auron/columnar/AuronArrowColumnVector.scala         |  3 +++
 .../execution/auron/plan/NativeParquetSinkBase.scala    |  6 ++++--
 .../auron/plan/NativeShuffleExchangeBase.scala          |  2 ++
 .../auron/plan/NativeShuffledHashJoinBase.scala         |  2 ++
 .../sql/execution/auron/plan/NativeUnionBase.scala      |  4 ++--
 .../shuffle/AuronBlockStoreShuffleReaderBase.scala      |  2 +-
 .../src/main/scala/org/apache/auron/sparkver.scala      |  6 ++++--
 .../shuffle/uniffle/AuronUniffleShuffleReader.scala     |  3 +++
 23 files changed, 73 insertions(+), 34 deletions(-)

diff --git 
a/auron-spark-tests/common/src/test/scala/org/apache/spark/sql/SparkQueryTestsBase.scala
 
b/auron-spark-tests/common/src/test/scala/org/apache/spark/sql/SparkQueryTestsBase.scala
index 6c6d839a..8498edcd 100644
--- 
a/auron-spark-tests/common/src/test/scala/org/apache/spark/sql/SparkQueryTestsBase.scala
+++ 
b/auron-spark-tests/common/src/test/scala/org/apache/spark/sql/SparkQueryTestsBase.scala
@@ -162,7 +162,7 @@ object AuronQueryTestUtil extends Assertions {
   private def genError(
       expectedAnswer: Seq[Row],
       sparkAnswer: Seq[Row],
-      isSorted: Boolean = false): String = {
+      isSorted: Boolean): String = {
     val getRowType: Option[Row] => String = row =>
       row
         .map(row =>
diff --git 
a/auron-spark-tests/spark33/src/test/scala/org/apache/spark/sql/AuronDataFrameSuite.scala
 
b/auron-spark-tests/spark33/src/test/scala/org/apache/spark/sql/AuronDataFrameSuite.scala
index 74e27698..b5374551 100644
--- 
a/auron-spark-tests/spark33/src/test/scala/org/apache/spark/sql/AuronDataFrameSuite.scala
+++ 
b/auron-spark-tests/spark33/src/test/scala/org/apache/spark/sql/AuronDataFrameSuite.scala
@@ -186,22 +186,22 @@ class AuronDataFrameSuite extends DataFrameSuite with 
SparkQueryTestsBase {
       val join = df.join(df, "id")
       checkAnswer(join, df)
       val shuffleCount = collect(join.queryExecution.executedPlan) {
-        case e: NativeShuffleExchangeExec =>
+        case _: NativeShuffleExchangeExec =>
           true
       }.size
       assert(shuffleCount === 1, s"Expected 1 shuffle exchange, got 
$shuffleCount")
-      assert(collect(join.queryExecution.executedPlan) { case e: 
ReusedExchangeExec =>
+      assert(collect(join.queryExecution.executedPlan) { case _: 
ReusedExchangeExec =>
         true
       }.size === 1)
       val broadcasted = broadcast(join)
       val join2 = join.join(broadcasted, "id").join(broadcasted, "id")
       checkAnswer(join2, df)
       val shuffleCount2 = collect(join2.queryExecution.executedPlan) {
-        case e: NativeShuffleExchangeExec =>
+        case _: NativeShuffleExchangeExec =>
           true
       }.size
       assert(shuffleCount2 == 1, s"Expected 1 shuffle exchange in join2, got 
$shuffleCount2")
-      assert(collect(join2.queryExecution.executedPlan) { case e: 
ReusedExchangeExec =>
+      assert(collect(join2.queryExecution.executedPlan) { case _: 
ReusedExchangeExec =>
         true
       }.size == 4)
     }
diff --git 
a/auron-spark-ui/src/main/scala/org/apache/spark/sql/execution/ui/AuronSQLAppStatusListener.scala
 
b/auron-spark-ui/src/main/scala/org/apache/spark/sql/execution/ui/AuronSQLAppStatusListener.scala
index 0da16d4f..651234ac 100644
--- 
a/auron-spark-ui/src/main/scala/org/apache/spark/sql/execution/ui/AuronSQLAppStatusListener.scala
+++ 
b/auron-spark-ui/src/main/scala/org/apache/spark/sql/execution/ui/AuronSQLAppStatusListener.scala
@@ -16,6 +16,8 @@
  */
 package org.apache.spark.sql.execution.ui
 
+import scala.annotation.nowarn
+
 import org.apache.spark.{SparkConf, SparkContext}
 import org.apache.spark.internal.Logging
 import org.apache.spark.scheduler.{SparkListener, SparkListenerEvent}
@@ -23,6 +25,7 @@ import org.apache.spark.status.ElementTrackingStore
 
 import org.apache.auron.spark.ui.AuronBuildInfoEvent
 
+@nowarn("cat=unused") // conf temporarily unused
 class AuronSQLAppStatusListener(conf: SparkConf, kvstore: ElementTrackingStore)
     extends SparkListener
     with Logging {
diff --git a/pom.xml b/pom.xml
index f59ee507..405b3474 100644
--- a/pom.xml
+++ b/pom.xml
@@ -379,7 +379,10 @@
           <target>${javaVersion}</target>
           <scalaVersion>${scalaLongVersion}</scalaVersion>
           <args>
+            <arg>-deprecation</arg>
+            <arg>-feature</arg>
             <arg>-Ywarn-unused</arg>
+            <arg>-Xfatal-warnings</arg>
           </args>
         </configuration>
         <dependencies>
@@ -927,11 +930,25 @@
             <artifactId>scala-maven-plugin</artifactId>
             <configuration>
               <args>
+                <arg>-deprecation</arg>
+                <arg>-feature</arg>
                 <arg>-Ywarn-unused</arg>
                 <!-- https://github.com/scalamacros/paradise is no longer 
actively developed. -->
                 <!-- In Scala 2.13, the plugin's functionality has been 
included in the compiler -->
                 <!-- directly under the -Ymacro-annotations flag.  -->
                 <arg>-Ymacro-annotations</arg>
+                <arg>-Xfatal-warnings</arg>
+
+                <arg>-Wconf:cat=deprecation:wv,any:e</arg>
+                <arg>-Wconf:cat=other-nullary-override:s</arg>
+                
<arg>-Wconf:msg=^(?=.*?method|value|type|object|trait|inheritance)(?=.*?deprecated)(?=.*?since
 2.13).+$:s</arg>
+                <arg>-Wconf:msg=Auto-application to \`\(\)\` is 
deprecated:s</arg>
+                <arg>-Wconf:msg=object JavaConverters in package collection is 
deprecated:s</arg>
+                <arg>-Wconf:cat=unchecked&amp;msg=outer reference:s</arg>
+                <arg>-Wconf:cat=unchecked&amp;msg=eliminated by erasure:s</arg>
+                <arg>-Wconf:cat=unused-nowarn:s</arg>
+                <arg>-Wconf:msg=early initializers are deprecated:s</arg>
+                <arg>-Wconf:cat=other-match-analysis:s</arg>
               </args>
               <compilerPlugins>
                 <compilerPlugin>
diff --git 
a/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/auron/InterceptedValidateSparkPlan.scala
 
b/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/auron/InterceptedValidateSparkPlan.scala
index 95aefa65..f90e4865 100644
--- 
a/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/auron/InterceptedValidateSparkPlan.scala
+++ 
b/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/auron/InterceptedValidateSparkPlan.scala
@@ -16,6 +16,8 @@
  */
 package org.apache.spark.sql.auron
 
+import scala.annotation.nowarn
+
 import org.apache.spark.internal.Logging
 import org.apache.spark.sql.execution.SparkPlan
 
@@ -71,6 +73,7 @@ object InterceptedValidateSparkPlan extends Logging {
     }
   }
 
+  @nowarn("cat=unused") // plan unused
   @sparkver("3.0 / 3.1")
   def validate(plan: SparkPlan): Unit = {
     throw new UnsupportedOperationException("validate is not supported in 
spark 3.0.3 or 3.1.3")
diff --git 
a/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/auron/ShimsImpl.scala
 
b/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/auron/ShimsImpl.scala
index 839f5655..c3a1861d 100644
--- 
a/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/auron/ShimsImpl.scala
+++ 
b/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/auron/ShimsImpl.scala
@@ -19,6 +19,7 @@ package org.apache.spark.sql.auron
 import java.io.File
 import java.util.UUID
 
+import scala.annotation.nowarn
 import scala.collection.mutable
 
 import org.apache.commons.lang3.reflect.FieldUtils
@@ -284,6 +285,7 @@ class ShimsImpl extends Shims with Logging {
       child: SparkPlan): NativeGenerateBase =
     NativeGenerateExec(generator, requiredChildOutput, outer, generatorOutput, 
child)
 
+  @sparkver("3.4 / 3.5")
   private def effectiveLimit(rawLimit: Int): Int =
     if (rawLimit == -1) Int.MaxValue else rawLimit
 
@@ -989,6 +991,7 @@ class ShimsImpl extends Shims with Logging {
     }
   }
 
+  @nowarn("cat=unused") // Some params temporarily unused
   @sparkver("3.4 / 3.5")
   private def convertPromotePrecision(
       e: Expression,
@@ -1021,6 +1024,7 @@ class ShimsImpl extends Shims with Logging {
     }
   }
 
+  @nowarn("cat=unused") // Some params temporarily unused
   @sparkver("3.0 / 3.1 / 3.2")
   private def convertBloomFilterAgg(agg: AggregateFunction): 
Option[pb.PhysicalAggExprNode] = None
 
@@ -1047,6 +1051,7 @@ class ShimsImpl extends Shims with Logging {
     }
   }
 
+  @nowarn("cat=unused") // Some params temporarily unused
   @sparkver("3.0 / 3.1 / 3.2")
   private def convertBloomFilterMightContain(
       e: Expression,
diff --git 
a/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronBlockStoreShuffleReader.scala
 
b/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronBlockStoreShuffleReader.scala
index e08d8490..14ebba48 100644
--- 
a/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronBlockStoreShuffleReader.scala
+++ 
b/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronBlockStoreShuffleReader.scala
@@ -37,6 +37,9 @@ class AuronBlockStoreShuffleReader[K, C](
     extends AuronBlockStoreShuffleReaderBase[K, C](handle, context)
     with Logging {
 
+  // Touch mapOutputTracker to suppress -Xfatal-warnings (used in Spark 3.2+, 
unused in 3.0/3.1)
+  private val _ = mapOutputTracker
+
   override def readBlocks(): Iterator[InputStream] = {
     @sparkver("3.2 / 3.3 / 3.4 / 3.5")
     def fetchIterator = new ShuffleBlockFetcherIterator(
diff --git 
a/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronRssShuffleManagerBase.scala
 
b/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronRssShuffleManagerBase.scala
index 98863bf6..af92b0e8 100644
--- 
a/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronRssShuffleManagerBase.scala
+++ 
b/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronRssShuffleManagerBase.scala
@@ -16,6 +16,8 @@
  */
 package org.apache.spark.sql.execution.auron.shuffle
 
+import scala.annotation.nowarn
+
 import org.apache.spark.{ShuffleDependency, SparkConf, TaskContext}
 import org.apache.spark.internal.Logging
 import org.apache.spark.shuffle._
@@ -23,6 +25,7 @@ import 
org.apache.spark.sql.execution.auron.shuffle.AuronShuffleDependency.isArr
 
 import org.apache.auron.sparkver
 
+@nowarn("cat=unused") // _conf temporarily unused
 abstract class AuronRssShuffleManagerBase(_conf: SparkConf) extends 
ShuffleManager with Logging {
   override def registerShuffle[K, V, C](
       shuffleId: Int,
diff --git 
a/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/execution/joins/auron/plan/NativeShuffledHashJoinExecProvider.scala
 
b/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/execution/joins/auron/plan/NativeShuffledHashJoinExecProvider.scala
index 0236dd26..5f763eec 100644
--- 
a/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/execution/joins/auron/plan/NativeShuffledHashJoinExecProvider.scala
+++ 
b/spark-extension-shims-spark/src/main/scala/org/apache/spark/sql/execution/joins/auron/plan/NativeShuffledHashJoinExecProvider.scala
@@ -16,6 +16,8 @@
  */
 package org.apache.spark.sql.execution.joins.auron.plan
 
+import scala.annotation.nowarn
+
 import org.apache.spark.sql.auron.join.JoinBuildSides.JoinBuildSide
 import org.apache.spark.sql.catalyst.expressions.Expression
 import org.apache.spark.sql.catalyst.plans.JoinType
@@ -80,6 +82,7 @@ case object NativeShuffledHashJoinExecProvider {
     NativeShuffledHashJoinExec(left, right, leftKeys, rightKeys, joinType, 
buildSide, isSkewJoin)
   }
 
+  @nowarn("cat=unused") // Some params temporarily unused
   @sparkver("3.1")
   def provide(
       left: SparkPlan,
@@ -127,6 +130,7 @@ case object NativeShuffledHashJoinExecProvider {
     NativeShuffledHashJoinExec(left, right, leftKeys, rightKeys, joinType, 
buildSide)
   }
 
+  @nowarn("cat=unused") // Some params temporarily unused
   @sparkver("3.0")
   def provide(
       left: SparkPlan,
diff --git 
a/spark-extension-shims-spark/src/test/scala/org/apache/auron/AuronFunctionSuite.scala
 
b/spark-extension-shims-spark/src/test/scala/org/apache/auron/AuronFunctionSuite.scala
index c7e2f6a9..ae1d2432 100644
--- 
a/spark-extension-shims-spark/src/test/scala/org/apache/auron/AuronFunctionSuite.scala
+++ 
b/spark-extension-shims-spark/src/test/scala/org/apache/auron/AuronFunctionSuite.scala
@@ -278,8 +278,6 @@ class AuronFunctionSuite extends AuronQueryTest with 
BaseAuronSQLSuite {
       val dateTimeStampMin = format.parse(dateStringMin).getTime
       val dateTimeStampMax = format.parse(dateStringMax).getTime
       format = new SimpleDateFormat("yyyy-MM-dd")
-      val dateString = "2015-01-01"
-      val date = format.parse(dateString)
 
       val functions =
         s"""
@@ -320,8 +318,6 @@ class AuronFunctionSuite extends AuronQueryTest with 
BaseAuronSQLSuite {
       val dateTimeStampMin = format.parse(dateStringMin).getTime
       val dateTimeStampMax = format.parse(dateStringMax).getTime
       format = new SimpleDateFormat("yyyy-MM-dd")
-      val dateString = "2015-07-01"
-      val date = format.parse(dateString)
 
       val functions =
         s"""
diff --git 
a/spark-extension-shims-spark/src/test/scala/org/apache/auron/AuronQuerySuite.scala
 
b/spark-extension-shims-spark/src/test/scala/org/apache/auron/AuronQuerySuite.scala
index 349b489a..3e990978 100644
--- 
a/spark-extension-shims-spark/src/test/scala/org/apache/auron/AuronQuerySuite.scala
+++ 
b/spark-extension-shims-spark/src/test/scala/org/apache/auron/AuronQuerySuite.scala
@@ -204,18 +204,10 @@ class AuronQuerySuite extends AuronQueryTest with 
BaseAuronSQLSuite with AuronSQ
               .toDF("c1", "c2")
               .write
               .orc(path)
-            val correctAnswer = Seq(Row(1, 2), Row(3, 4), Row(5, 6), Row(null, 
null))
             checkSparkAnswerAndOperator(() => spark.read.orc(path))
 
             withTable("t") {
               sql(s"CREATE EXTERNAL TABLE t(c3 INT, c2 INT) USING ORC LOCATION 
'$path'")
-
-              val expected = if (forcePositionalEvolution) {
-                correctAnswer
-              } else {
-                Seq(Row(null, 2), Row(null, 4), Row(null, 6), Row(null, null))
-              }
-
               checkSparkAnswerAndOperator(() => spark.table("t"))
             }
           }
@@ -236,7 +228,6 @@ class AuronQuerySuite extends AuronQueryTest with 
BaseAuronSQLSuite with AuronSQ
               .write
               .partitionBy("p")
               .orc(path)
-            val correctAnswer = Seq(Row(1, 2, 1), Row(3, 4, 2), Row(5, 6, 3), 
Row(null, null, 4))
             checkSparkAnswerAndOperator(() => spark.read.orc(path))
 
             withTable("t") {
@@ -247,12 +238,6 @@ class AuronQuerySuite extends AuronQueryTest with 
BaseAuronSQLSuite with AuronSQ
                      |LOCATION '$path'
                      |""".stripMargin)
               sql("MSCK REPAIR TABLE t")
-              if (forcePositionalEvolution) {
-                correctAnswer
-              } else {
-                Seq(Row(null, 2, 1), Row(null, 4, 2), Row(null, 6, 3), 
Row(null, null, 4))
-              }
-
               checkSparkAnswerAndOperator(() => spark.table("t"))
             }
           }
diff --git 
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronCallNativeWrapper.scala
 
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronCallNativeWrapper.scala
index bf0918c4..b4028c1e 100644
--- 
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronCallNativeWrapper.scala
+++ 
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronCallNativeWrapper.scala
@@ -22,6 +22,7 @@ import java.nio.file.Files
 import java.nio.file.StandardCopyOption
 import java.util.concurrent.atomic.AtomicReference
 
+import scala.annotation.nowarn
 import scala.collection.mutable.ArrayBuffer
 
 import org.apache.arrow.c.ArrowArray
@@ -53,6 +54,7 @@ import org.apache.auron.protobuf.TaskDefinition
  * This class has been deprecated and migrated to {@link
  * org.apache.auron.jni.AuronCallNativeWrapper}. Will be removed in the future.
  */
+@nowarn("cat=deprecation") // JniBridge is temporarily used (deprecated)
 @Deprecated
 case class AuronCallNativeWrapper(
     nativePlan: PhysicalPlanNode,
@@ -193,6 +195,7 @@ case class AuronCallNativeWrapper(
   }
 }
 
+@nowarn("cat=deprecation") // JniBridge is temporarily used (deprecated)
 object AuronCallNativeWrapper extends Logging {
   def initNative(): Unit = {
     lazyInitNative
diff --git 
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
 
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
index f88c03f2..8b48c39a 100644
--- 
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
+++ 
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
@@ -1010,12 +1010,12 @@ object NativeConverters extends Logging {
         val children = e.children.map(Cast(_, e.dataType))
         buildScalarFunction(pb.ScalarFunction.Coalesce, children, e.dataType)
 
-      case e @ StringLPad(str, len, pad) =>
+      case _ @StringLPad(str, len, pad) =>
         buildScalarFunction(
           pb.ScalarFunction.Lpad,
           Seq(str, castIfNecessary(len, LongType), pad),
           StringType)
-      case e @ StringRPad(str, len, pad) =>
+      case _ @StringRPad(str, len, pad) =>
         buildScalarFunction(
           pb.ScalarFunction.Rpad,
           Seq(str, castIfNecessary(len, LongType), pad),
diff --git 
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeHelper.scala 
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeHelper.scala
index 3ae7669e..e1665647 100644
--- 
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeHelper.scala
+++ 
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeHelper.scala
@@ -100,7 +100,7 @@ object NativeHelper extends Logging {
     if (nativePlan == null) {
       return Iterator.empty
     }
-    var auronCallNativeWrapper = new 
org.apache.auron.jni.AuronCallNativeWrapper(
+    val auronCallNativeWrapper = new 
org.apache.auron.jni.AuronCallNativeWrapper(
       ROOT_ALLOCATOR,
       nativePlan,
       metrics,
diff --git 
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/Shims.scala 
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/Shims.scala
index 85ab65d5..8d817521 100644
--- a/spark-extension/src/main/scala/org/apache/spark/sql/auron/Shims.scala
+++ b/spark-extension/src/main/scala/org/apache/spark/sql/auron/Shims.scala
@@ -18,6 +18,8 @@ package org.apache.spark.sql.auron
 
 import java.io.File
 
+import scala.annotation.nowarn
+
 import org.apache.spark.ShuffleDependency
 import org.apache.spark.SparkContext
 import org.apache.spark.TaskContext
@@ -272,6 +274,7 @@ abstract class Shims {
 
   def getMinPartitionNum(sparkSession: SparkSession): Int
 
+  @nowarn("cat=unused") // Some params temporarily unused
   def postTransform(plan: SparkPlan, sc: SparkContext): Unit = {}
 
   def getAdaptiveInputPlan(exec: AdaptiveSparkPlanExec): SparkPlan
diff --git 
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/columnar/AuronArrowColumnVector.scala
 
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/columnar/AuronArrowColumnVector.scala
index d8be57e6..3dbde795 100644
--- 
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/columnar/AuronArrowColumnVector.scala
+++ 
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/columnar/AuronArrowColumnVector.scala
@@ -16,6 +16,8 @@
  */
 package org.apache.spark.sql.execution.auron.columnar
 
+import scala.annotation.nowarn
+
 import org.apache.arrow.vector.BigIntVector
 import org.apache.arrow.vector.BitVector
 import org.apache.arrow.vector.DateDayVector
@@ -143,6 +145,7 @@ class AuronArrowColumnVector(vector: ValueVector)
 }
 
 object AuronArrowColumnVector {
+  @nowarn("cat=unused") // Data type get methods unimplemented (placeholder)
   abstract private class ArrowVectorAccessor(private val vector: ValueVector) {
     def isNullAt(rowId: Int): Boolean =
       if (vector.getValueCount > 0 && vector.getValidityBuffer.capacity == 0) 
false
diff --git 
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeParquetSinkBase.scala
 
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeParquetSinkBase.scala
index 260ee249..2bd50523 100644
--- 
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeParquetSinkBase.scala
+++ 
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeParquetSinkBase.scala
@@ -21,6 +21,7 @@ import java.security.PrivilegedExceptionAction
 import java.util
 import java.util.UUID
 
+import scala.annotation.nowarn
 import scala.collection.JavaConverters._
 
 import org.apache.hadoop.conf.Configuration
@@ -80,7 +81,7 @@ abstract class NativeParquetSinkBase(
       hiveQlTable.getMetadata)
     val tableSchema = table.schema
     val hadoopConf = newHadoopConf(tableDesc)
-    val job = new Job(hadoopConf)
+    val job = Job.getInstance(hadoopConf)
     val parquetFileFormat = new ParquetFileFormat()
     parquetFileFormat.prepareWrite(sparkSession, job, Map(), tableSchema)
 
@@ -114,7 +115,7 @@ abstract class NativeParquetSinkBase(
           })
 
         // init parquet schema
-        val job = new Job(new JobConf(serializableConf.value))
+        val job = Job.getInstance(new JobConf(serializableConf.value))
         val tableProperties = tableDesc.getProperties
         val columnNameProperty: String = 
tableProperties.getProperty(IOConstants.COLUMNS)
         val columnTypeProperty: String = 
tableProperties.getProperty(IOConstants.COLUMNS_TYPES)
@@ -157,6 +158,7 @@ abstract class NativeParquetSinkBase(
       friendlyName = "NativeRDD.ParquetSink")
   }
 
+  @nowarn("cat=unused") // _tableDesc temporarily unused
   protected def newHadoopConf(_tableDesc: TableDesc): Configuration =
     sparkSession.sessionState.newHadoopConf()
 }
diff --git 
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeShuffleExchangeBase.scala
 
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeShuffleExchangeBase.scala
index 552dcff3..8b6aa04d 100644
--- 
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeShuffleExchangeBase.scala
+++ 
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeShuffleExchangeBase.scala
@@ -18,6 +18,7 @@ package org.apache.spark.sql.execution.auron.plan
 
 import java.util.UUID
 
+import scala.annotation.nowarn
 import scala.collection.JavaConverters._
 import scala.collection.mutable
 import scala.collection.mutable.ArrayBuffer
@@ -308,6 +309,7 @@ abstract class NativeShuffleExchangeBase(
     dependency
   }
 
+  @nowarn("cat=unused") // Some params temporarily unused
   private def rangePartitioningBound[K: Ordering: ClassTag, V](
       partitions: Int,
       rdd: RDD[_ <: Product2[K, V]],
diff --git 
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeShuffledHashJoinBase.scala
 
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeShuffledHashJoinBase.scala
index 1f8a06c8..17dfbe25 100644
--- 
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeShuffledHashJoinBase.scala
+++ 
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeShuffledHashJoinBase.scala
@@ -82,6 +82,8 @@ abstract class NativeShuffledHashJoinBase(
   private def nativeBuildSide = buildSide match {
     case JoinBuildLeft => pb.JoinSide.LEFT_SIDE
     case JoinBuildRight => pb.JoinSide.RIGHT_SIDE
+    case other =>
+      throw new IllegalArgumentException(s"Unknown Join buildSide: $other")
   }
 
   protected def rewriteKeyExprToLong(exprs: Seq[Expression]): Seq[Expression]
diff --git 
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeUnionBase.scala
 
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeUnionBase.scala
index a62ea6f1..bf3a1549 100644
--- 
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeUnionBase.scala
+++ 
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeUnionBase.scala
@@ -70,7 +70,7 @@ abstract class NativeUnionBase(
         val unionInputs = ArrayBuffer[(PhysicalPlanNode, Int)]()
         partition match {
           case p: UnionPartition[_] =>
-            val rdds = unionRDD.asInstanceOf[UnionRDD[_]].rdds
+            val rdds = unionRDD.asInstanceOf[UnionRDD[Any]].rdds
             val nativeRDD = rdds(p.parentRddIndex).asInstanceOf[NativeRDD]
             val input = nativeRDD.nativePlan(p.parentPartition, taskContext)
             for (childIndex <- rdds.indices) {
@@ -81,7 +81,7 @@ abstract class NativeUnionBase(
               }
             }
           case p: PartitionerAwareUnionRDDPartition =>
-            val rdds = unionRDD.asInstanceOf[PartitionerAwareUnionRDD[_]].rdds
+            val rdds = 
unionRDD.asInstanceOf[PartitionerAwareUnionRDD[Any]].rdds
             for ((rdd, partition) <- rdds.zip(p.parents)) {
               val nativeRDD = rdd.asInstanceOf[NativeRDD]
               unionInputs.append((nativeRDD.nativePlan(partition, 
taskContext), partition.index))
diff --git 
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronBlockStoreShuffleReaderBase.scala
 
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronBlockStoreShuffleReaderBase.scala
index 477c1b5e..eaba3640 100644
--- 
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronBlockStoreShuffleReaderBase.scala
+++ 
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronBlockStoreShuffleReaderBase.scala
@@ -182,5 +182,5 @@ trait BlockObject extends AutoCloseable {
   def getFileLength: Long = throw new UnsupportedOperationException
   def getByteBuffer: ByteBuffer = throw new UnsupportedOperationException
   def getChannel: ReadableByteChannel = throw new UnsupportedOperationException
-  def throwFetchFailed(errmsg: String): Unit = throw new 
UnsupportedOperationException
+  def throwFetchFailed(errmsg: String): Unit = throw new 
UnsupportedOperationException(errmsg)
 }
diff --git 
a/spark-version-annotation-macros/src/main/scala/org/apache/auron/sparkver.scala
 
b/spark-version-annotation-macros/src/main/scala/org/apache/auron/sparkver.scala
index 7d1e494d..fd7d4ed9 100644
--- 
a/spark-version-annotation-macros/src/main/scala/org/apache/auron/sparkver.scala
+++ 
b/spark-version-annotation-macros/src/main/scala/org/apache/auron/sparkver.scala
@@ -16,8 +16,7 @@
  */
 package org.apache.auron
 
-import scala.annotation.StaticAnnotation
-import scala.annotation.compileTimeOnly
+import scala.annotation.{compileTimeOnly, nowarn, StaticAnnotation}
 import scala.language.experimental._
 import scala.reflect.macros.whitebox
 
@@ -96,16 +95,19 @@ object sparkver {
   }
 }
 
+@nowarn("cat=unused") // 'vers' is used by macro
 @compileTimeOnly("enable macro paradise to expand macro annotations")
 final class sparkver(vers: String) extends StaticAnnotation {
   def macroTransform(annottees: Any*): Any = macro sparkver.Macros.verEnable
 }
 
+@nowarn("cat=unused") // 'vers' is used by macro
 @compileTimeOnly("enable macro paradise to expand macro annotations")
 final class sparkverEnableMembers(vers: String) extends StaticAnnotation {
   def macroTransform(annottees: Any*): Any = macro 
sparkver.Macros.verEnableMembers
 }
 
+@nowarn("cat=unused") // 'vers' is used by macro
 @compileTimeOnly("enable macro paradise to expand macro annotations")
 final class sparkverEnableOverride(vers: String) extends StaticAnnotation {
   def macroTransform(annottees: Any*): Any = macro 
sparkver.Macros.verEnableOverride
diff --git 
a/thirdparty/auron-uniffle/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/uniffle/AuronUniffleShuffleReader.scala
 
b/thirdparty/auron-uniffle/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/uniffle/AuronUniffleShuffleReader.scala
index 18745f5c..c4b1ca31 100644
--- 
a/thirdparty/auron-uniffle/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/uniffle/AuronUniffleShuffleReader.scala
+++ 
b/thirdparty/auron-uniffle/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/uniffle/AuronUniffleShuffleReader.scala
@@ -20,6 +20,7 @@ import java.io.InputStream
 import java.nio.ByteBuffer
 import java.util
 
+import scala.annotation.nowarn
 import scala.collection.AbstractIterator
 
 import org.apache.commons.lang3.reflect.FieldUtils
@@ -40,6 +41,7 @@ import org.apache.uniffle.common.config.RssConf
 import org.apache.uniffle.common.exception.RssException
 import 
org.apache.uniffle.shaded.org.roaringbitmap.longlong.Roaring64NavigableMap
 
+@nowarn("cat=unused") // Some params temporarily unused
 class AuronUniffleShuffleReader[K, C](
     reader: RssShuffleReader[K, C],
     handle: RssShuffleHandleWrapper[K, _, C],
@@ -207,6 +209,7 @@ class AuronUniffleShuffleReader[K, C](
     }
   }
 
+  @nowarn("cat=unused") // Some params temporarily unused
   private class UniffleInputStream(
       iterator: MultiPartitionIterator[_, _],
       shuffleId: Int,

Reply via email to