This is an automated email from the ASF dual-hosted git repository.
richox pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/auron.git
The following commit(s) were added to refs/heads/master by this push:
new 4574dfbc [AURON #1523] Fix CheckStyle Violations in Auron Codebase.
(#1523)
4574dfbc is described below
commit 4574dfbcbc7b1f6df3dfc655cdde1f69478ee6fa
Author: slfan1989 <[email protected]>
AuthorDate: Tue Oct 28 14:13:27 2025 +0800
[AURON #1523] Fix CheckStyle Violations in Auron Codebase. (#1523)
* Fix CheckStyle Violations in Auron Codebase.
Signed-off-by: slfan1989 <[email protected]>
* Fix CheckStyle Violations in Auron Codebase.
Signed-off-by: slfan1989 <[email protected]>
---------
Signed-off-by: slfan1989 <[email protected]>
---
.../test/scala/org/apache/spark/sql/auron/EmptyNativeRddSuite.scala | 2 +-
.../src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala | 2 +-
.../src/main/scala/org/apache/spark/sql/auron/NativeRDD.scala | 2 +-
.../apache/spark/sql/execution/auron/plan/NativeParquetScanBase.scala | 2 +-
.../execution/auron/shuffle/AuronBlockStoreShuffleReaderBase.scala | 2 +-
.../src/main/scala/org/apache/auron/sparkver.scala | 4 ++--
6 files changed, 7 insertions(+), 7 deletions(-)
diff --git
a/spark-extension-shims-spark/src/test/scala/org/apache/spark/sql/auron/EmptyNativeRddSuite.scala
b/spark-extension-shims-spark/src/test/scala/org/apache/spark/sql/auron/EmptyNativeRddSuite.scala
index 2ef2e371..59263544 100644
---
a/spark-extension-shims-spark/src/test/scala/org/apache/spark/sql/auron/EmptyNativeRddSuite.scala
+++
b/spark-extension-shims-spark/src/test/scala/org/apache/spark/sql/auron/EmptyNativeRddSuite.scala
@@ -28,7 +28,7 @@ class EmptyNativeRddSuite
assert(empty.collect().size === 0)
val thrown = intercept[UnsupportedOperationException] {
- empty.reduce((row1, row2) => {
+ empty.reduce((row1, _) => {
row1
})
}
diff --git
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala
index 6c7f348a..dcb4b672 100644
---
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala
+++
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/AuronConverters.scala
@@ -1099,7 +1099,7 @@ object AuronConverters extends Logging {
rddPartitioner = None,
rddDependencies = Nil,
false,
- (_partition, _taskContext) => {
+ (_, _) => {
val nativeEmptyExec = EmptyPartitionsExecNode
.newBuilder()
.setNumPartitions(outputPartitioning.numPartitions)
diff --git
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeRDD.scala
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeRDD.scala
index 4000035e..c6205d17 100644
--- a/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeRDD.scala
+++ b/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeRDD.scala
@@ -104,7 +104,7 @@ class NativePlanWrapper(var p: (Partition, TaskContext) =>
PhysicalPlanNode)
@throws[IOException]
@throws[ClassNotFoundException]
private def readObject(in: ObjectInputStream): Unit = {
- val _init: Unit = NativePlanWrapper.changeProtobufDefaultRecursionLimit
+ NativePlanWrapper.changeProtobufDefaultRecursionLimit
p = in.readObject.asInstanceOf[(Partition, TaskContext) =>
PhysicalPlanNode]
}
}
diff --git
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeParquetScanBase.scala
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeParquetScanBase.scala
index 591e31a1..ddee72d3 100644
---
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeParquetScanBase.scala
+++
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/plan/NativeParquetScanBase.scala
@@ -63,7 +63,7 @@ abstract class NativeParquetScanBase(basedFileScan:
FileSourceScanExec)
None,
Nil,
rddShuffleReadFull = true,
- (partition, _context) => {
+ (partition, _) => {
val resourceId =
s"NativeParquetScanExec:${UUID.randomUUID().toString}"
putJniBridgeResource(resourceId, broadcastedHadoopConf)
diff --git
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronBlockStoreShuffleReaderBase.scala
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronBlockStoreShuffleReaderBase.scala
index 30f4abb8..b64ad045 100644
---
a/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronBlockStoreShuffleReaderBase.scala
+++
b/spark-extension/src/main/scala/org/apache/spark/sql/execution/auron/shuffle/AuronBlockStoreShuffleReaderBase.scala
@@ -162,7 +162,7 @@ object AuronBlockStoreShuffleReaderBase extends Logging {
case in: InputStreamToByteBuffer =>
Some(in.toByteBuffer)
- case in =>
+ case _ =>
None
}
}
diff --git
a/spark-version-annotation-macros/src/main/scala/org/apache/auron/sparkver.scala
b/spark-version-annotation-macros/src/main/scala/org/apache/auron/sparkver.scala
index b50fc1fc..7d1e494d 100644
---
a/spark-version-annotation-macros/src/main/scala/org/apache/auron/sparkver.scala
+++
b/spark-version-annotation-macros/src/main/scala/org/apache/auron/sparkver.scala
@@ -62,9 +62,9 @@ object sparkver {
impl(c)(annottees: _*) {
val head = annottees.head.tree match {
- case ClassDef(mods, name, tparams, Template(parents, self, _body)) =>
+ case ClassDef(mods, name, tparams, Template(parents, self, _)) =>
ClassDef(mods, name, tparams, Template(parents, self,
List(EmptyTree)))
- case ModuleDef(mods, name, Template(parents, self, _body)) =>
+ case ModuleDef(mods, name, Template(parents, self, _)) =>
ModuleDef(mods, name, Template(parents, self, List(EmptyTree)))
}
c.Expr(q"$head; ..${annottees.tail}")