This is an automated email from the ASF dual-hosted git repository.
maxgekk pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push:
new d072833d50a [SPARK-45698][CORE][SQL][SS] Clean up the deprecated API
usage related to `Buffer`
d072833d50a is described below
commit d072833d50abd1a6630dd629e3560e3d0ad929cf
Author: yangjie01 <[email protected]>
AuthorDate: Fri Oct 27 12:59:22 2023 +0300
[SPARK-45698][CORE][SQL][SS] Clean up the deprecated API usage related to
`Buffer`
### What changes were proposed in this pull request?
This PR cleans up the use of the following APIs in `s.c.mutable.Buffer`, as
they have been deprecated after Scala 2.13.0 or Scala 2.13.4:
- `trimStart` -> `dropInPlace`
- `trimEnd` -> `dropRightInPlace`
- `append(elems: A*)` -> `appendAll`
- `prepend(elems: A*)` -> `prependAll(elems)`
```
deprecated("use dropInPlace instead", since = "2.13.4")
def trimStart(n: Int): Unit = dropInPlace(n)
deprecated("use dropRightInPlace instead", since = "2.13.4")
def trimEnd(n: Int): Unit = dropRightInPlace(n)
deprecated("Use appendAll instead", "2.13.0")
`inline` final def append(elems: A*): this.type = addAll(elems)
deprecated("Use prependAll instead", "2.13.0")
`inline` final def prepend(elems: A*): this.type = prependAll(elems)
```
### Why are the changes needed?
Clean up deprecated API usage.
### Does this PR introduce _any_ user-facing change?
No
### How was this patch tested?
Pass GitHub Acitons.
### Was this patch authored or co-authored using generative AI tooling?
No
Closes #43551 from LuciferYang/buffer-deprecated.
Authored-by: yangjie01 <[email protected]>
Signed-off-by: Max Gekk <[email protected]>
---
core/src/main/scala/org/apache/spark/deploy/master/Master.scala | 4 ++--
core/src/main/scala/org/apache/spark/util/SizeEstimator.scala | 2 +-
.../org/apache/spark/util/collection/ExternalAppendOnlyMap.scala | 2 +-
core/src/test/scala/org/apache/spark/deploy/IvyTestUtils.scala | 2 +-
.../src/main/scala/org/apache/spark/sql/execution/SparkPlan.scala | 4 ++--
.../org/apache/spark/sql/execution/arrow/ArrowConverters.scala | 2 +-
.../streaming/continuous/ContinuousTextSocketSource.scala | 2 +-
.../scala/org/apache/spark/sql/execution/streaming/memory.scala | 2 +-
.../execution/streaming/sources/TextSocketMicroBatchStream.scala | 2 +-
sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala | 8 ++++----
10 files changed, 15 insertions(+), 15 deletions(-)
diff --git a/core/src/main/scala/org/apache/spark/deploy/master/Master.scala
b/core/src/main/scala/org/apache/spark/deploy/master/Master.scala
index 46503e017ea..29022c7419b 100644
--- a/core/src/main/scala/org/apache/spark/deploy/master/Master.scala
+++ b/core/src/main/scala/org/apache/spark/deploy/master/Master.scala
@@ -1042,7 +1042,7 @@ private[deploy] class Master(
completedApps.take(toRemove).foreach { a =>
applicationMetricsSystem.removeSource(a.appSource)
}
- completedApps.trimStart(toRemove)
+ completedApps.dropInPlace(toRemove)
}
completedApps += app // Remember it in our history
waitingApps -= app
@@ -1204,7 +1204,7 @@ private[deploy] class Master(
drivers -= driver
if (completedDrivers.size >= retainedDrivers) {
val toRemove = math.max(retainedDrivers / 10, 1)
- completedDrivers.trimStart(toRemove)
+ completedDrivers.dropInPlace(toRemove)
}
completedDrivers += driver
persistenceEngine.removeDriver(driver)
diff --git a/core/src/main/scala/org/apache/spark/util/SizeEstimator.scala
b/core/src/main/scala/org/apache/spark/util/SizeEstimator.scala
index 704aeaefa55..1447a3e752d 100644
--- a/core/src/main/scala/org/apache/spark/util/SizeEstimator.scala
+++ b/core/src/main/scala/org/apache/spark/util/SizeEstimator.scala
@@ -180,7 +180,7 @@ object SizeEstimator extends Logging {
def dequeue(): AnyRef = {
val elem = stack.last
- stack.trimEnd(1)
+ stack.dropRightInPlace(1)
elem
}
}
diff --git
a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
index 3afbe322b6e..8224472b754 100644
---
a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
+++
b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
@@ -362,7 +362,7 @@ class ExternalAppendOnlyMap[K, V, C](
private def removeFromBuffer[T](buffer: ArrayBuffer[T], index: Int): T = {
val elem = buffer(index)
buffer(index) = buffer(buffer.size - 1) // This also works if index ==
buffer.size - 1
- buffer.trimEnd(1)
+ buffer.dropRightInPlace(1)
elem
}
diff --git a/core/src/test/scala/org/apache/spark/deploy/IvyTestUtils.scala
b/core/src/test/scala/org/apache/spark/deploy/IvyTestUtils.scala
index 0dcdba3dfb8..0b970a03ad8 100644
--- a/core/src/test/scala/org/apache/spark/deploy/IvyTestUtils.scala
+++ b/core/src/test/scala/org/apache/spark/deploy/IvyTestUtils.scala
@@ -316,7 +316,7 @@ private[deploy] object IvyTestUtils {
}
if (withR) {
val rFiles = createRFiles(root, className, artifact.groupId)
- allFiles.append(rFiles: _*)
+ allFiles.appendAll(rFiles)
}
val jarFile = packJar(jarPath, artifact, allFiles.toSeq, useIvyLayout,
withR)
assert(jarFile.exists(), "Problem creating Jar file")
diff --git
a/sql/core/src/main/scala/org/apache/spark/sql/execution/SparkPlan.scala
b/sql/core/src/main/scala/org/apache/spark/sql/execution/SparkPlan.scala
index bbd74a1fe74..d93a83dec44 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/execution/SparkPlan.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/SparkPlan.scala
@@ -536,13 +536,13 @@ abstract class SparkPlan extends QueryPlan[SparkPlan]
with Logging with Serializ
while (buf.length < n && i < res.length) {
val rows = decodeUnsafeRows(res(i)._2)
if (n - buf.length >= res(i)._1) {
- buf.prepend(rows.toArray[InternalRow]: _*)
+ buf.prependAll(rows.toArray[InternalRow])
} else {
val dropUntil = res(i)._1 - (n - buf.length)
// Same as Iterator.drop but this only takes a long.
var j: Long = 0L
while (j < dropUntil) { rows.next(); j += 1L}
- buf.prepend(rows.toArray[InternalRow]: _*)
+ buf.prependAll(rows.toArray[InternalRow])
}
i += 1
}
diff --git
a/sql/core/src/main/scala/org/apache/spark/sql/execution/arrow/ArrowConverters.scala
b/sql/core/src/main/scala/org/apache/spark/sql/execution/arrow/ArrowConverters.scala
index a0dd939dda2..d6bf1e29edd 100644
---
a/sql/core/src/main/scala/org/apache/spark/sql/execution/arrow/ArrowConverters.scala
+++
b/sql/core/src/main/scala/org/apache/spark/sql/execution/arrow/ArrowConverters.scala
@@ -315,7 +315,7 @@ private[sql] object ArrowConverters extends Logging {
val reader =
new ArrowStreamReader(new ByteArrayInputStream(arrowBatchIter.next()),
allocator)
val root = if (reader.loadNextBatch()) reader.getVectorSchemaRoot else
null
- resources.append(reader, root)
+ resources.appendAll(Seq(reader, root))
if (root == null) {
(Iterator.empty, null)
} else {
diff --git
a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/continuous/ContinuousTextSocketSource.scala
b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/continuous/ContinuousTextSocketSource.scala
index d1b346b4174..3f5f81bee4e 100644
---
a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/continuous/ContinuousTextSocketSource.scala
+++
b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/continuous/ContinuousTextSocketSource.scala
@@ -144,7 +144,7 @@ class TextSocketContinuousStream(
" for partition " + partition + ". Max valid offset: " + max)
}
val n = offset - startOffset.offsets(partition)
- buckets(partition).trimStart(n)
+ buckets(partition).dropInPlace(n)
}
startOffset = endOffset
recordEndpoint.setStartOffsets(startOffset.offsets)
diff --git
a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/memory.scala
b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/memory.scala
index 8fe5319ef4c..826543fd565 100644
---
a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/memory.scala
+++
b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/memory.scala
@@ -279,7 +279,7 @@ case class MemoryStream[A : Encoder](
s"Offsets committed out of order: $lastOffsetCommitted followed by
$end")
}
- batches.trimStart(offsetDiff)
+ batches.dropInPlace(offsetDiff)
lastOffsetCommitted = newOffset
}
diff --git
a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/sources/TextSocketMicroBatchStream.scala
b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/sources/TextSocketMicroBatchStream.scala
index 580f7066e44..a01f40bead8 100644
---
a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/sources/TextSocketMicroBatchStream.scala
+++
b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/sources/TextSocketMicroBatchStream.scala
@@ -159,7 +159,7 @@ class TextSocketMicroBatchStream(host: String, port: Int,
numPartitions: Int)
s"Offsets committed out of order: $lastOffsetCommitted followed by
$end")
}
- batches.trimStart(offsetDiff)
+ batches.dropInPlace(offsetDiff)
lastOffsetCommitted = newOffset
}
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala
b/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala
index 45b80d81e73..8c293bab641 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala
@@ -756,10 +756,10 @@ class JoinSuite extends QueryTest with SharedSparkSession
with AdaptiveSparkPlan
}
val expected = new ListBuffer[Row]()
- expected.append(
+ expected.appendAll(Seq(
Row(1, "1", 1, 1), Row(1, "1", 1, 2),
Row(2, "2", 2, 1), Row(2, "2", 2, 2),
- Row(3, "3", 3, 1), Row(3, "3", 3, 2)
+ Row(3, "3", 3, 1), Row(3, "3", 3, 2))
)
for (i <- 4 to 100) {
expected.append(Row(i, i.toString, null, null))
@@ -830,10 +830,10 @@ class JoinSuite extends QueryTest with SharedSparkSession
with AdaptiveSparkPlan
}
val expected = new ListBuffer[Row]()
- expected.append(
+ expected.appendAll(Seq(
Row(1, "1", 1, 1), Row(1, "1", 1, 2),
Row(2, "2", 2, 1), Row(2, "2", 2, 2),
- Row(3, "3", 3, 1), Row(3, "3", 3, 2)
+ Row(3, "3", 3, 1), Row(3, "3", 3, 2))
)
for (i <- 4 to 100) {
expected.append(Row(i, i.toString, null, null))
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]