This is an automated email from the ASF dual-hosted git repository.
yumwang pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push:
new 0996a1588fe [SPARK-40401][CORE] Remove the support of deprecated
`spark.akka.*` configs
0996a1588fe is described below
commit 0996a1588feabea0aa3578bd551dcc1a1627eab7
Author: Dongjoon Hyun <[email protected]>
AuthorDate: Sat Sep 10 12:17:25 2022 +0800
[SPARK-40401][CORE] Remove the support of deprecated `spark.akka.*` configs
### What changes were proposed in this pull request?
This PR aims to remove the support of `spark.akka.*` configs and related
warnings.
### Why are the changes needed?
- Apache Spark 2.0+ is not using Akka, but some of `spark.akka.*` configs
are still accepted as alternative values from deprecated configs.
- Due to the Akka's License change, we had better be clear that Apache
Spark is irrelevant to Akka even in config names.
- https://www.lightbend.com/blog/why-we-are-changing-the-license-for-akka
After this PR, we don't have a word `spark.akka`.
```
$ git grep spark.akka
```
### Does this PR introduce _any_ user-facing change?
Yes, we will ignore `spark.akka.*`, but it's inevitable in order to remove
any user confusions due to the config naming, `spark.akka.*`.
### How was this patch tested?
Manually review.
Closes #37849 from dongjoon-hyun/SPARK-40401.
Authored-by: Dongjoon Hyun <[email protected]>
Signed-off-by: Yuming Wang <[email protected]>
---
core/src/main/scala/org/apache/spark/SparkConf.scala | 11 -----------
.../src/test/scala/org/apache/spark/SparkConfSuite.scala | 16 +---------------
2 files changed, 1 insertion(+), 26 deletions(-)
diff --git a/core/src/main/scala/org/apache/spark/SparkConf.scala
b/core/src/main/scala/org/apache/spark/SparkConf.scala
index f296b1408fa..08344d8e547 100644
--- a/core/src/main/scala/org/apache/spark/SparkConf.scala
+++ b/core/src/main/scala/org/apache/spark/SparkConf.scala
@@ -682,18 +682,12 @@ private[spark] object SparkConf extends Logging {
AlternateConfig("spark.io.compression.snappy.block.size", "1.4")),
IO_COMPRESSION_LZ4_BLOCKSIZE.key -> Seq(
AlternateConfig("spark.io.compression.lz4.block.size", "1.4")),
- RPC_ASK_TIMEOUT.key -> Seq(
- AlternateConfig("spark.akka.askTimeout", "1.4")),
- RPC_LOOKUP_TIMEOUT.key -> Seq(
- AlternateConfig("spark.akka.lookupTimeout", "1.4")),
"spark.streaming.fileStream.minRememberDuration" -> Seq(
AlternateConfig("spark.streaming.minRememberDuration", "1.5")),
"spark.yarn.max.executor.failures" -> Seq(
AlternateConfig("spark.yarn.max.worker.failures", "1.5")),
MEMORY_OFFHEAP_ENABLED.key -> Seq(
AlternateConfig("spark.unsafe.offHeap", "1.6")),
- RPC_MESSAGE_MAX_SIZE.key -> Seq(
- AlternateConfig("spark.akka.frameSize", "1.6")),
"spark.yarn.jars" -> Seq(
AlternateConfig("spark.yarn.jar", "2.0")),
MAX_REMOTE_BLOCK_SIZE_FETCH_TO_MEM.key -> Seq(
@@ -780,11 +774,6 @@ private[spark] object SparkConf extends Logging {
s"may be removed in the future. Please use the new key '$newKey'
instead.")
return
}
- if (key.startsWith("spark.akka") || key.startsWith("spark.ssl.akka")) {
- logWarning(
- s"The configuration key $key is not supported anymore " +
- s"because Spark doesn't use Akka since 2.0")
- }
}
/**
diff --git a/core/src/test/scala/org/apache/spark/SparkConfSuite.scala
b/core/src/test/scala/org/apache/spark/SparkConfSuite.scala
index 7ae1eac1db1..74fd7816221 100644
--- a/core/src/test/scala/org/apache/spark/SparkConfSuite.scala
+++ b/core/src/test/scala/org/apache/spark/SparkConfSuite.scala
@@ -20,7 +20,6 @@ package org.apache.spark
import java.util.concurrent.{Executors, TimeUnit}
import scala.collection.JavaConverters._
-import scala.concurrent.duration._
import scala.util.{Random, Try}
import com.esotericsoftware.kryo.Kryo
@@ -34,7 +33,7 @@ import org.apache.spark.resource.ResourceID
import org.apache.spark.resource.ResourceUtils._
import org.apache.spark.resource.TestResourceIDs._
import org.apache.spark.serializer.{JavaSerializer, KryoRegistrator,
KryoSerializer}
-import org.apache.spark.util.{ResetSystemProperties, RpcUtils, Utils}
+import org.apache.spark.util.{ResetSystemProperties, Utils}
class SparkConfSuite extends SparkFunSuite with LocalSparkContext with
ResetSystemProperties {
test("Test byteString conversion") {
@@ -281,19 +280,6 @@ class SparkConfSuite extends SparkFunSuite with
LocalSparkContext with ResetSyst
assert(conf.get(KERBEROS_FILESYSTEMS_TO_ACCESS) === Array("testNode"))
}
- test("akka deprecated configs") {
- val conf = new SparkConf()
-
- assert(!conf.contains(RPC_ASK_TIMEOUT))
- assert(!conf.contains(RPC_LOOKUP_TIMEOUT))
-
- conf.set("spark.akka.askTimeout", "3")
- assert(RpcUtils.askRpcTimeout(conf).duration === 3.seconds)
-
- conf.set("spark.akka.lookupTimeout", "4")
- assert(RpcUtils.lookupRpcTimeout(conf).duration === 4.seconds)
- }
-
test("SPARK-13727") {
val conf = new SparkConf()
// set the conf in the deprecated way
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]