This is an automated email from the ASF dual-hosted git repository. huaxingao pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push: new d10db29ca16 [SPARK-39711][TESTS] Remove redundant trait: BeforeAndAfterAll & BeforeAndAfterEach & Logging d10db29ca16 is described below commit d10db29ca1609c34f082068f0cc7419c5ecef190 Author: panbingkun <pbk1...@gmail.com> AuthorDate: Mon Jul 11 19:30:30 2022 -0700 [SPARK-39711][TESTS] Remove redundant trait: BeforeAndAfterAll & BeforeAndAfterEach & Logging ### What changes were proposed in this pull request? SparkFunSuite declare as follow: ``` abstract class SparkFunSuite extends AnyFunSuite with BeforeAndAfterAll with BeforeAndAfterEach with ThreadAudit with Logging ``` some suite extends SparkFunSuite and meanwhile with BeforeAndAfterAll or BeforeAndAfterEach or Logging, it is redundant. ### Why are the changes needed? Eliminate redundant information and make the code cleaner. ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? Pass GA. Closes #37123 from panbingkun/remove_BeforeAndAfterAll. Authored-by: panbingkun <pbk1...@gmail.com> Signed-off-by: huaxingao <huaxin_...@apple.com> --- .../scala/org/apache/spark/kafka010/KafkaTokenSparkConfSuite.scala | 3 +-- .../apache/spark/streaming/kafka010/KafkaDataConsumerSuite.scala | 3 +-- .../scala/org/apache/spark/streaming/kafka010/KafkaRDDSuite.scala | 3 +-- core/src/test/scala/org/apache/spark/SSLOptionsSuite.scala | 3 +-- .../org/apache/spark/SparkContextSchedulerCreationSuite.scala | 4 +--- core/src/test/scala/org/apache/spark/ThreadingSuite.scala | 4 +--- .../test/scala/org/apache/spark/deploy/SparkSubmitUtilsSuite.scala | 3 +-- .../org/apache/spark/deploy/history/ApplicationCacheSuite.scala | 2 +- .../org/apache/spark/deploy/history/FsHistoryProviderSuite.scala | 4 +--- .../spark/deploy/history/RealBrowserUIHistoryServerSuite.scala | 4 +--- .../scala/org/apache/spark/deploy/master/ui/MasterWebUISuite.scala | 4 +--- .../org/apache/spark/deploy/rest/StandaloneRestSubmitSuite.scala | 3 +-- .../org/apache/spark/input/WholeTextFileInputFormatSuite.scala | 5 +---- .../org/apache/spark/input/WholeTextFileRecordReaderSuite.scala | 4 +--- .../org/apache/spark/internal/plugin/PluginContainerSuite.scala | 3 +-- .../test/scala/org/apache/spark/memory/MemoryManagerSuite.scala | 4 +--- .../src/test/scala/org/apache/spark/rdd/AsyncRDDActionsSuite.scala | 3 +-- core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala | 3 +-- core/src/test/scala/org/apache/spark/rpc/RpcEnvSuite.scala | 3 +-- .../org/apache/spark/scheduler/EventLoggingListenerSuite.scala | 5 +---- .../test/scala/org/apache/spark/scheduler/HealthTrackerSuite.scala | 4 +--- .../scala/org/apache/spark/scheduler/TaskSchedulerImplSuite.scala | 6 ++---- .../scala/org/apache/spark/scheduler/TaskSetExcludelistSuite.scala | 3 +-- .../org/apache/spark/serializer/SerializationDebuggerSuite.scala | 5 +---- .../scala/org/apache/spark/shuffle/ShuffleBlockPusherSuite.scala | 3 +-- .../org/apache/spark/shuffle/ShuffleDriverComponentsSuite.scala | 4 +--- .../apache/spark/shuffle/sort/IndexShuffleBlockResolverSuite.scala | 3 +-- .../shuffle/sort/io/LocalDiskShuffleMapOutputWriterSuite.scala | 3 +-- .../scala/org/apache/spark/storage/BlockInfoManagerSuite.scala | 4 +--- .../test/scala/org/apache/spark/storage/BlockManagerSuite.scala | 7 +++---- .../scala/org/apache/spark/storage/DiskBlockManagerSuite.scala | 4 +--- .../org/apache/spark/storage/DiskBlockObjectWriterSuite.scala | 4 +--- .../scala/org/apache/spark/ui/RealBrowserUISeleniumSuite.scala | 3 +-- core/src/test/scala/org/apache/spark/ui/UISeleniumSuite.scala | 3 +-- core/src/test/scala/org/apache/spark/util/FileAppenderSuite.scala | 4 ++-- core/src/test/scala/org/apache/spark/util/UtilsSuite.scala | 3 +-- .../apache/spark/util/collection/ExternalSorterSpillSuite.scala | 3 +-- .../test/scala/org/apache/spark/util/collection/SorterSuite.scala | 3 +-- .../apache/spark/util/collection/unsafe/sort/RadixSortSuite.scala | 3 +-- mllib/src/test/scala/org/apache/spark/ml/MLEventsSuite.scala | 5 +---- .../test/scala/org/apache/spark/ml/recommendation/ALSSuite.scala | 3 +-- .../src/test/scala/org/apache/spark/ml/stat/CorrelationSuite.scala | 4 +--- .../org/apache/spark/ml/tree/impl/GradientBoostedTreesSuite.scala | 3 +-- .../test/scala/org/apache/spark/mllib/linalg/VectorsSuite.scala | 3 +-- .../test/scala/org/apache/spark/mllib/stat/CorrelationSuite.scala | 3 +-- .../org/apache/spark/mllib/tree/GradientBoostedTreesSuite.scala | 3 +-- repl/src/test/scala-2.12/org/apache/spark/repl/Repl2Suite.scala | 4 +--- repl/src/test/scala-2.13/org/apache/spark/repl/Repl2Suite.scala | 4 +--- repl/src/test/scala/org/apache/spark/repl/ReplSuite.scala | 3 +-- repl/src/test/scala/org/apache/spark/repl/SparkShellSuite.scala | 5 +---- .../scala/org/apache/spark/deploy/yarn/BaseYarnClusterSuite.scala | 6 ++---- .../apache/spark/deploy/yarn/ContainerPlacementStrategySuite.scala | 3 +-- .../scala/org/apache/spark/deploy/yarn/YarnAllocatorSuite.scala | 3 +-- .../org/apache/spark/deploy/yarn/YarnSparkHadoopUtilSuite.scala | 4 +--- .../org/apache/spark/network/yarn/YarnShuffleServiceSuite.scala | 3 +-- .../apache/spark/sql/catalyst/catalog/ExternalCatalogSuite.scala | 4 +--- .../sql/catalyst/expressions/aggregate/HistogramNumericSuite.scala | 3 +-- .../test/scala/org/apache/spark/sql/SparkSessionBuilderSuite.scala | 3 +-- .../spark/sql/execution/CoalesceShufflePartitionsSuite.scala | 4 +--- .../org/apache/spark/sql/execution/columnar/ColumnTypeSuite.scala | 3 +-- .../spark/sql/execution/streaming/state/StateStoreRDDSuite.scala | 4 ++-- .../apache/spark/sql/execution/vectorized/ColumnVectorSuite.scala | 4 +--- .../org/apache/spark/sql/sources/DataSourceAnalysisSuite.scala | 4 +--- .../scala/org/apache/spark/sql/streaming/ui/UISeleniumSuite.scala | 3 +-- .../scala/org/apache/spark/sql/hive/thriftserver/CliSuite.scala | 4 +--- .../org/apache/spark/sql/hive/HiveContextCompatibilitySuite.scala | 5 +---- .../test/scala/org/apache/spark/sql/hive/client/FiltersSuite.scala | 3 +-- .../org/apache/spark/sql/hive/execution/ConcurrentHiveSuite.scala | 4 +--- .../org/apache/spark/sql/hive/execution/Hive_2_1_DDLSuite.scala | 4 +--- .../scala/org/apache/spark/streaming/DStreamClosureSuite.scala | 4 +--- .../src/test/scala/org/apache/spark/streaming/FailureSuite.scala | 3 +-- .../org/apache/spark/streaming/ReceivedBlockTrackerSuite.scala | 4 +--- .../test/scala/org/apache/spark/streaming/UISeleniumSuite.scala | 4 +--- .../org/apache/spark/streaming/rdd/MapWithStateRDDSuite.scala | 4 +--- .../spark/streaming/rdd/WriteAheadLogBackedBlockRDDSuite.scala | 4 +--- 75 files changed, 81 insertions(+), 195 deletions(-) diff --git a/connector/kafka-0-10-token-provider/src/test/scala/org/apache/spark/kafka010/KafkaTokenSparkConfSuite.scala b/connector/kafka-0-10-token-provider/src/test/scala/org/apache/spark/kafka010/KafkaTokenSparkConfSuite.scala index 17caf96818e..e42704b8590 100644 --- a/connector/kafka-0-10-token-provider/src/test/scala/org/apache/spark/kafka010/KafkaTokenSparkConfSuite.scala +++ b/connector/kafka-0-10-token-provider/src/test/scala/org/apache/spark/kafka010/KafkaTokenSparkConfSuite.scala @@ -18,11 +18,10 @@ package org.apache.spark.kafka010 import org.apache.kafka.common.security.auth.SecurityProtocol.{SASL_SSL, SSL} -import org.scalatest.BeforeAndAfterEach import org.apache.spark.{SparkConf, SparkFunSuite} -class KafkaTokenSparkConfSuite extends SparkFunSuite with BeforeAndAfterEach { +class KafkaTokenSparkConfSuite extends SparkFunSuite { private val identifier1 = "cluster1" private val identifier2 = "cluster2" private val authBootStrapServers = "127.0.0.1:0" diff --git a/connector/kafka-0-10/src/test/scala/org/apache/spark/streaming/kafka010/KafkaDataConsumerSuite.scala b/connector/kafka-0-10/src/test/scala/org/apache/spark/streaming/kafka010/KafkaDataConsumerSuite.scala index a885c8064a7..c7712b1aaee 100644 --- a/connector/kafka-0-10/src/test/scala/org/apache/spark/streaming/kafka010/KafkaDataConsumerSuite.scala +++ b/connector/kafka-0-10/src/test/scala/org/apache/spark/streaming/kafka010/KafkaDataConsumerSuite.scala @@ -26,12 +26,11 @@ import org.apache.kafka.clients.consumer.ConsumerConfig._ import org.apache.kafka.common.TopicPartition import org.apache.kafka.common.serialization.ByteArrayDeserializer import org.mockito.Mockito.when -import org.scalatest.BeforeAndAfterAll import org.scalatestplus.mockito.MockitoSugar import org.apache.spark._ -class KafkaDataConsumerSuite extends SparkFunSuite with MockitoSugar with BeforeAndAfterAll { +class KafkaDataConsumerSuite extends SparkFunSuite with MockitoSugar { private var testUtils: KafkaTestUtils = _ private val topic = "topic" + Random.nextInt() private val topicPartition = new TopicPartition(topic, 0) diff --git a/connector/kafka-0-10/src/test/scala/org/apache/spark/streaming/kafka010/KafkaRDDSuite.scala b/connector/kafka-0-10/src/test/scala/org/apache/spark/streaming/kafka010/KafkaRDDSuite.scala index a5c8719212a..b25e6c8e459 100644 --- a/connector/kafka-0-10/src/test/scala/org/apache/spark/streaming/kafka010/KafkaRDDSuite.scala +++ b/connector/kafka-0-10/src/test/scala/org/apache/spark/streaming/kafka010/KafkaRDDSuite.scala @@ -30,14 +30,13 @@ import kafka.utils.Pool import org.apache.kafka.common.TopicPartition import org.apache.kafka.common.record.{CompressionType, MemoryRecords, SimpleRecord} import org.apache.kafka.common.serialization.StringDeserializer -import org.scalatest.BeforeAndAfterAll import org.scalatest.concurrent.Eventually.{eventually, interval, timeout} import org.apache.spark._ import org.apache.spark.scheduler.ExecutorCacheTaskLocation import org.apache.spark.streaming.kafka010.mocks.MockTime -class KafkaRDDSuite extends SparkFunSuite with BeforeAndAfterAll { +class KafkaRDDSuite extends SparkFunSuite { private var kafkaTestUtils: KafkaTestUtils = _ diff --git a/core/src/test/scala/org/apache/spark/SSLOptionsSuite.scala b/core/src/test/scala/org/apache/spark/SSLOptionsSuite.scala index 57d33971a57..c990d81de2e 100644 --- a/core/src/test/scala/org/apache/spark/SSLOptionsSuite.scala +++ b/core/src/test/scala/org/apache/spark/SSLOptionsSuite.scala @@ -23,11 +23,10 @@ import javax.net.ssl.SSLContext import org.apache.hadoop.conf.Configuration import org.apache.hadoop.security.alias.{CredentialProvider, CredentialProviderFactory} -import org.scalatest.BeforeAndAfterAll import org.apache.spark.util.SparkConfWithEnv -class SSLOptionsSuite extends SparkFunSuite with BeforeAndAfterAll { +class SSLOptionsSuite extends SparkFunSuite { test("test resolving property file as spark conf ") { val keyStorePath = new File(this.getClass.getResource("/keystore").toURI).getAbsolutePath diff --git a/core/src/test/scala/org/apache/spark/SparkContextSchedulerCreationSuite.scala b/core/src/test/scala/org/apache/spark/SparkContextSchedulerCreationSuite.scala index 3a615d0ea6c..5dfd6ea23cb 100644 --- a/core/src/test/scala/org/apache/spark/SparkContextSchedulerCreationSuite.scala +++ b/core/src/test/scala/org/apache/spark/SparkContextSchedulerCreationSuite.scala @@ -19,15 +19,13 @@ package org.apache.spark import org.scalatest.PrivateMethodTester -import org.apache.spark.internal.Logging import org.apache.spark.scheduler.{SchedulerBackend, TaskScheduler, TaskSchedulerImpl} import org.apache.spark.scheduler.cluster.StandaloneSchedulerBackend import org.apache.spark.scheduler.local.LocalSchedulerBackend import org.apache.spark.util.Utils - class SparkContextSchedulerCreationSuite - extends SparkFunSuite with LocalSparkContext with PrivateMethodTester with Logging { + extends SparkFunSuite with LocalSparkContext with PrivateMethodTester { def noOp(taskSchedulerImpl: TaskSchedulerImpl): Unit = {} diff --git a/core/src/test/scala/org/apache/spark/ThreadingSuite.scala b/core/src/test/scala/org/apache/spark/ThreadingSuite.scala index bb04d0d2632..0ee6523e5b8 100644 --- a/core/src/test/scala/org/apache/spark/ThreadingSuite.scala +++ b/core/src/test/scala/org/apache/spark/ThreadingSuite.scala @@ -20,8 +20,6 @@ package org.apache.spark import java.util.concurrent.{Semaphore, TimeUnit} import java.util.concurrent.atomic.{AtomicBoolean, AtomicInteger} -import org.apache.spark.internal.Logging - /** * Holds state shared across task threads in some ThreadingSuite tests. */ @@ -35,7 +33,7 @@ object ThreadingSuiteState { } } -class ThreadingSuite extends SparkFunSuite with LocalSparkContext with Logging { +class ThreadingSuite extends SparkFunSuite with LocalSparkContext { test("accessing SparkContext form a different thread") { sc = new SparkContext("local", "test") diff --git a/core/src/test/scala/org/apache/spark/deploy/SparkSubmitUtilsSuite.scala b/core/src/test/scala/org/apache/spark/deploy/SparkSubmitUtilsSuite.scala index c118c58cafe..db99a020bc9 100644 --- a/core/src/test/scala/org/apache/spark/deploy/SparkSubmitUtilsSuite.scala +++ b/core/src/test/scala/org/apache/spark/deploy/SparkSubmitUtilsSuite.scala @@ -28,13 +28,12 @@ import scala.collection.mutable.ArrayBuffer import org.apache.ivy.core.module.descriptor.MDArtifact import org.apache.ivy.core.settings.IvySettings import org.apache.ivy.plugins.resolver.{AbstractResolver, ChainResolver, FileSystemResolver, IBiblioResolver} -import org.scalatest.BeforeAndAfterAll import org.apache.spark.SparkFunSuite import org.apache.spark.deploy.SparkSubmitUtils.MavenCoordinate import org.apache.spark.util.{DependencyUtils, Utils} -class SparkSubmitUtilsSuite extends SparkFunSuite with BeforeAndAfterAll { +class SparkSubmitUtilsSuite extends SparkFunSuite { private var tempIvyPath: String = _ diff --git a/core/src/test/scala/org/apache/spark/deploy/history/ApplicationCacheSuite.scala b/core/src/test/scala/org/apache/spark/deploy/history/ApplicationCacheSuite.scala index 7cf533e58b6..25d668ad75c 100644 --- a/core/src/test/scala/org/apache/spark/deploy/history/ApplicationCacheSuite.scala +++ b/core/src/test/scala/org/apache/spark/deploy/history/ApplicationCacheSuite.scala @@ -37,7 +37,7 @@ import org.apache.spark.status.api.v1.{ApplicationAttemptInfo => AttemptInfo, Ap import org.apache.spark.ui.SparkUI import org.apache.spark.util.ManualClock -class ApplicationCacheSuite extends SparkFunSuite with Logging with MockitoSugar with Matchers { +class ApplicationCacheSuite extends SparkFunSuite with MockitoSugar with Matchers { /** * Stub cache operations. diff --git a/core/src/test/scala/org/apache/spark/deploy/history/FsHistoryProviderSuite.scala b/core/src/test/scala/org/apache/spark/deploy/history/FsHistoryProviderSuite.scala index afe8c61a524..e25a2cf3b67 100644 --- a/core/src/test/scala/org/apache/spark/deploy/history/FsHistoryProviderSuite.scala +++ b/core/src/test/scala/org/apache/spark/deploy/history/FsHistoryProviderSuite.scala @@ -40,7 +40,6 @@ import org.scalatest.matchers.should.Matchers._ import org.apache.spark.{JobExecutionStatus, SecurityManager, SPARK_VERSION, SparkConf, SparkFunSuite} import org.apache.spark.deploy.SparkHadoopUtil import org.apache.spark.deploy.history.EventLogTestHelper._ -import org.apache.spark.internal.Logging import org.apache.spark.internal.config.DRIVER_LOG_DFS_DIR import org.apache.spark.internal.config.History._ import org.apache.spark.internal.config.UI.{ADMIN_ACLS, ADMIN_ACLS_GROUPS, UI_VIEW_ACLS, UI_VIEW_ACLS_GROUPS, USER_GROUPS_MAPPING} @@ -57,8 +56,7 @@ import org.apache.spark.util.{Clock, JsonProtocol, ManualClock, Utils} import org.apache.spark.util.kvstore.InMemoryStore import org.apache.spark.util.logging.DriverLogger -abstract class FsHistoryProviderSuite extends SparkFunSuite with Matchers with Logging - with PrivateMethodTester { +abstract class FsHistoryProviderSuite extends SparkFunSuite with Matchers with PrivateMethodTester { private var testDir: File = null diff --git a/core/src/test/scala/org/apache/spark/deploy/history/RealBrowserUIHistoryServerSuite.scala b/core/src/test/scala/org/apache/spark/deploy/history/RealBrowserUIHistoryServerSuite.scala index ba8bd8d4404..7d2d31b4ada 100644 --- a/core/src/test/scala/org/apache/spark/deploy/history/RealBrowserUIHistoryServerSuite.scala +++ b/core/src/test/scala/org/apache/spark/deploy/history/RealBrowserUIHistoryServerSuite.scala @@ -22,7 +22,6 @@ import javax.servlet.http.HttpServletRequest import org.eclipse.jetty.proxy.ProxyServlet import org.eclipse.jetty.servlet.{ServletContextHandler, ServletHolder} import org.openqa.selenium.WebDriver -import org.scalatest.{BeforeAndAfterAll, BeforeAndAfterEach} import org.scalatest.matchers.must.Matchers import org.scalatest.matchers.should.Matchers._ import org.scalatestplus.selenium.WebBrowser @@ -37,8 +36,7 @@ import org.apache.spark.util.{ResetSystemProperties, Utils} * Tests for HistoryServer with real web browsers. */ abstract class RealBrowserUIHistoryServerSuite(val driverProp: String) - extends SparkFunSuite with WebBrowser with Matchers with BeforeAndAfterAll - with BeforeAndAfterEach with ResetSystemProperties { + extends SparkFunSuite with WebBrowser with Matchers with ResetSystemProperties { implicit var webDriver: WebDriver diff --git a/core/src/test/scala/org/apache/spark/deploy/master/ui/MasterWebUISuite.scala b/core/src/test/scala/org/apache/spark/deploy/master/ui/MasterWebUISuite.scala index b28651ea79c..024511189ac 100644 --- a/core/src/test/scala/org/apache/spark/deploy/master/ui/MasterWebUISuite.scala +++ b/core/src/test/scala/org/apache/spark/deploy/master/ui/MasterWebUISuite.scala @@ -25,7 +25,6 @@ import java.util.Date import scala.collection.mutable.HashMap import org.mockito.Mockito.{mock, times, verify, when} -import org.scalatest.BeforeAndAfterAll import org.apache.spark.{SecurityManager, SparkConf, SparkFunSuite} import org.apache.spark.deploy.DeployMessages.{DecommissionWorkersOnHosts, KillDriverResponse, RequestKillDriver} @@ -34,8 +33,7 @@ import org.apache.spark.deploy.master._ import org.apache.spark.rpc.{RpcEndpointRef, RpcEnv} import org.apache.spark.util.Utils - -class MasterWebUISuite extends SparkFunSuite with BeforeAndAfterAll { +class MasterWebUISuite extends SparkFunSuite { val conf = new SparkConf() val securityMgr = new SecurityManager(conf) diff --git a/core/src/test/scala/org/apache/spark/deploy/rest/StandaloneRestSubmitSuite.scala b/core/src/test/scala/org/apache/spark/deploy/rest/StandaloneRestSubmitSuite.scala index d5312845a3b..696765297e7 100644 --- a/core/src/test/scala/org/apache/spark/deploy/rest/StandaloneRestSubmitSuite.scala +++ b/core/src/test/scala/org/apache/spark/deploy/rest/StandaloneRestSubmitSuite.scala @@ -26,7 +26,6 @@ import scala.collection.mutable import org.json4s.JsonAST._ import org.json4s.jackson.JsonMethods._ -import org.scalatest.BeforeAndAfterEach import org.apache.spark._ import org.apache.spark.deploy.{SparkSubmit, SparkSubmitArguments} @@ -38,7 +37,7 @@ import org.apache.spark.util.Utils /** * Tests for the REST application submission protocol used in standalone cluster mode. */ -class StandaloneRestSubmitSuite extends SparkFunSuite with BeforeAndAfterEach { +class StandaloneRestSubmitSuite extends SparkFunSuite { private var rpcEnv: Option[RpcEnv] = None private var server: Option[RestSubmissionServer] = None diff --git a/core/src/test/scala/org/apache/spark/input/WholeTextFileInputFormatSuite.scala b/core/src/test/scala/org/apache/spark/input/WholeTextFileInputFormatSuite.scala index 9a21ea6dafc..f8217b10a77 100644 --- a/core/src/test/scala/org/apache/spark/input/WholeTextFileInputFormatSuite.scala +++ b/core/src/test/scala/org/apache/spark/input/WholeTextFileInputFormatSuite.scala @@ -21,17 +21,14 @@ import java.io.{DataOutputStream, File, FileOutputStream} import scala.collection.immutable.IndexedSeq -import org.scalatest.BeforeAndAfterAll - import org.apache.spark.{SparkConf, SparkContext, SparkFunSuite} -import org.apache.spark.internal.Logging /** * Tests the correctness of * [[org.apache.spark.input.WholeTextFileInputFormat WholeTextFileInputFormat]]. A temporary * directory containing files is created as fake input which is deleted in the end. */ -class WholeTextFileInputFormatSuite extends SparkFunSuite with BeforeAndAfterAll with Logging { +class WholeTextFileInputFormatSuite extends SparkFunSuite { private var sc: SparkContext = _ override def beforeAll(): Unit = { diff --git a/core/src/test/scala/org/apache/spark/input/WholeTextFileRecordReaderSuite.scala b/core/src/test/scala/org/apache/spark/input/WholeTextFileRecordReaderSuite.scala index 5c31d7f0a85..c833d22b3be 100644 --- a/core/src/test/scala/org/apache/spark/input/WholeTextFileRecordReaderSuite.scala +++ b/core/src/test/scala/org/apache/spark/input/WholeTextFileRecordReaderSuite.scala @@ -26,17 +26,15 @@ import scala.collection.immutable.IndexedSeq import org.apache.hadoop.conf.Configuration import org.apache.hadoop.io.Text import org.apache.hadoop.io.compress.{CompressionCodecFactory, GzipCodec} -import org.scalatest.BeforeAndAfterAll import org.apache.spark.{SparkConf, SparkContext, SparkFunSuite} -import org.apache.spark.internal.Logging /** * Tests the correctness of * [[org.apache.spark.input.WholeTextFileRecordReader WholeTextFileRecordReader]]. A temporary * directory is created as fake input. Temporal storage would be deleted in the end. */ -class WholeTextFileRecordReaderSuite extends SparkFunSuite with BeforeAndAfterAll with Logging { +class WholeTextFileRecordReaderSuite extends SparkFunSuite { private var sc: SparkContext = _ private var factory: CompressionCodecFactory = _ diff --git a/core/src/test/scala/org/apache/spark/internal/plugin/PluginContainerSuite.scala b/core/src/test/scala/org/apache/spark/internal/plugin/PluginContainerSuite.scala index 9ef81d30ff1..2b8515d52d1 100644 --- a/core/src/test/scala/org/apache/spark/internal/plugin/PluginContainerSuite.scala +++ b/core/src/test/scala/org/apache/spark/internal/plugin/PluginContainerSuite.scala @@ -29,7 +29,6 @@ import com.codahale.metrics.Gauge import com.google.common.io.Files import org.mockito.ArgumentMatchers.{any, eq => meq} import org.mockito.Mockito.{mock, spy, verify, when} -import org.scalatest.BeforeAndAfterEach import org.scalatest.concurrent.Eventually.{eventually, interval, timeout} import org.apache.spark._ @@ -42,7 +41,7 @@ import org.apache.spark.resource.ResourceUtils.GPU import org.apache.spark.resource.TestResourceIDs.{DRIVER_GPU_ID, EXECUTOR_GPU_ID, WORKER_GPU_ID} import org.apache.spark.util.Utils -class PluginContainerSuite extends SparkFunSuite with BeforeAndAfterEach with LocalSparkContext { +class PluginContainerSuite extends SparkFunSuite with LocalSparkContext { override def afterEach(): Unit = { TestSparkPlugin.reset() diff --git a/core/src/test/scala/org/apache/spark/memory/MemoryManagerSuite.scala b/core/src/test/scala/org/apache/spark/memory/MemoryManagerSuite.scala index 1ecb50d2ad9..642216a7a47 100644 --- a/core/src/test/scala/org/apache/spark/memory/MemoryManagerSuite.scala +++ b/core/src/test/scala/org/apache/spark/memory/MemoryManagerSuite.scala @@ -27,7 +27,6 @@ import org.mockito.ArgumentMatchers.{any, anyLong} import org.mockito.Mockito.{mock, when, RETURNS_SMART_NULLS} import org.mockito.invocation.InvocationOnMock import org.mockito.stubbing.Answer -import org.scalatest.BeforeAndAfterEach import org.scalatest.time.SpanSugar._ import org.apache.spark.SparkFunSuite @@ -35,11 +34,10 @@ import org.apache.spark.storage.{BlockId, BlockStatus, StorageLevel} import org.apache.spark.storage.memory.MemoryStore import org.apache.spark.util.ThreadUtils - /** * Helper trait for sharing code among [[MemoryManager]] tests. */ -private[memory] trait MemoryManagerSuite extends SparkFunSuite with BeforeAndAfterEach { +private[memory] trait MemoryManagerSuite extends SparkFunSuite { protected val evictedBlocks = new mutable.ArrayBuffer[(BlockId, BlockStatus)] diff --git a/core/src/test/scala/org/apache/spark/rdd/AsyncRDDActionsSuite.scala b/core/src/test/scala/org/apache/spark/rdd/AsyncRDDActionsSuite.scala index 93daf903232..5e66ca962ea 100644 --- a/core/src/test/scala/org/apache/spark/rdd/AsyncRDDActionsSuite.scala +++ b/core/src/test/scala/org/apache/spark/rdd/AsyncRDDActionsSuite.scala @@ -25,14 +25,13 @@ import scala.concurrent.ExecutionContext.Implicits.global // scalastyle:on executioncontextglobal import scala.concurrent.duration.Duration -import org.scalatest.BeforeAndAfterAll import org.scalatest.concurrent.{Signaler, ThreadSignaler, TimeLimits} import org.scalatest.time.SpanSugar._ import org.apache.spark._ import org.apache.spark.util.ThreadUtils -class AsyncRDDActionsSuite extends SparkFunSuite with BeforeAndAfterAll with TimeLimits { +class AsyncRDDActionsSuite extends SparkFunSuite with TimeLimits { @transient private var sc: SparkContext = _ diff --git a/core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala b/core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala index 5b01b54a0a9..1ae2d51b125 100644 --- a/core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala +++ b/core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala @@ -21,9 +21,8 @@ import org.scalatest.matchers.must.Matchers import org.scalatest.matchers.should.Matchers._ import org.apache.spark.{SharedSparkContext, SparkFunSuite} -import org.apache.spark.internal.Logging -class SortingSuite extends SparkFunSuite with SharedSparkContext with Matchers with Logging { +class SortingSuite extends SparkFunSuite with SharedSparkContext with Matchers { test("sortByKey") { val pairs = sc.parallelize(Seq((1, 0), (2, 0), (0, 0), (3, 0)), 2) diff --git a/core/src/test/scala/org/apache/spark/rpc/RpcEnvSuite.scala b/core/src/test/scala/org/apache/spark/rpc/RpcEnvSuite.scala index 1681bf9b57b..c70dde79b3c 100644 --- a/core/src/test/scala/org/apache/spark/rpc/RpcEnvSuite.scala +++ b/core/src/test/scala/org/apache/spark/rpc/RpcEnvSuite.scala @@ -30,7 +30,6 @@ import scala.concurrent.duration._ import com.google.common.io.Files import org.mockito.ArgumentMatchers.any import org.mockito.Mockito.{mock, never, verify, when} -import org.scalatest.BeforeAndAfterAll import org.scalatest.concurrent.Eventually._ import org.apache.spark.{SparkConf, SparkEnv, SparkException, SparkFunSuite} @@ -41,7 +40,7 @@ import org.apache.spark.util.{ThreadUtils, Utils} /** * Common tests for an RpcEnv implementation. */ -abstract class RpcEnvSuite extends SparkFunSuite with BeforeAndAfterAll { +abstract class RpcEnvSuite extends SparkFunSuite { var env: RpcEnv = _ diff --git a/core/src/test/scala/org/apache/spark/scheduler/EventLoggingListenerSuite.scala b/core/src/test/scala/org/apache/spark/scheduler/EventLoggingListenerSuite.scala index 2a5a44ad783..31db0328f81 100644 --- a/core/src/test/scala/org/apache/spark/scheduler/EventLoggingListenerSuite.scala +++ b/core/src/test/scala/org/apache/spark/scheduler/EventLoggingListenerSuite.scala @@ -20,7 +20,6 @@ package org.apache.spark.scheduler import java.io.{File, InputStream} import java.util.{Arrays, Properties} -import scala.collection.immutable.Map import scala.collection.mutable import scala.collection.mutable.Set import scala.io.{Codec, Source} @@ -34,7 +33,6 @@ import org.apache.spark.deploy.SparkHadoopUtil import org.apache.spark.deploy.history.{EventLogFileReader, SingleEventLogFileWriter} import org.apache.spark.deploy.history.EventLogTestHelper._ import org.apache.spark.executor.{ExecutorMetrics, TaskMetrics} -import org.apache.spark.internal.Logging import org.apache.spark.internal.config.{EVENT_LOG_DIR, EVENT_LOG_ENABLED} import org.apache.spark.io._ import org.apache.spark.metrics.{ExecutorMetricType, MetricsSystem} @@ -49,8 +47,7 @@ import org.apache.spark.util.{JsonProtocol, Utils} * logging events, whether the parsing of the file names is correct, and whether the logged events * can be read and deserialized into actual SparkListenerEvents. */ -class EventLoggingListenerSuite extends SparkFunSuite with LocalSparkContext with BeforeAndAfter - with Logging { +class EventLoggingListenerSuite extends SparkFunSuite with LocalSparkContext with BeforeAndAfter { private val fileSystem = Utils.getHadoopFileSystem("/", SparkHadoopUtil.get.newConfiguration(new SparkConf())) diff --git a/core/src/test/scala/org/apache/spark/scheduler/HealthTrackerSuite.scala b/core/src/test/scala/org/apache/spark/scheduler/HealthTrackerSuite.scala index 4986aee2acc..c098f8d9a02 100644 --- a/core/src/test/scala/org/apache/spark/scheduler/HealthTrackerSuite.scala +++ b/core/src/test/scala/org/apache/spark/scheduler/HealthTrackerSuite.scala @@ -20,15 +20,13 @@ package org.apache.spark.scheduler import org.mockito.ArgumentMatchers.any import org.mockito.Mockito.{never, verify, when} import org.mockito.invocation.InvocationOnMock -import org.scalatest.BeforeAndAfterEach import org.scalatestplus.mockito.MockitoSugar import org.apache.spark._ import org.apache.spark.internal.config import org.apache.spark.util.ManualClock -class HealthTrackerSuite extends SparkFunSuite with BeforeAndAfterEach with MockitoSugar - with LocalSparkContext { +class HealthTrackerSuite extends SparkFunSuite with MockitoSugar with LocalSparkContext { private val clock = new ManualClock(0) diff --git a/core/src/test/scala/org/apache/spark/scheduler/TaskSchedulerImplSuite.scala b/core/src/test/scala/org/apache/spark/scheduler/TaskSchedulerImplSuite.scala index 71eda0063fe..6b71bc8a6ed 100644 --- a/core/src/test/scala/org/apache/spark/scheduler/TaskSchedulerImplSuite.scala +++ b/core/src/test/scala/org/apache/spark/scheduler/TaskSchedulerImplSuite.scala @@ -28,12 +28,10 @@ import scala.language.reflectiveCalls import org.mockito.ArgumentMatchers.{any, anyInt, anyString, eq => meq} import org.mockito.Mockito.{atLeast, atMost, never, spy, times, verify, when} -import org.scalatest.BeforeAndAfterEach import org.scalatest.concurrent.Eventually import org.scalatestplus.mockito.MockitoSugar import org.apache.spark._ -import org.apache.spark.internal.Logging import org.apache.spark.internal.config import org.apache.spark.resource.{ExecutorResourceRequests, ResourceProfile, TaskResourceRequests} import org.apache.spark.resource.ResourceUtils._ @@ -48,8 +46,8 @@ class FakeSchedulerBackend extends SchedulerBackend { def maxNumConcurrentTasks(rp: ResourceProfile): Int = 0 } -class TaskSchedulerImplSuite extends SparkFunSuite with LocalSparkContext with BeforeAndAfterEach - with Logging with MockitoSugar with Eventually { +class TaskSchedulerImplSuite extends SparkFunSuite with LocalSparkContext + with MockitoSugar with Eventually { var failedTaskSetException: Option[Throwable] = None var failedTaskSetReason: String = null diff --git a/core/src/test/scala/org/apache/spark/scheduler/TaskSetExcludelistSuite.scala b/core/src/test/scala/org/apache/spark/scheduler/TaskSetExcludelistSuite.scala index d20768d7cd1..de10cb9b163 100644 --- a/core/src/test/scala/org/apache/spark/scheduler/TaskSetExcludelistSuite.scala +++ b/core/src/test/scala/org/apache/spark/scheduler/TaskSetExcludelistSuite.scala @@ -18,14 +18,13 @@ package org.apache.spark.scheduler import org.mockito.ArgumentMatchers.isA import org.mockito.Mockito.{never, verify} -import org.scalatest.BeforeAndAfterEach import org.scalatestplus.mockito.MockitoSugar import org.apache.spark.{SparkConf, SparkFunSuite} import org.apache.spark.internal.config import org.apache.spark.util.ManualClock -class TaskSetExcludelistSuite extends SparkFunSuite with BeforeAndAfterEach with MockitoSugar { +class TaskSetExcludelistSuite extends SparkFunSuite with MockitoSugar { private var listenerBusMock: LiveListenerBus = _ diff --git a/core/src/test/scala/org/apache/spark/serializer/SerializationDebuggerSuite.scala b/core/src/test/scala/org/apache/spark/serializer/SerializationDebuggerSuite.scala index 912a516dff0..e903cf31d69 100644 --- a/core/src/test/scala/org/apache/spark/serializer/SerializationDebuggerSuite.scala +++ b/core/src/test/scala/org/apache/spark/serializer/SerializationDebuggerSuite.scala @@ -21,12 +21,9 @@ import java.io._ import scala.annotation.meta.param -import org.scalatest.BeforeAndAfterEach - import org.apache.spark.SparkFunSuite - -class SerializationDebuggerSuite extends SparkFunSuite with BeforeAndAfterEach { +class SerializationDebuggerSuite extends SparkFunSuite { import SerializationDebugger.find diff --git a/core/src/test/scala/org/apache/spark/shuffle/ShuffleBlockPusherSuite.scala b/core/src/test/scala/org/apache/spark/shuffle/ShuffleBlockPusherSuite.scala index 94c0417b375..c8d89625dd8 100644 --- a/core/src/test/scala/org/apache/spark/shuffle/ShuffleBlockPusherSuite.scala +++ b/core/src/test/scala/org/apache/spark/shuffle/ShuffleBlockPusherSuite.scala @@ -29,7 +29,6 @@ import org.mockito.Answers.RETURNS_SMART_NULLS import org.mockito.ArgumentMatchers.any import org.mockito.Mockito._ import org.mockito.invocation.InvocationOnMock -import org.scalatest.BeforeAndAfterEach import org.apache.spark._ import org.apache.spark.executor.CoarseGrainedExecutorBackend @@ -44,7 +43,7 @@ import org.apache.spark.shuffle.ShuffleBlockPusher.PushRequest import org.apache.spark.storage._ import org.apache.spark.util.ThreadUtils -class ShuffleBlockPusherSuite extends SparkFunSuite with BeforeAndAfterEach { +class ShuffleBlockPusherSuite extends SparkFunSuite { @Mock(answer = RETURNS_SMART_NULLS) private var blockManager: BlockManager = _ @Mock(answer = RETURNS_SMART_NULLS) private var dependency: ShuffleDependency[Int, Int, Int] = _ diff --git a/core/src/test/scala/org/apache/spark/shuffle/ShuffleDriverComponentsSuite.scala b/core/src/test/scala/org/apache/spark/shuffle/ShuffleDriverComponentsSuite.scala index 3d70ff1fed2..f67204649ce 100644 --- a/core/src/test/scala/org/apache/spark/shuffle/ShuffleDriverComponentsSuite.scala +++ b/core/src/test/scala/org/apache/spark/shuffle/ShuffleDriverComponentsSuite.scala @@ -22,15 +22,13 @@ import java.util.concurrent.atomic.AtomicBoolean import com.google.common.collect.ImmutableMap import org.scalatest.Assertions._ -import org.scalatest.BeforeAndAfterEach import org.apache.spark.{LocalSparkContext, SparkConf, SparkContext, SparkFunSuite} import org.apache.spark.internal.config.SHUFFLE_IO_PLUGIN_CLASS import org.apache.spark.shuffle.api.{ShuffleDataIO, ShuffleDriverComponents, ShuffleExecutorComponents, ShuffleMapOutputWriter} import org.apache.spark.shuffle.sort.io.LocalDiskShuffleDataIO -class ShuffleDriverComponentsSuite - extends SparkFunSuite with LocalSparkContext with BeforeAndAfterEach { +class ShuffleDriverComponentsSuite extends SparkFunSuite with LocalSparkContext { test("test serialization of shuffle initialization conf to executors") { val testConf = new SparkConf() diff --git a/core/src/test/scala/org/apache/spark/shuffle/sort/IndexShuffleBlockResolverSuite.scala b/core/src/test/scala/org/apache/spark/shuffle/sort/IndexShuffleBlockResolverSuite.scala index de12f6840a1..31b255cff72 100644 --- a/core/src/test/scala/org/apache/spark/shuffle/sort/IndexShuffleBlockResolverSuite.scala +++ b/core/src/test/scala/org/apache/spark/shuffle/sort/IndexShuffleBlockResolverSuite.scala @@ -25,7 +25,6 @@ import org.mockito.ArgumentMatchers.any import org.mockito.Mockito._ import org.mockito.invocation.InvocationOnMock import org.roaringbitmap.RoaringBitmap -import org.scalatest.BeforeAndAfterEach import org.apache.spark.{SparkConf, SparkFunSuite} import org.apache.spark.internal.config @@ -33,7 +32,7 @@ import org.apache.spark.shuffle.{IndexShuffleBlockResolver, ShuffleBlockInfo} import org.apache.spark.storage._ import org.apache.spark.util.Utils -class IndexShuffleBlockResolverSuite extends SparkFunSuite with BeforeAndAfterEach { +class IndexShuffleBlockResolverSuite extends SparkFunSuite { @Mock(answer = RETURNS_SMART_NULLS) private var blockManager: BlockManager = _ @Mock(answer = RETURNS_SMART_NULLS) private var diskBlockManager: DiskBlockManager = _ diff --git a/core/src/test/scala/org/apache/spark/shuffle/sort/io/LocalDiskShuffleMapOutputWriterSuite.scala b/core/src/test/scala/org/apache/spark/shuffle/sort/io/LocalDiskShuffleMapOutputWriterSuite.scala index 6c9ec8b71a4..3db75272625 100644 --- a/core/src/test/scala/org/apache/spark/shuffle/sort/io/LocalDiskShuffleMapOutputWriterSuite.scala +++ b/core/src/test/scala/org/apache/spark/shuffle/sort/io/LocalDiskShuffleMapOutputWriterSuite.scala @@ -27,13 +27,12 @@ import org.mockito.ArgumentMatchers.{any, anyInt, anyLong} import org.mockito.Mock import org.mockito.Mockito.when import org.mockito.MockitoAnnotations -import org.scalatest.BeforeAndAfterEach import org.apache.spark.{SparkConf, SparkFunSuite} import org.apache.spark.shuffle.IndexShuffleBlockResolver import org.apache.spark.util.Utils -class LocalDiskShuffleMapOutputWriterSuite extends SparkFunSuite with BeforeAndAfterEach { +class LocalDiskShuffleMapOutputWriterSuite extends SparkFunSuite { @Mock(answer = RETURNS_SMART_NULLS) private var blockResolver: IndexShuffleBlockResolver = _ diff --git a/core/src/test/scala/org/apache/spark/storage/BlockInfoManagerSuite.scala b/core/src/test/scala/org/apache/spark/storage/BlockInfoManagerSuite.scala index 5321931ed92..a1cd1dbc9be 100644 --- a/core/src/test/scala/org/apache/spark/storage/BlockInfoManagerSuite.scala +++ b/core/src/test/scala/org/apache/spark/storage/BlockInfoManagerSuite.scala @@ -23,15 +23,13 @@ import scala.concurrent.{ExecutionContext, Future} import scala.language.implicitConversions import scala.reflect.ClassTag -import org.scalatest.BeforeAndAfterEach import org.scalatest.time.SpanSugar._ import org.apache.spark.{SparkException, SparkFunSuite, TaskContext, TaskContextImpl} import org.apache.spark.executor.TaskMetrics import org.apache.spark.util.ThreadUtils - -class BlockInfoManagerSuite extends SparkFunSuite with BeforeAndAfterEach { +class BlockInfoManagerSuite extends SparkFunSuite { private implicit val ec = ExecutionContext.global private var blockInfoManager: BlockInfoManager = _ diff --git a/core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala b/core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala index 545bbee6671..ff11fc8ebb1 100644 --- a/core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala +++ b/core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala @@ -33,7 +33,7 @@ import com.esotericsoftware.kryo.KryoException import org.apache.commons.lang3.RandomUtils import org.mockito.{ArgumentCaptor, ArgumentMatchers => mc} import org.mockito.Mockito.{doAnswer, mock, never, spy, times, verify, when} -import org.scalatest.{BeforeAndAfterAll, BeforeAndAfterEach, PrivateMethodTester} +import org.scalatest.PrivateMethodTester import org.scalatest.concurrent.{Signaler, ThreadSignaler, TimeLimits} import org.scalatest.concurrent.Eventually._ import org.scalatest.matchers.must.Matchers @@ -66,9 +66,8 @@ import org.apache.spark.storage.BlockManagerMessages._ import org.apache.spark.util._ import org.apache.spark.util.io.ChunkedByteBuffer -class BlockManagerSuite extends SparkFunSuite with Matchers with BeforeAndAfterEach - with PrivateMethodTester with LocalSparkContext with ResetSystemProperties - with EncryptionFunSuite with TimeLimits with BeforeAndAfterAll { +class BlockManagerSuite extends SparkFunSuite with Matchers with PrivateMethodTester + with LocalSparkContext with ResetSystemProperties with EncryptionFunSuite with TimeLimits { import BlockManagerSuite._ diff --git a/core/src/test/scala/org/apache/spark/storage/DiskBlockManagerSuite.scala b/core/src/test/scala/org/apache/spark/storage/DiskBlockManagerSuite.scala index 3e4002614ca..ac896c0b17a 100644 --- a/core/src/test/scala/org/apache/spark/storage/DiskBlockManagerSuite.scala +++ b/core/src/test/scala/org/apache/spark/storage/DiskBlockManagerSuite.scala @@ -26,14 +26,12 @@ import com.fasterxml.jackson.core.`type`.TypeReference import com.fasterxml.jackson.databind.ObjectMapper import jnr.posix.{POSIX, POSIXFactory} import org.apache.commons.io.FileUtils -import org.scalatest.{BeforeAndAfterAll, BeforeAndAfterEach} import org.apache.spark.{SparkConf, SparkFunSuite} import org.apache.spark.internal.config import org.apache.spark.util.Utils - -class DiskBlockManagerSuite extends SparkFunSuite with BeforeAndAfterEach with BeforeAndAfterAll { +class DiskBlockManagerSuite extends SparkFunSuite { private val testConf = new SparkConf(false) private var rootDir0: File = _ private var rootDir1: File = _ diff --git a/core/src/test/scala/org/apache/spark/storage/DiskBlockObjectWriterSuite.scala b/core/src/test/scala/org/apache/spark/storage/DiskBlockObjectWriterSuite.scala index b1f9032c247..e6bf01b4b65 100644 --- a/core/src/test/scala/org/apache/spark/storage/DiskBlockObjectWriterSuite.scala +++ b/core/src/test/scala/org/apache/spark/storage/DiskBlockObjectWriterSuite.scala @@ -18,14 +18,12 @@ package org.apache.spark.storage import java.io.File -import org.scalatest.BeforeAndAfterEach - import org.apache.spark.{SparkConf, SparkFunSuite} import org.apache.spark.executor.ShuffleWriteMetrics import org.apache.spark.serializer.{JavaSerializer, SerializerManager} import org.apache.spark.util.Utils -class DiskBlockObjectWriterSuite extends SparkFunSuite with BeforeAndAfterEach { +class DiskBlockObjectWriterSuite extends SparkFunSuite { var tempDir: File = _ diff --git a/core/src/test/scala/org/apache/spark/ui/RealBrowserUISeleniumSuite.scala b/core/src/test/scala/org/apache/spark/ui/RealBrowserUISeleniumSuite.scala index 3f296acdeb3..185d91ec25e 100644 --- a/core/src/test/scala/org/apache/spark/ui/RealBrowserUISeleniumSuite.scala +++ b/core/src/test/scala/org/apache/spark/ui/RealBrowserUISeleniumSuite.scala @@ -18,7 +18,6 @@ package org.apache.spark.ui import org.openqa.selenium.{By, JavascriptExecutor, WebDriver} -import org.scalatest.BeforeAndAfterAll import org.scalatest.concurrent.Eventually._ import org.scalatest.matchers.must.Matchers import org.scalatest.matchers.should.Matchers._ @@ -35,7 +34,7 @@ import org.apache.spark.util.CallSite * Selenium tests for the Spark Web UI with real web browsers. */ abstract class RealBrowserUISeleniumSuite(val driverProp: String) - extends SparkFunSuite with WebBrowser with Matchers with BeforeAndAfterAll { + extends SparkFunSuite with WebBrowser with Matchers { implicit var webDriver: WebDriver with JavascriptExecutor private val driverPropPrefix = "spark.test." diff --git a/core/src/test/scala/org/apache/spark/ui/UISeleniumSuite.scala b/core/src/test/scala/org/apache/spark/ui/UISeleniumSuite.scala index 47fffa0572b..874ef1552d8 100644 --- a/core/src/test/scala/org/apache/spark/ui/UISeleniumSuite.scala +++ b/core/src/test/scala/org/apache/spark/ui/UISeleniumSuite.scala @@ -30,7 +30,6 @@ import org.json4s._ import org.json4s.jackson.JsonMethods import org.openqa.selenium.{By, WebDriver} import org.openqa.selenium.htmlunit.HtmlUnitDriver -import org.scalatest.BeforeAndAfterAll import org.scalatest.concurrent.Eventually._ import org.scalatest.matchers.must.Matchers import org.scalatest.matchers.should.Matchers._ @@ -80,7 +79,7 @@ private[spark] class SparkUICssErrorHandler extends DefaultCssErrorHandler { /** * Selenium tests for the Spark Web UI. */ -class UISeleniumSuite extends SparkFunSuite with WebBrowser with Matchers with BeforeAndAfterAll { +class UISeleniumSuite extends SparkFunSuite with WebBrowser with Matchers { implicit var webDriver: WebDriver = _ implicit val formats = DefaultFormats diff --git a/core/src/test/scala/org/apache/spark/util/FileAppenderSuite.scala b/core/src/test/scala/org/apache/spark/util/FileAppenderSuite.scala index e79e3256fe1..ef46983afb4 100644 --- a/core/src/test/scala/org/apache/spark/util/FileAppenderSuite.scala +++ b/core/src/test/scala/org/apache/spark/util/FileAppenderSuite.scala @@ -34,10 +34,10 @@ import org.mockito.Mockito.{atLeast, mock, verify, when} import org.scalatest.BeforeAndAfter import org.apache.spark.{SparkConf, SparkFunSuite} -import org.apache.spark.internal.{config, Logging} +import org.apache.spark.internal.config import org.apache.spark.util.logging.{FileAppender, RollingFileAppender, SizeBasedRollingPolicy, TimeBasedRollingPolicy} -class FileAppenderSuite extends SparkFunSuite with BeforeAndAfter with Logging { +class FileAppenderSuite extends SparkFunSuite with BeforeAndAfter { val testFile = new File(Utils.createTempDir(), "FileAppenderSuite-test").getAbsoluteFile diff --git a/core/src/test/scala/org/apache/spark/util/UtilsSuite.scala b/core/src/test/scala/org/apache/spark/util/UtilsSuite.scala index 973c09884a6..3a1f0862e94 100644 --- a/core/src/test/scala/org/apache/spark/util/UtilsSuite.scala +++ b/core/src/test/scala/org/apache/spark/util/UtilsSuite.scala @@ -39,7 +39,6 @@ import org.apache.hadoop.fs.Path import org.apache.logging.log4j.Level import org.apache.spark.{SparkConf, SparkException, SparkFunSuite, TaskContext} -import org.apache.spark.internal.Logging import org.apache.spark.internal.config._ import org.apache.spark.internal.config.Tests.IS_TESTING import org.apache.spark.launcher.SparkLauncher @@ -47,7 +46,7 @@ import org.apache.spark.network.util.ByteUnit import org.apache.spark.scheduler.SparkListener import org.apache.spark.util.io.ChunkedByteBufferInputStream -class UtilsSuite extends SparkFunSuite with ResetSystemProperties with Logging { +class UtilsSuite extends SparkFunSuite with ResetSystemProperties { test("timeConversion") { // Test -1 diff --git a/core/src/test/scala/org/apache/spark/util/collection/ExternalSorterSpillSuite.scala b/core/src/test/scala/org/apache/spark/util/collection/ExternalSorterSpillSuite.scala index 959d5d813df..ab8999c86c3 100644 --- a/core/src/test/scala/org/apache/spark/util/collection/ExternalSorterSpillSuite.scala +++ b/core/src/test/scala/org/apache/spark/util/collection/ExternalSorterSpillSuite.scala @@ -25,7 +25,6 @@ import scala.collection.mutable.ArrayBuffer import org.mockito.ArgumentMatchers.{any, anyInt} import org.mockito.Mockito.{mock, when} import org.mockito.invocation.InvocationOnMock -import org.scalatest.BeforeAndAfterEach import org.apache.spark.{SparkConf, SparkEnv, SparkFunSuite, TaskContext} import org.apache.spark.executor.ShuffleWriteMetrics @@ -35,7 +34,7 @@ import org.apache.spark.serializer.{KryoSerializer, SerializerInstance, Serializ import org.apache.spark.storage.{BlockId, BlockManager, DiskBlockManager, DiskBlockObjectWriter, TempShuffleBlockId} import org.apache.spark.util.{Utils => UUtils} -class ExternalSorterSpillSuite extends SparkFunSuite with BeforeAndAfterEach { +class ExternalSorterSpillSuite extends SparkFunSuite { private val spillFilesCreated = ArrayBuffer.empty[File] diff --git a/core/src/test/scala/org/apache/spark/util/collection/SorterSuite.scala b/core/src/test/scala/org/apache/spark/util/collection/SorterSuite.scala index 82b4f2eac62..e3a57a70e2d 100644 --- a/core/src/test/scala/org/apache/spark/util/collection/SorterSuite.scala +++ b/core/src/test/scala/org/apache/spark/util/collection/SorterSuite.scala @@ -22,11 +22,10 @@ import java.util.Arrays import java.util.concurrent.TimeUnit import org.apache.spark.SparkFunSuite -import org.apache.spark.internal.Logging import org.apache.spark.util.Utils.timeIt import org.apache.spark.util.random.XORShiftRandom -class SorterSuite extends SparkFunSuite with Logging { +class SorterSuite extends SparkFunSuite { test("equivalent to Arrays.sort") { val rand = new XORShiftRandom(123) diff --git a/core/src/test/scala/org/apache/spark/util/collection/unsafe/sort/RadixSortSuite.scala b/core/src/test/scala/org/apache/spark/util/collection/unsafe/sort/RadixSortSuite.scala index b3e5e0a73dd..d33ac97553d 100644 --- a/core/src/test/scala/org/apache/spark/util/collection/unsafe/sort/RadixSortSuite.scala +++ b/core/src/test/scala/org/apache/spark/util/collection/unsafe/sort/RadixSortSuite.scala @@ -25,13 +25,12 @@ import scala.util.Random import com.google.common.primitives.Ints import org.apache.spark.SparkFunSuite -import org.apache.spark.internal.Logging import org.apache.spark.unsafe.array.LongArray import org.apache.spark.unsafe.memory.MemoryBlock import org.apache.spark.util.collection.Sorter import org.apache.spark.util.random.XORShiftRandom -class RadixSortSuite extends SparkFunSuite with Logging { +class RadixSortSuite extends SparkFunSuite { private val N = 10000L // scale this down for more readable results /** diff --git a/mllib/src/test/scala/org/apache/spark/ml/MLEventsSuite.scala b/mllib/src/test/scala/org/apache/spark/ml/MLEventsSuite.scala index f7e7e7fe2a7..d58c938b7d0 100644 --- a/mllib/src/test/scala/org/apache/spark/ml/MLEventsSuite.scala +++ b/mllib/src/test/scala/org/apache/spark/ml/MLEventsSuite.scala @@ -23,7 +23,6 @@ import scala.concurrent.duration._ import org.apache.hadoop.fs.Path import org.mockito.ArgumentMatchers.{any, eq => meq} import org.mockito.Mockito.when -import org.scalatest.BeforeAndAfterEach import org.scalatest.concurrent.Eventually import org.scalatestplus.mockito.MockitoSugar.mock @@ -35,9 +34,7 @@ import org.apache.spark.scheduler.{SparkListener, SparkListenerEvent} import org.apache.spark.sql._ import org.apache.spark.util.JsonProtocol - -class MLEventsSuite - extends SparkFunSuite with BeforeAndAfterEach with MLlibTestSparkContext with Eventually { +class MLEventsSuite extends SparkFunSuite with MLlibTestSparkContext with Eventually { private val events = mutable.ArrayBuffer.empty[MLEvent] private val listener: SparkListener = new SparkListener { diff --git a/mllib/src/test/scala/org/apache/spark/ml/recommendation/ALSSuite.scala b/mllib/src/test/scala/org/apache/spark/ml/recommendation/ALSSuite.scala index 6f93365059d..f852a4243ea 100644 --- a/mllib/src/test/scala/org/apache/spark/ml/recommendation/ALSSuite.scala +++ b/mllib/src/test/scala/org/apache/spark/ml/recommendation/ALSSuite.scala @@ -1072,8 +1072,7 @@ class ALSCleanerSuite extends SparkFunSuite with LocalRootDirsTest { } } -class ALSStorageSuite - extends SparkFunSuite with MLlibTestSparkContext with DefaultReadWriteTest with Logging { +class ALSStorageSuite extends SparkFunSuite with MLlibTestSparkContext with DefaultReadWriteTest { test("invalid storage params") { intercept[IllegalArgumentException] { diff --git a/mllib/src/test/scala/org/apache/spark/ml/stat/CorrelationSuite.scala b/mllib/src/test/scala/org/apache/spark/ml/stat/CorrelationSuite.scala index 7d935e651f2..6813a50d7b1 100644 --- a/mllib/src/test/scala/org/apache/spark/ml/stat/CorrelationSuite.scala +++ b/mllib/src/test/scala/org/apache/spark/ml/stat/CorrelationSuite.scala @@ -20,14 +20,12 @@ package org.apache.spark.ml.stat import breeze.linalg.{DenseMatrix => BDM} import org.apache.spark.SparkFunSuite -import org.apache.spark.internal.Logging import org.apache.spark.ml.linalg.{Matrices, Matrix, Vectors} import org.apache.spark.ml.util.TestingUtils._ import org.apache.spark.mllib.util.MLlibTestSparkContext import org.apache.spark.sql.{DataFrame, Row} - -class CorrelationSuite extends SparkFunSuite with MLlibTestSparkContext with Logging { +class CorrelationSuite extends SparkFunSuite with MLlibTestSparkContext { val xData = Array(1.0, 0.0, -2.0) val yData = Array(4.0, 5.0, 3.0) diff --git a/mllib/src/test/scala/org/apache/spark/ml/tree/impl/GradientBoostedTreesSuite.scala b/mllib/src/test/scala/org/apache/spark/ml/tree/impl/GradientBoostedTreesSuite.scala index 18fc1407557..3a8fcd9cc4e 100644 --- a/mllib/src/test/scala/org/apache/spark/ml/tree/impl/GradientBoostedTreesSuite.scala +++ b/mllib/src/test/scala/org/apache/spark/ml/tree/impl/GradientBoostedTreesSuite.scala @@ -18,7 +18,6 @@ package org.apache.spark.ml.tree.impl import org.apache.spark.SparkFunSuite -import org.apache.spark.internal.Logging import org.apache.spark.ml.feature.Instance import org.apache.spark.mllib.tree.{GradientBoostedTreesSuite => OldGBTSuite} import org.apache.spark.mllib.tree.configuration.{BoostingStrategy, Strategy} @@ -30,7 +29,7 @@ import org.apache.spark.mllib.util.MLlibTestSparkContext /** * Test suite for [[GradientBoostedTrees]]. */ -class GradientBoostedTreesSuite extends SparkFunSuite with MLlibTestSparkContext with Logging { +class GradientBoostedTreesSuite extends SparkFunSuite with MLlibTestSparkContext { test("runWithValidation stops early and performs better on a validation dataset") { // Set numIterations large enough so that it stops early. diff --git a/mllib/src/test/scala/org/apache/spark/mllib/linalg/VectorsSuite.scala b/mllib/src/test/scala/org/apache/spark/mllib/linalg/VectorsSuite.scala index 70ba4d3049b..ca452a8a69b 100644 --- a/mllib/src/test/scala/org/apache/spark/mllib/linalg/VectorsSuite.scala +++ b/mllib/src/test/scala/org/apache/spark/mllib/linalg/VectorsSuite.scala @@ -25,13 +25,12 @@ import breeze.linalg.{DenseMatrix => BDM} import org.json4s.jackson.JsonMethods.{parse => parseJson} import org.apache.spark.{SparkConf, SparkException, SparkFunSuite} -import org.apache.spark.internal.Logging import org.apache.spark.internal.config.Kryo._ import org.apache.spark.ml.{linalg => newlinalg} import org.apache.spark.mllib.util.TestingUtils._ import org.apache.spark.serializer.KryoSerializer -class VectorsSuite extends SparkFunSuite with Logging { +class VectorsSuite extends SparkFunSuite { val arr = Array(0.1, 0.0, 0.3, 0.4) val n = 4 diff --git a/mllib/src/test/scala/org/apache/spark/mllib/stat/CorrelationSuite.scala b/mllib/src/test/scala/org/apache/spark/mllib/stat/CorrelationSuite.scala index 4613f7fb6f4..eec54bec327 100644 --- a/mllib/src/test/scala/org/apache/spark/mllib/stat/CorrelationSuite.scala +++ b/mllib/src/test/scala/org/apache/spark/mllib/stat/CorrelationSuite.scala @@ -20,7 +20,6 @@ package org.apache.spark.mllib.stat import breeze.linalg.{DenseMatrix => BDM, Matrix => BM} import org.apache.spark.SparkFunSuite -import org.apache.spark.internal.Logging import org.apache.spark.mllib.linalg.Vectors import org.apache.spark.mllib.random.RandomRDDs import org.apache.spark.mllib.stat.correlation.{Correlations, PearsonCorrelation, @@ -28,7 +27,7 @@ import org.apache.spark.mllib.stat.correlation.{Correlations, PearsonCorrelation import org.apache.spark.mllib.util.MLlibTestSparkContext import org.apache.spark.mllib.util.TestingUtils._ -class CorrelationSuite extends SparkFunSuite with MLlibTestSparkContext with Logging { +class CorrelationSuite extends SparkFunSuite with MLlibTestSparkContext { // test input data val xData = Array(1.0, 0.0, -2.0) diff --git a/mllib/src/test/scala/org/apache/spark/mllib/tree/GradientBoostedTreesSuite.scala b/mllib/src/test/scala/org/apache/spark/mllib/tree/GradientBoostedTreesSuite.scala index c61f89322d3..206aef78884 100644 --- a/mllib/src/test/scala/org/apache/spark/mllib/tree/GradientBoostedTreesSuite.scala +++ b/mllib/src/test/scala/org/apache/spark/mllib/tree/GradientBoostedTreesSuite.scala @@ -18,7 +18,6 @@ package org.apache.spark.mllib.tree import org.apache.spark.SparkFunSuite -import org.apache.spark.internal.Logging import org.apache.spark.mllib.regression.LabeledPoint import org.apache.spark.mllib.tree.configuration.{BoostingStrategy, Strategy} import org.apache.spark.mllib.tree.configuration.Algo._ @@ -31,7 +30,7 @@ import org.apache.spark.util.Utils /** * Test suite for [[GradientBoostedTrees]]. */ -class GradientBoostedTreesSuite extends SparkFunSuite with MLlibTestSparkContext with Logging { +class GradientBoostedTreesSuite extends SparkFunSuite with MLlibTestSparkContext { test("Regression with continuous features: SquaredError") { GradientBoostedTreesSuite.testCombinations.foreach { diff --git a/repl/src/test/scala-2.12/org/apache/spark/repl/Repl2Suite.scala b/repl/src/test/scala-2.12/org/apache/spark/repl/Repl2Suite.scala index 90af9ec299e..15b45ad797e 100644 --- a/repl/src/test/scala-2.12/org/apache/spark/repl/Repl2Suite.scala +++ b/repl/src/test/scala-2.12/org/apache/spark/repl/Repl2Suite.scala @@ -21,11 +21,9 @@ import java.io._ import scala.tools.nsc.interpreter.SimpleReader -import org.scalatest.BeforeAndAfterAll - import org.apache.spark.{SparkContext, SparkFunSuite} -class Repl2Suite extends SparkFunSuite with BeforeAndAfterAll { +class Repl2Suite extends SparkFunSuite { test("propagation of local properties") { // A mock ILoop that doesn't install the SIGINT handler. class ILoop(out: PrintWriter) extends SparkILoop(None, out) { diff --git a/repl/src/test/scala-2.13/org/apache/spark/repl/Repl2Suite.scala b/repl/src/test/scala-2.13/org/apache/spark/repl/Repl2Suite.scala index ea57ca38ad7..d55ac91e466 100644 --- a/repl/src/test/scala-2.13/org/apache/spark/repl/Repl2Suite.scala +++ b/repl/src/test/scala-2.13/org/apache/spark/repl/Repl2Suite.scala @@ -19,11 +19,9 @@ package org.apache.spark.repl import java.io._ -import org.scalatest.BeforeAndAfterAll - import org.apache.spark.{SparkContext, SparkFunSuite} -class Repl2Suite extends SparkFunSuite with BeforeAndAfterAll { +class Repl2Suite extends SparkFunSuite { test("propagation of local properties") { // A mock ILoop that doesn't install the SIGINT handler. class ILoop(out: PrintWriter) extends SparkILoop(null, out) diff --git a/repl/src/test/scala/org/apache/spark/repl/ReplSuite.scala b/repl/src/test/scala/org/apache/spark/repl/ReplSuite.scala index 69e1273f5fa..1f44826c100 100644 --- a/repl/src/test/scala/org/apache/spark/repl/ReplSuite.scala +++ b/repl/src/test/scala/org/apache/spark/repl/ReplSuite.scala @@ -22,14 +22,13 @@ import java.nio.file.Files import org.apache.logging.log4j.{Level, LogManager} import org.apache.logging.log4j.core.{Logger, LoggerContext} -import org.scalatest.BeforeAndAfterAll import org.apache.spark.SparkFunSuite import org.apache.spark.internal.Logging import org.apache.spark.sql.SparkSession import org.apache.spark.sql.internal.StaticSQLConf.CATALOG_IMPLEMENTATION -class ReplSuite extends SparkFunSuite with BeforeAndAfterAll { +class ReplSuite extends SparkFunSuite { private var originalClassLoader: ClassLoader = null diff --git a/repl/src/test/scala/org/apache/spark/repl/SparkShellSuite.scala b/repl/src/test/scala/org/apache/spark/repl/SparkShellSuite.scala index ba815d72237..39544beec41 100644 --- a/repl/src/test/scala/org/apache/spark/repl/SparkShellSuite.scala +++ b/repl/src/test/scala/org/apache/spark/repl/SparkShellSuite.scala @@ -26,14 +26,11 @@ import scala.collection.mutable.ArrayBuffer import scala.concurrent.Promise import scala.concurrent.duration._ -import org.scalatest.BeforeAndAfterAll - import org.apache.spark.ProcessTestUtils.ProcessOutputCapturer import org.apache.spark.SparkFunSuite -import org.apache.spark.internal.Logging import org.apache.spark.util.ThreadUtils -class SparkShellSuite extends SparkFunSuite with BeforeAndAfterAll with Logging { +class SparkShellSuite extends SparkFunSuite { /** * Run a spark-shell operation and expect all the script and expected answers to be returned. * This method refers to [[runCliWithin()]] method in [[CliSuite]]. diff --git a/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/BaseYarnClusterSuite.scala b/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/BaseYarnClusterSuite.scala index d8867848278..ce935917456 100644 --- a/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/BaseYarnClusterSuite.scala +++ b/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/BaseYarnClusterSuite.scala @@ -29,19 +29,17 @@ import com.google.common.io.Files import org.apache.hadoop.yarn.conf.YarnConfiguration import org.apache.hadoop.yarn.server.MiniYARNCluster import org.scalactic.source.Position -import org.scalatest.{BeforeAndAfterAll, Tag} +import org.scalatest.Tag import org.scalatest.concurrent.Eventually._ import org.scalatest.matchers.must.Matchers import org.apache.spark._ import org.apache.spark.deploy.yarn.config._ -import org.apache.spark.internal.Logging import org.apache.spark.internal.config._ import org.apache.spark.launcher._ import org.apache.spark.util.Utils -abstract class BaseYarnClusterSuite - extends SparkFunSuite with BeforeAndAfterAll with Matchers with Logging { +abstract class BaseYarnClusterSuite extends SparkFunSuite with Matchers { private var isBindSuccessful = true // log4j configuration for the YARN containers, so that their output is collected diff --git a/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/ContainerPlacementStrategySuite.scala b/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/ContainerPlacementStrategySuite.scala index d64aad9ad47..0cabef77cf9 100644 --- a/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/ContainerPlacementStrategySuite.scala +++ b/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/ContainerPlacementStrategySuite.scala @@ -19,13 +19,12 @@ package org.apache.spark.deploy.yarn import org.apache.hadoop.yarn.api.records._ import org.apache.hadoop.yarn.client.api.AMRMClient.ContainerRequest -import org.scalatest.BeforeAndAfterEach import org.scalatest.matchers.must.Matchers import org.apache.spark.SparkFunSuite import org.apache.spark.resource.ResourceProfile -class ContainerPlacementStrategySuite extends SparkFunSuite with Matchers with BeforeAndAfterEach { +class ContainerPlacementStrategySuite extends SparkFunSuite with Matchers { private val yarnAllocatorSuite = new YarnAllocatorSuite import yarnAllocatorSuite._ diff --git a/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/YarnAllocatorSuite.scala b/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/YarnAllocatorSuite.scala index 2c7d7ac26c5..5a80aa9c610 100644 --- a/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/YarnAllocatorSuite.scala +++ b/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/YarnAllocatorSuite.scala @@ -33,7 +33,6 @@ import org.mockito.ArgumentCaptor import org.mockito.Mockito._ import org.mockito.invocation.InvocationOnMock import org.mockito.stubbing.Answer -import org.scalatest.BeforeAndAfterEach import org.scalatest.matchers.must.Matchers import org.scalatest.matchers.should.Matchers._ @@ -61,7 +60,7 @@ class MockResolver extends SparkRackResolver(SparkHadoopUtil.get.conf) { } -class YarnAllocatorSuite extends SparkFunSuite with Matchers with BeforeAndAfterEach { +class YarnAllocatorSuite extends SparkFunSuite with Matchers { val conf = new YarnConfiguration() val sparkConf = new SparkConf() sparkConf.set(DRIVER_HOST_ADDRESS, "localhost") diff --git a/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/YarnSparkHadoopUtilSuite.scala b/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/YarnSparkHadoopUtilSuite.scala index a36d1fa14d8..562ce3b5bdb 100644 --- a/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/YarnSparkHadoopUtilSuite.scala +++ b/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/YarnSparkHadoopUtilSuite.scala @@ -28,12 +28,10 @@ import org.scalatest.matchers.should.Matchers._ import org.apache.spark.{SecurityManager, SparkConf, SparkFunSuite} import org.apache.spark.deploy.SparkHadoopUtil -import org.apache.spark.internal.Logging import org.apache.spark.internal.config.UI._ import org.apache.spark.util.{ResetSystemProperties, Utils} -class YarnSparkHadoopUtilSuite extends SparkFunSuite with Matchers with Logging - with ResetSystemProperties { +class YarnSparkHadoopUtilSuite extends SparkFunSuite with Matchers with ResetSystemProperties { val hasBash = try { diff --git a/resource-managers/yarn/src/test/scala/org/apache/spark/network/yarn/YarnShuffleServiceSuite.scala b/resource-managers/yarn/src/test/scala/org/apache/spark/network/yarn/YarnShuffleServiceSuite.scala index 38d2247cc0d..dded1c77390 100644 --- a/resource-managers/yarn/src/test/scala/org/apache/spark/network/yarn/YarnShuffleServiceSuite.scala +++ b/resource-managers/yarn/src/test/scala/org/apache/spark/network/yarn/YarnShuffleServiceSuite.scala @@ -35,7 +35,6 @@ import org.apache.hadoop.yarn.api.records.ApplicationId import org.apache.hadoop.yarn.conf.YarnConfiguration import org.apache.hadoop.yarn.server.api.{ApplicationInitializationContext, ApplicationTerminationContext} import org.mockito.Mockito.{mock, when} -import org.scalatest.BeforeAndAfterEach import org.scalatest.concurrent.Eventually._ import org.scalatest.matchers.must.Matchers import org.scalatest.matchers.should.Matchers._ @@ -50,7 +49,7 @@ import org.apache.spark.tags.ExtendedLevelDBTest import org.apache.spark.util.Utils @ExtendedLevelDBTest -class YarnShuffleServiceSuite extends SparkFunSuite with Matchers with BeforeAndAfterEach { +class YarnShuffleServiceSuite extends SparkFunSuite with Matchers { private[yarn] var yarnConfig: YarnConfiguration = null private[yarn] val SORT_MANAGER = "org.apache.spark.shuffle.sort.SortShuffleManager" diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalogSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalogSuite.scala index 08ff128e512..a1218a00a39 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalogSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalogSuite.scala @@ -22,7 +22,6 @@ import java.util.TimeZone import org.apache.hadoop.conf.Configuration import org.apache.hadoop.fs.Path -import org.scalatest.BeforeAndAfterEach import org.apache.spark.SparkFunSuite import org.apache.spark.sql.AnalysisException @@ -36,13 +35,12 @@ import org.apache.spark.sql.connector.catalog.SupportsNamespaces.PROP_OWNER import org.apache.spark.sql.types._ import org.apache.spark.util.Utils - /** * A reasonable complete test suite (i.e. behaviors) for a [[ExternalCatalog]]. * * Implementations of the [[ExternalCatalog]] interface can create test suites by extending this. */ -abstract class ExternalCatalogSuite extends SparkFunSuite with BeforeAndAfterEach { +abstract class ExternalCatalogSuite extends SparkFunSuite { protected val utils: CatalogTestUtils import utils._ diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/aggregate/HistogramNumericSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/aggregate/HistogramNumericSuite.scala index 617226463f4..1cb3988b4c6 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/aggregate/HistogramNumericSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/aggregate/HistogramNumericSuite.scala @@ -21,7 +21,6 @@ import java.sql.Timestamp import java.time.{Duration, Period} import org.apache.spark.SparkFunSuite -import org.apache.spark.internal.Logging import org.apache.spark.sql.catalyst.InternalRow import org.apache.spark.sql.catalyst.analysis.TypeCheckResult.TypeCheckFailure import org.apache.spark.sql.catalyst.analysis.UnresolvedAttribute @@ -35,7 +34,7 @@ import org.apache.spark.sql.internal.SQLConf import org.apache.spark.sql.types._ import org.apache.spark.sql.util.NumericHistogram -class HistogramNumericSuite extends SparkFunSuite with SQLHelper with Logging { +class HistogramNumericSuite extends SparkFunSuite with SQLHelper { private val random = new java.util.Random() diff --git a/sql/core/src/test/scala/org/apache/spark/sql/SparkSessionBuilderSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/SparkSessionBuilderSuite.scala index ee8b6d28abc..7789d168792 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/SparkSessionBuilderSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/SparkSessionBuilderSuite.scala @@ -21,7 +21,6 @@ import scala.collection.JavaConverters._ import org.apache.hadoop.fs.Path import org.apache.logging.log4j.Level -import org.scalatest.BeforeAndAfterEach import org.scalatest.concurrent.Eventually import org.scalatest.time.SpanSugar._ @@ -36,7 +35,7 @@ import org.apache.spark.util.ThreadUtils /** * Test cases for the builder pattern of [[SparkSession]]. */ -class SparkSessionBuilderSuite extends SparkFunSuite with BeforeAndAfterEach with Eventually { +class SparkSessionBuilderSuite extends SparkFunSuite with Eventually { override def afterEach(): Unit = { // This suite should not interfere with the other test suites. diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/CoalesceShufflePartitionsSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/CoalesceShufflePartitionsSuite.scala index ddf4d421f3e..81777f67f37 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/CoalesceShufflePartitionsSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/CoalesceShufflePartitionsSuite.scala @@ -17,8 +17,6 @@ package org.apache.spark.sql.execution -import org.scalatest.BeforeAndAfterAll - import org.apache.spark.{SparkConf, SparkFunSuite} import org.apache.spark.internal.config.IO_ENCRYPTION_ENABLED import org.apache.spark.internal.config.UI.UI_ENABLED @@ -29,7 +27,7 @@ import org.apache.spark.sql.execution.exchange.ReusedExchangeExec import org.apache.spark.sql.functions._ import org.apache.spark.sql.internal.SQLConf -class CoalesceShufflePartitionsSuite extends SparkFunSuite with BeforeAndAfterAll { +class CoalesceShufflePartitionsSuite extends SparkFunSuite { private var originalActiveSparkSession: Option[SparkSession] = _ private var originalInstantiatedSparkSession: Option[SparkSession] = _ diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/ColumnTypeSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/ColumnTypeSuite.scala index e6dfc7c99ab..55fab5a7d66 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/ColumnTypeSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/ColumnTypeSuite.scala @@ -22,7 +22,6 @@ import java.nio.charset.StandardCharsets import java.time.{Duration, Period} import org.apache.spark.SparkFunSuite -import org.apache.spark.internal.Logging import org.apache.spark.sql.Row import org.apache.spark.sql.catalyst.CatalystTypeConverters import org.apache.spark.sql.catalyst.expressions.{GenericInternalRow, UnsafeProjection} @@ -30,7 +29,7 @@ import org.apache.spark.sql.execution.columnar.ColumnarTestUtils._ import org.apache.spark.sql.types._ import org.apache.spark.unsafe.types.CalendarInterval -class ColumnTypeSuite extends SparkFunSuite with Logging { +class ColumnTypeSuite extends SparkFunSuite { private val DEFAULT_BUFFER_SIZE = 512 private val MAP_TYPE = MAP(MapType(IntegerType, StringType)) private val ARRAY_TYPE = ARRAY(ArrayType(IntegerType)) diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/state/StateStoreRDDSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/state/StateStoreRDDSuite.scala index 6bb8ebe5149..135bd324cc9 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/state/StateStoreRDDSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/state/StateStoreRDDSuite.scala @@ -23,7 +23,7 @@ import java.util.UUID import scala.util.Random -import org.scalatest.{BeforeAndAfter, BeforeAndAfterAll} +import org.scalatest.BeforeAndAfter import org.apache.spark.{SparkConf, SparkContext, SparkFunSuite} import org.apache.spark.rdd.RDD @@ -34,7 +34,7 @@ import org.apache.spark.sql.catalyst.util.quietly import org.apache.spark.sql.execution.streaming.StatefulOperatorStateInfo import org.apache.spark.util.{CompletionIterator, Utils} -class StateStoreRDDSuite extends SparkFunSuite with BeforeAndAfter with BeforeAndAfterAll { +class StateStoreRDDSuite extends SparkFunSuite with BeforeAndAfter { import StateStoreTestsHelper._ diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/vectorized/ColumnVectorSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/vectorized/ColumnVectorSuite.scala index cdf41ed651d..910e3e682de 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/vectorized/ColumnVectorSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/vectorized/ColumnVectorSuite.scala @@ -17,8 +17,6 @@ package org.apache.spark.sql.execution.vectorized -import org.scalatest.BeforeAndAfterEach - import org.apache.spark.SparkFunSuite import org.apache.spark.sql.catalyst.expressions.SpecificInternalRow import org.apache.spark.sql.execution.columnar.ColumnAccessor @@ -27,7 +25,7 @@ import org.apache.spark.sql.types._ import org.apache.spark.sql.vectorized.ColumnarArray import org.apache.spark.unsafe.types.UTF8String -class ColumnVectorSuite extends SparkFunSuite with BeforeAndAfterEach { +class ColumnVectorSuite extends SparkFunSuite { private def withVector( vector: WritableColumnVector)( block: WritableColumnVector => Unit): Unit = { diff --git a/sql/core/src/test/scala/org/apache/spark/sql/sources/DataSourceAnalysisSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/sources/DataSourceAnalysisSuite.scala index 152d096afdf..ba0dfa95178 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/sources/DataSourceAnalysisSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/sources/DataSourceAnalysisSuite.scala @@ -17,8 +17,6 @@ package org.apache.spark.sql.sources -import org.scalatest.BeforeAndAfterAll - import org.apache.spark.SparkFunSuite import org.apache.spark.sql.AnalysisException import org.apache.spark.sql.catalyst.analysis.SimpleAnalyzer @@ -30,7 +28,7 @@ import org.apache.spark.sql.internal.SQLConf import org.apache.spark.sql.internal.SQLConf.StoreAssignmentPolicy import org.apache.spark.sql.types.{DataType, IntegerType, StructType} -class DataSourceAnalysisSuite extends SparkFunSuite with BeforeAndAfterAll with SQLHelper { +class DataSourceAnalysisSuite extends SparkFunSuite with SQLHelper { private var targetAttributes: Seq[Attribute] = _ private var targetPartitionSchema: StructType = _ diff --git a/sql/core/src/test/scala/org/apache/spark/sql/streaming/ui/UISeleniumSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/streaming/ui/UISeleniumSuite.scala index db3d6529c99..83f4ff15eb4 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/streaming/ui/UISeleniumSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/streaming/ui/UISeleniumSuite.scala @@ -19,7 +19,6 @@ package org.apache.spark.sql.streaming.ui import org.openqa.selenium.WebDriver import org.openqa.selenium.htmlunit.HtmlUnitDriver -import org.scalatest.BeforeAndAfterAll import org.scalatest.concurrent.Eventually._ import org.scalatest.matchers.must.Matchers import org.scalatest.matchers.should.Matchers._ @@ -37,7 +36,7 @@ import org.apache.spark.sql.internal.StaticSQLConf.ENABLED_STREAMING_UI_CUSTOM_M import org.apache.spark.sql.streaming.{StreamingQueryException, Trigger} import org.apache.spark.ui.SparkUICssErrorHandler -class UISeleniumSuite extends SparkFunSuite with WebBrowser with Matchers with BeforeAndAfterAll { +class UISeleniumSuite extends SparkFunSuite with WebBrowser with Matchers { implicit var webDriver: WebDriver = _ diff --git a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/CliSuite.scala b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/CliSuite.scala index f6e8b49d555..e64492c3128 100644 --- a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/CliSuite.scala +++ b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/CliSuite.scala @@ -31,12 +31,10 @@ import scala.concurrent.duration._ import org.apache.hadoop.hive.cli.CliSessionState import org.apache.hadoop.hive.conf.HiveConf.ConfVars import org.apache.hadoop.hive.ql.session.SessionState -import org.scalatest.BeforeAndAfterAll import org.apache.spark.{SparkConf, SparkContext, SparkFunSuite} import org.apache.spark.ProcessTestUtils.ProcessOutputCapturer import org.apache.spark.deploy.SparkHadoopUtil -import org.apache.spark.internal.Logging import org.apache.spark.sql.hive.HiveUtils import org.apache.spark.sql.hive.HiveUtils._ import org.apache.spark.sql.hive.client.HiveClientImpl @@ -47,7 +45,7 @@ import org.apache.spark.util.{ThreadUtils, Utils} /** * A test suite for the `spark-sql` CLI tool. */ -class CliSuite extends SparkFunSuite with BeforeAndAfterAll with Logging { +class CliSuite extends SparkFunSuite { val warehousePath = Utils.createTempDir() val metastorePath = Utils.createTempDir() val scratchDirPath = Utils.createTempDir() diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/HiveContextCompatibilitySuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/HiveContextCompatibilitySuite.scala index a80db765846..935ee90731a 100644 --- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/HiveContextCompatibilitySuite.scala +++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/HiveContextCompatibilitySuite.scala @@ -17,12 +17,9 @@ package org.apache.spark.sql.hive -import org.scalatest.BeforeAndAfterEach - import org.apache.spark.{SparkConf, SparkContext, SparkFunSuite} - -class HiveContextCompatibilitySuite extends SparkFunSuite with BeforeAndAfterEach { +class HiveContextCompatibilitySuite extends SparkFunSuite { override protected val enableAutoThreadAudit = false private var sc: SparkContext = null diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/client/FiltersSuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/client/FiltersSuite.scala index 29b51e1d17b..e610218e1d7 100644 --- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/client/FiltersSuite.scala +++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/client/FiltersSuite.scala @@ -24,7 +24,6 @@ import org.apache.hadoop.hive.metastore.api.FieldSchema import org.apache.hadoop.hive.serde.serdeConstants import org.apache.spark.SparkFunSuite -import org.apache.spark.internal.Logging import org.apache.spark.sql.catalyst.dsl.expressions._ import org.apache.spark.sql.catalyst.expressions._ import org.apache.spark.sql.catalyst.plans.PlanTest @@ -36,7 +35,7 @@ import org.apache.spark.unsafe.types.UTF8String * A set of tests for the filter conversion logic used when pushing partition pruning into the * metastore */ -class FiltersSuite extends SparkFunSuite with Logging with PlanTest { +class FiltersSuite extends SparkFunSuite with PlanTest { private val shim = new Shim_v0_13 private val testTable = new org.apache.hadoop.hive.ql.metadata.Table("default", "test") diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/ConcurrentHiveSuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/ConcurrentHiveSuite.scala index 76134d23d18..b63fe256a93 100644 --- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/ConcurrentHiveSuite.scala +++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/ConcurrentHiveSuite.scala @@ -17,13 +17,11 @@ package org.apache.spark.sql.hive.execution -import org.scalatest.BeforeAndAfterAll - import org.apache.spark.{SparkConf, SparkContext, SparkFunSuite} import org.apache.spark.internal.config.UI.UI_ENABLED import org.apache.spark.sql.hive.test.TestHiveContext -class ConcurrentHiveSuite extends SparkFunSuite with BeforeAndAfterAll { +class ConcurrentHiveSuite extends SparkFunSuite { ignore("multiple instances not supported") { test("Multiple Hive Instances") { (1 to 10).map { i => diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/Hive_2_1_DDLSuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/Hive_2_1_DDLSuite.scala index 6f37e39a532..cdef3c01087 100644 --- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/Hive_2_1_DDLSuite.scala +++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/Hive_2_1_DDLSuite.scala @@ -18,7 +18,6 @@ package org.apache.spark.sql.hive.execution import org.apache.hadoop.conf.Configuration -import org.scalatest.{BeforeAndAfterAll, BeforeAndAfterEach} import org.apache.spark.{SparkConf, SparkFunSuite} import org.apache.spark.launcher.SparkLauncher @@ -36,8 +35,7 @@ import org.apache.spark.util.Utils */ @SlowHiveTest @ExtendedHiveTest -class Hive_2_1_DDLSuite extends SparkFunSuite with TestHiveSingleton with BeforeAndAfterEach - with BeforeAndAfterAll { +class Hive_2_1_DDLSuite extends SparkFunSuite with TestHiveSingleton { // Create a custom HiveExternalCatalog instance with the desired configuration. We cannot // use SparkSession here since there's already an active on managed by the TestHive object. diff --git a/streaming/src/test/scala/org/apache/spark/streaming/DStreamClosureSuite.scala b/streaming/src/test/scala/org/apache/spark/streaming/DStreamClosureSuite.scala index dad324b53dd..7fcc67b4614 100644 --- a/streaming/src/test/scala/org/apache/spark/streaming/DStreamClosureSuite.scala +++ b/streaming/src/test/scala/org/apache/spark/streaming/DStreamClosureSuite.scala @@ -19,8 +19,6 @@ package org.apache.spark.streaming import java.io.NotSerializableException -import org.scalatest.BeforeAndAfterAll - import org.apache.spark.{HashPartitioner, SparkContext, SparkException, SparkFunSuite} import org.apache.spark.rdd.RDD import org.apache.spark.streaming.dstream.DStream @@ -29,7 +27,7 @@ import org.apache.spark.util.ReturnStatementInClosureException /** * Test that closures passed to DStream operations are actually cleaned. */ -class DStreamClosureSuite extends SparkFunSuite with LocalStreamingContext with BeforeAndAfterAll { +class DStreamClosureSuite extends SparkFunSuite with LocalStreamingContext { override protected def beforeEach(): Unit = { super.beforeEach() diff --git a/streaming/src/test/scala/org/apache/spark/streaming/FailureSuite.scala b/streaming/src/test/scala/org/apache/spark/streaming/FailureSuite.scala index 19ceb748e07..895b9576b6a 100644 --- a/streaming/src/test/scala/org/apache/spark/streaming/FailureSuite.scala +++ b/streaming/src/test/scala/org/apache/spark/streaming/FailureSuite.scala @@ -22,14 +22,13 @@ import java.io.File import org.scalatest.BeforeAndAfter import org.apache.spark._ -import org.apache.spark.internal.Logging import org.apache.spark.util.Utils /** * This testsuite tests master failures at random times while the stream is running using * the real clock. */ -class FailureSuite extends SparkFunSuite with BeforeAndAfter with Logging { +class FailureSuite extends SparkFunSuite with BeforeAndAfter { private val batchDuration: Duration = Milliseconds(1000) private val numBatches = 30 diff --git a/streaming/src/test/scala/org/apache/spark/streaming/ReceivedBlockTrackerSuite.scala b/streaming/src/test/scala/org/apache/spark/streaming/ReceivedBlockTrackerSuite.scala index 241ad1b8675..7bdb8bd9a02 100644 --- a/streaming/src/test/scala/org/apache/spark/streaming/ReceivedBlockTrackerSuite.scala +++ b/streaming/src/test/scala/org/apache/spark/streaming/ReceivedBlockTrackerSuite.scala @@ -34,7 +34,6 @@ import org.scalatest.matchers.must.Matchers import org.scalatest.matchers.should.Matchers._ import org.apache.spark.{SparkConf, SparkFunSuite} -import org.apache.spark.internal.Logging import org.apache.spark.storage.StreamBlockId import org.apache.spark.streaming.receiver.BlockManagerBasedStoreResult import org.apache.spark.streaming.scheduler.{AllocatedBlocks, _} @@ -42,8 +41,7 @@ import org.apache.spark.streaming.util._ import org.apache.spark.streaming.util.WriteAheadLogSuite._ import org.apache.spark.util.{Clock, ManualClock, SystemClock, Utils} -class ReceivedBlockTrackerSuite - extends SparkFunSuite with BeforeAndAfter with Matchers with Logging { +class ReceivedBlockTrackerSuite extends SparkFunSuite with BeforeAndAfter with Matchers { val hadoopConf = new Configuration() val streamId = 1 diff --git a/streaming/src/test/scala/org/apache/spark/streaming/UISeleniumSuite.scala b/streaming/src/test/scala/org/apache/spark/streaming/UISeleniumSuite.scala index f7971019925..979cfdeefaf 100644 --- a/streaming/src/test/scala/org/apache/spark/streaming/UISeleniumSuite.scala +++ b/streaming/src/test/scala/org/apache/spark/streaming/UISeleniumSuite.scala @@ -21,7 +21,6 @@ import scala.collection.mutable.Queue import org.openqa.selenium.WebDriver import org.openqa.selenium.htmlunit.HtmlUnitDriver -import org.scalatest.BeforeAndAfterAll import org.scalatest.concurrent.Eventually._ import org.scalatest.matchers.must.Matchers import org.scalatest.matchers.should.Matchers._ @@ -35,8 +34,7 @@ import org.apache.spark.ui.SparkUICssErrorHandler /** * Selenium tests for the Spark Streaming Web UI. */ -class UISeleniumSuite - extends SparkFunSuite with WebBrowser with Matchers with BeforeAndAfterAll with TestSuiteBase { +class UISeleniumSuite extends SparkFunSuite with WebBrowser with Matchers with TestSuiteBase { implicit var webDriver: WebDriver = _ diff --git a/streaming/src/test/scala/org/apache/spark/streaming/rdd/MapWithStateRDDSuite.scala b/streaming/src/test/scala/org/apache/spark/streaming/rdd/MapWithStateRDDSuite.scala index f06b1feb8c0..1b59a8c6b8b 100644 --- a/streaming/src/test/scala/org/apache/spark/streaming/rdd/MapWithStateRDDSuite.scala +++ b/streaming/src/test/scala/org/apache/spark/streaming/rdd/MapWithStateRDDSuite.scala @@ -22,15 +22,13 @@ import java.io.File import scala.collection.mutable.ArrayBuffer import scala.reflect.ClassTag -import org.scalatest.BeforeAndAfterAll - import org.apache.spark._ import org.apache.spark.rdd.RDD import org.apache.spark.streaming.{State, Time} import org.apache.spark.streaming.util.OpenHashMapBasedStateMap import org.apache.spark.util.Utils -class MapWithStateRDDSuite extends SparkFunSuite with RDDCheckpointTester with BeforeAndAfterAll { +class MapWithStateRDDSuite extends SparkFunSuite with RDDCheckpointTester { private var sc: SparkContext = null private var checkpointDir: File = _ diff --git a/streaming/src/test/scala/org/apache/spark/streaming/rdd/WriteAheadLogBackedBlockRDDSuite.scala b/streaming/src/test/scala/org/apache/spark/streaming/rdd/WriteAheadLogBackedBlockRDDSuite.scala index 86a8dc47098..ad35841caa9 100644 --- a/streaming/src/test/scala/org/apache/spark/streaming/rdd/WriteAheadLogBackedBlockRDDSuite.scala +++ b/streaming/src/test/scala/org/apache/spark/streaming/rdd/WriteAheadLogBackedBlockRDDSuite.scala @@ -21,7 +21,6 @@ import java.io.File import scala.util.Random import org.apache.hadoop.conf.Configuration -import org.scalatest.{BeforeAndAfterAll, BeforeAndAfterEach} import org.apache.spark.{SparkConf, SparkContext, SparkException, SparkFunSuite} import org.apache.spark.internal.config._ @@ -30,8 +29,7 @@ import org.apache.spark.storage.{BlockId, BlockManager, StorageLevel, StreamBloc import org.apache.spark.streaming.util.{FileBasedWriteAheadLogSegment, FileBasedWriteAheadLogWriter} import org.apache.spark.util.Utils -class WriteAheadLogBackedBlockRDDSuite - extends SparkFunSuite with BeforeAndAfterAll with BeforeAndAfterEach { +class WriteAheadLogBackedBlockRDDSuite extends SparkFunSuite { val conf = new SparkConf() .setMaster("local[2]") --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org