This is an automated email from the ASF dual-hosted git repository.
dongjoon pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push:
new 7efeab3696f [SPARK-39754][CORE][SQL] Remove unused `import` or
unnecessary `{}`
7efeab3696f is described below
commit 7efeab3696f3ffe341b4499d6ecac920f0994b3c
Author: panbingkun <[email protected]>
AuthorDate: Tue Jul 12 09:12:56 2022 -0700
[SPARK-39754][CORE][SQL] Remove unused `import` or unnecessary `{}`
### What changes were proposed in this pull request?
Review code and found some issue about import in java & scala basecode,
mainly focus on:
> 1.unnecessary braces in single import
> 2.unused import
### Why are the changes needed?
Eliminate redundant information and make the code cleaner
### Does this PR introduce _any_ user-facing change?
No.
### How was this patch tested?
Pass GA.
Closes #37166 from panbingkun/remove_unused_import.
Authored-by: panbingkun <[email protected]>
Signed-off-by: Dongjoon Hyun <[email protected]>
---
.../src/main/java/org/apache/spark/network/shuffle/ErrorHandler.java | 2 --
.../org/apache/spark/network/shuffle/ExternalBlockHandlerSuite.java | 1 -
.../main/scala/org/apache/spark/deploy/history/ApplicationCache.scala | 1 -
core/src/main/scala/org/apache/spark/resource/ResourceUtils.scala | 2 +-
launcher/src/main/java/org/apache/spark/launcher/SparkLauncher.java | 1 -
.../java/org/apache/spark/launcher/SparkSubmitOptionParserSuite.java | 1 -
.../scala/org/apache/spark/sql/catalyst/analysis/ResolveUnion.scala | 2 +-
.../apache/spark/sql/execution/command/AnalyzePartitionCommand.scala | 2 +-
.../scala/org/apache/spark/sql/execution/command/CommandUtils.scala | 2 +-
.../spark/sql/execution/datasources/v2/DataSourceV2Strategy.scala | 2 +-
10 files changed, 5 insertions(+), 11 deletions(-)
diff --git
a/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ErrorHandler.java
b/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ErrorHandler.java
index 519b02d1242..abd2348cdaf 100644
---
a/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ErrorHandler.java
+++
b/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ErrorHandler.java
@@ -25,8 +25,6 @@ import com.google.common.base.Throwables;
import org.apache.spark.annotation.Evolving;
import org.apache.spark.network.server.BlockPushNonFatalFailure;
-import static
org.apache.spark.network.server.BlockPushNonFatalFailure.ReturnCode.*;
-
/**
* Plugs into {@link RetryingBlockTransferor} to further control when an
exception should be retried
* and logged.
diff --git
a/common/network-shuffle/src/test/java/org/apache/spark/network/shuffle/ExternalBlockHandlerSuite.java
b/common/network-shuffle/src/test/java/org/apache/spark/network/shuffle/ExternalBlockHandlerSuite.java
index f681af71e40..54f29fedf83 100644
---
a/common/network-shuffle/src/test/java/org/apache/spark/network/shuffle/ExternalBlockHandlerSuite.java
+++
b/common/network-shuffle/src/test/java/org/apache/spark/network/shuffle/ExternalBlockHandlerSuite.java
@@ -34,7 +34,6 @@ import org.mockito.ArgumentCaptor;
import org.roaringbitmap.RoaringBitmap;
import static org.junit.Assert.*;
-import static org.mockito.ArgumentMatchers.any;
import static org.mockito.Mockito.*;
import org.apache.spark.network.buffer.ManagedBuffer;
diff --git
a/core/src/main/scala/org/apache/spark/deploy/history/ApplicationCache.scala
b/core/src/main/scala/org/apache/spark/deploy/history/ApplicationCache.scala
index 89b30a35ebe..829631a0454 100644
--- a/core/src/main/scala/org/apache/spark/deploy/history/ApplicationCache.scala
+++ b/core/src/main/scala/org/apache/spark/deploy/history/ApplicationCache.scala
@@ -17,7 +17,6 @@
package org.apache.spark.deploy.history
-import java.util.NoSuchElementException
import java.util.concurrent.ExecutionException
import javax.servlet.{DispatcherType, Filter, FilterChain, FilterConfig,
ServletException, ServletRequest, ServletResponse}
import javax.servlet.http.{HttpServletRequest, HttpServletResponse}
diff --git a/core/src/main/scala/org/apache/spark/resource/ResourceUtils.scala
b/core/src/main/scala/org/apache/spark/resource/ResourceUtils.scala
index 5b5a51fc760..58b37269be4 100644
--- a/core/src/main/scala/org/apache/spark/resource/ResourceUtils.scala
+++ b/core/src/main/scala/org/apache/spark/resource/ResourceUtils.scala
@@ -30,7 +30,7 @@ import org.apache.spark.annotation.DeveloperApi
import org.apache.spark.api.resource.ResourceDiscoveryPlugin
import org.apache.spark.internal.Logging
import org.apache.spark.internal.config.{EXECUTOR_CORES,
RESOURCES_DISCOVERY_PLUGIN, SPARK_TASK_PREFIX}
-import org.apache.spark.internal.config.Tests.{RESOURCES_WARNING_TESTING}
+import org.apache.spark.internal.config.Tests.RESOURCES_WARNING_TESTING
import org.apache.spark.util.Utils
/**
diff --git
a/launcher/src/main/java/org/apache/spark/launcher/SparkLauncher.java
b/launcher/src/main/java/org/apache/spark/launcher/SparkLauncher.java
index 12febc5441b..b2541768856 100644
--- a/launcher/src/main/java/org/apache/spark/launcher/SparkLauncher.java
+++ b/launcher/src/main/java/org/apache/spark/launcher/SparkLauncher.java
@@ -30,7 +30,6 @@ import java.util.logging.Level;
import java.util.logging.Logger;
import static org.apache.spark.launcher.CommandBuilderUtils.*;
-import static org.apache.spark.launcher.CommandBuilderUtils.join;
/**
* Launcher for Spark applications.
diff --git
a/launcher/src/test/java/org/apache/spark/launcher/SparkSubmitOptionParserSuite.java
b/launcher/src/test/java/org/apache/spark/launcher/SparkSubmitOptionParserSuite.java
index d7035dfcc56..9d566e85d9f 100644
---
a/launcher/src/test/java/org/apache/spark/launcher/SparkSubmitOptionParserSuite.java
+++
b/launcher/src/test/java/org/apache/spark/launcher/SparkSubmitOptionParserSuite.java
@@ -24,7 +24,6 @@ import java.util.List;
import org.junit.Before;
import org.junit.Test;
import static org.junit.Assert.assertThrows;
-import static org.mockito.ArgumentMatchers.isNull;
import static org.mockito.Mockito.*;
public class SparkSubmitOptionParserSuite extends BaseSuite {
diff --git
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveUnion.scala
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveUnion.scala
index fff38bbcd03..6209e076573 100644
---
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveUnion.scala
+++
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveUnion.scala
@@ -20,7 +20,7 @@ package org.apache.spark.sql.catalyst.analysis
import scala.collection.mutable
import org.apache.spark.sql.catalyst.expressions._
-import org.apache.spark.sql.catalyst.optimizer.{CombineUnions}
+import org.apache.spark.sql.catalyst.optimizer.CombineUnions
import org.apache.spark.sql.catalyst.plans.logical.{LogicalPlan, Project,
Union}
import org.apache.spark.sql.catalyst.rules.Rule
import org.apache.spark.sql.catalyst.trees.TreePattern.UNION
diff --git
a/sql/core/src/main/scala/org/apache/spark/sql/execution/command/AnalyzePartitionCommand.scala
b/sql/core/src/main/scala/org/apache/spark/sql/execution/command/AnalyzePartitionCommand.scala
index 38d92ba752a..433f9fcdd08 100644
---
a/sql/core/src/main/scala/org/apache/spark/sql/execution/command/AnalyzePartitionCommand.scala
+++
b/sql/core/src/main/scala/org/apache/spark/sql/execution/command/AnalyzePartitionCommand.scala
@@ -19,7 +19,7 @@ package org.apache.spark.sql.execution.command
import org.apache.spark.sql.{Column, Row, SparkSession}
import org.apache.spark.sql.catalyst.TableIdentifier
-import org.apache.spark.sql.catalyst.analysis.{UnresolvedAttribute}
+import org.apache.spark.sql.catalyst.analysis.UnresolvedAttribute
import org.apache.spark.sql.catalyst.catalog.{CatalogTable, CatalogTableType,
ExternalCatalogUtils}
import org.apache.spark.sql.catalyst.catalog.CatalogTypes.TablePartitionSpec
import org.apache.spark.sql.catalyst.expressions.{And, EqualTo, Literal}
diff --git
a/sql/core/src/main/scala/org/apache/spark/sql/execution/command/CommandUtils.scala
b/sql/core/src/main/scala/org/apache/spark/sql/execution/command/CommandUtils.scala
index 2154a5893dd..41f60bfa2ff 100644
---
a/sql/core/src/main/scala/org/apache/spark/sql/execution/command/CommandUtils.scala
+++
b/sql/core/src/main/scala/org/apache/spark/sql/execution/command/CommandUtils.scala
@@ -25,7 +25,7 @@ import scala.util.control.NonFatal
import org.apache.hadoop.fs.{FileSystem, Path, PathFilter}
import org.apache.spark.internal.Logging
-import org.apache.spark.sql.{SparkSession}
+import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.catalyst.{InternalRow, TableIdentifier}
import org.apache.spark.sql.catalyst.catalog.{CatalogStatistics, CatalogTable,
CatalogTablePartition, CatalogTableType}
import org.apache.spark.sql.catalyst.expressions._
diff --git
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DataSourceV2Strategy.scala
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DataSourceV2Strategy.scala
index b678effbea9..93445eea51b 100644
---
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DataSourceV2Strategy.scala
+++
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DataSourceV2Strategy.scala
@@ -31,7 +31,7 @@ import org.apache.spark.sql.catalyst.plans.logical._
import org.apache.spark.sql.catalyst.util.{toPrettySQL, ResolveDefaultColumns,
V2ExpressionBuilder}
import org.apache.spark.sql.connector.catalog.{Identifier,
StagingTableCatalog, SupportsDelete, SupportsNamespaces,
SupportsPartitionManagement, SupportsWrite, Table, TableCapability,
TableCatalog, TruncatableTable}
import org.apache.spark.sql.connector.catalog.index.SupportsIndex
-import org.apache.spark.sql.connector.expressions.{FieldReference}
+import org.apache.spark.sql.connector.expressions.FieldReference
import org.apache.spark.sql.connector.expressions.filter.{And => V2And, Not =>
V2Not, Or => V2Or, Predicate}
import org.apache.spark.sql.connector.read.LocalScan
import org.apache.spark.sql.connector.read.streaming.{ContinuousStream,
MicroBatchStream}
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]