This is an automated email from the ASF dual-hosted git repository.
benjobs pushed a commit to branch dev
in repository https://gitbox.apache.org/repos/asf/incubator-streampark.git
The following commit(s) were added to refs/heads/dev by this push:
new 9998ca604 [Improve] scala implicits some improvements (#3902)
9998ca604 is described below
commit 9998ca6042eea0dc3af178bf7794a6f76b2880e4
Author: benjobs <[email protected]>
AuthorDate: Sun Jul 21 00:43:15 2024 +0800
[Improve] scala implicits some improvements (#3902)
* [Improve] scala implicits some improvements
* [Improve] check-LICENSE.sh minor improvement
* [Improve] FlinkVersion and SparkVersion improvements
---
.../streampark/common/conf/ConfigOption.scala | 2 +-
.../streampark/common/conf/FlinkVersion.scala | 8 +--
.../common/conf/InternalConfigHolder.scala | 6 +--
.../streampark/common/conf/SparkVersion.scala | 60 +++++++++-------------
.../apache/streampark/common/conf/Workspace.scala | 2 +-
.../streampark/common/util/AssertUtils.scala | 3 +-
.../streampark/common/util/CURLBuilder.scala | 4 +-
.../streampark/common/util/CommandUtils.scala | 3 +-
.../streampark/common/util/ConfigUtils.scala | 4 +-
.../apache/streampark/common/util/FileUtils.scala | 5 +-
.../streampark/common/util/HBaseClient.scala | 3 +-
.../streampark/common/util/HadoopConfigUtils.scala | 5 +-
.../streampark/common/util/HadoopUtils.scala | 2 +-
.../apache/streampark/common/util/HostsUtils.scala | 4 +-
.../streampark/common/util/HttpClientUtils.scala | 5 +-
.../util/{ImplicitsUtils.scala => Implicits.scala} | 42 ++++++++++++---
.../apache/streampark/common/util/JdbcUtils.scala | 2 +-
.../streampark/common/util/MongoConfig.scala | 3 +-
.../streampark/common/util/PropertiesUtils.scala | 33 ++++++------
.../streampark/common/util/RedisClient.scala | 2 +-
.../streampark/common/util/RedisEndpoint.scala | 3 +-
.../apache/streampark/common/util/RedisUtils.scala | 4 +-
.../streampark/common/util/SqlConvertUtils.scala | 3 +-
.../org/apache/streampark/common/util/Utils.scala | 6 +--
.../apache/streampark/common/util/YarnUtils.scala | 3 +-
.../streampark/common/util/ZooKeeperUtils.scala | 3 +-
...mplicitsUtilsTest.scala => ImplicitsTest.scala} | 4 +-
.../console/core/runner/QuickStartRunner.java | 4 +-
.../flink/client/bean/CancelRequest.scala | 3 +-
.../flink/client/bean/DeployRequest.scala | 2 +-
.../flink/client/bean/DeployRequestTrait.scala | 3 +-
.../flink/client/bean/SavepointRequestTrait.scala | 3 +-
.../flink/client/bean/ShutDownRequest.scala | 3 +-
.../flink/client/bean/SubmitRequest.scala | 3 +-
.../flink/client/bean/SubmitResponse.scala | 4 +-
.../client/bean/TriggerSavepointRequest.scala | 3 +-
.../impl/KubernetesNativeApplicationClient.scala | 2 -
.../impl/KubernetesNativeSessionClient.scala | 4 +-
.../flink/client/impl/YarnApplicationClient.scala | 2 +-
.../flink/client/impl/YarnPerJobClient.scala | 3 +-
.../flink/client/impl/YarnSessionClient.scala | 3 +-
.../flink/client/trait/FlinkClientTrait.scala | 4 +-
.../flink/client/trait/YarnClientTrait.scala | 2 +-
.../flink/client/test/YarnPerJobTestCase.scala | 4 +-
.../flink/connector/conf/ThresholdConf.scala | 3 +-
.../flink/connector/failover/FailoverChecker.scala | 2 +-
.../flink/connector/failover/FailoverWriter.scala | 2 +-
.../flink/connector/failover/SinkBuffer.scala | 2 +-
.../flink/connector/failover/SinkRequestTest.scala | 2 +-
.../clickhouse/conf/ClickHouseHttpConfig.scala | 2 +-
.../conf/ClickHouseSinkConfigOption.scala | 2 +-
.../internal/ClickHouseSinkFunction.scala | 2 +-
.../clickhouse/internal/ClickHouseSinkWriter.scala | 2 +-
.../clickhouse/internal/ClickHouseWriterTask.scala | 2 +-
.../connector/doris/conf/DorisConfig.scala | 2 +-
.../doris/conf/DorisSinkConfigOption.scala | 2 +-
.../elasticsearch5/conf/ESSinkConfigOption.scala | 5 +-
.../connector/elasticsearch5/sink/ES5Sink.scala | 4 +-
.../elasticsearch6/conf/ES6SinkConfigOption.scala | 4 +-
.../connector/elasticsearch6/sink/ES6Sink.scala | 2 +-
.../elasticsearch7/conf/ES7SinkConfigOption.scala | 4 +-
.../connector/elasticsearch7/sink/ES7Sink.scala | 2 +-
.../hbase/internal/HBaseSinkFunction.scala | 2 +-
.../hbase/internal/HBaseSourceFunction.scala | 2 +-
.../connector/hbase/request/HBaseRequest.scala | 2 +-
.../connector/http/internal/HttpSinkWriter.scala | 2 +-
.../connector/http/internal/HttpWriterTask.scala | 2 +-
.../flink/connector/http/sink/HttpSink.scala | 2 +-
.../connector/influx/function/InfluxFunction.scala | 4 +-
.../jdbc/internal/JdbcSourceFunction.scala | 2 -
.../flink/connector/kafka/source/KafkaSource.scala | 2 +-
.../mongo/internal/MongoSourceFunction.scala | 2 +-
.../redis/conf/RedisSinkConfigOption.scala | 2 +-
.../flink/connector/redis/sink/RedisSink.scala | 2 +-
.../flink/kubernetes/FlinkK8sWatchController.scala | 3 +-
.../flink/kubernetes/KubernetesRetriever.scala | 3 +-
.../flink/kubernetes/PodTemplateParser.scala | 4 +-
.../helper/KubernetesDeploymentHelper.scala | 3 +-
.../kubernetes/ingress/IngressController.scala | 2 +-
.../kubernetes/ingress/IngressStrategyV1.scala | 3 +-
.../ingress/IngressStrategyV1beta1.scala | 3 +-
.../kubernetes/watcher/FlinkJobStatusWatcher.scala | 2 +-
.../flink/kubernetes/FlinkRestJsonTest.scala | 2 +-
.../flink/kubernetes/PodTemplateParserTest.scala | 4 +-
.../flink/packer/maven/DependencyInfo.scala | 4 +-
.../streampark/flink/packer/maven/MavenTool.scala | 3 +-
.../flink/packer/pipeline/BuildPipeline.scala | 2 +-
.../packer/pipeline/DockerResolveSnapshot.scala | 5 +-
.../flink/packer/pipeline/PipelineSnapshot.scala | 4 +-
.../pipeline/impl/FlinkRemoteBuildPipeline.scala | 3 +-
.../impl/FlinkYarnApplicationBuildPipeline.scala | 4 +-
.../impl/SparkYarnApplicationBuildPipeline.scala | 4 +-
.../streampark/flink/packer/MavenToolSpec.scala | 3 +-
.../streampark/flink/proxy/FlinkShimsProxy.scala | 2 +-
.../flink/core/FlinkStreamTableTrait.scala | 3 +-
.../flink/core/FlinkStreamingInitializer.scala | 3 +-
.../flink/core/FlinkTableInitializer.scala | 2 +-
.../flink/core/conf/FlinkRunOption.scala | 3 +-
.../streampark/flink/core/StreamTableContext.scala | 4 +-
.../streampark/flink/core/TableContext.scala | 4 +-
.../streampark/flink/core/StreamTableContext.scala | 4 +-
.../streampark/flink/core/TableContext.scala | 4 +-
.../streampark/flink/core/StreamTableContext.scala | 4 +-
.../streampark/flink/core/TableContext.scala | 4 +-
.../streampark/spark/client/bean/StopRequest.scala | 3 +-
.../spark/client/bean/SubmitRequest.scala | 4 +-
.../spark/client/bean/SubmitResponse.scala | 2 +-
.../spark/client/proxy/SparkShimsProxy.scala | 2 +-
.../streampark/spark/client/impl/YarnClient.scala | 2 +-
.../spark/client/trait/SparkClientTrait.scala | 2 +-
.../spark/connector/kafka/offset/HBaseOffset.scala | 2 +-
.../spark/connector/kafka/offset/Offset.scala | 2 +-
.../spark/connector/kafka/offset/RedisOffset.scala | 2 +-
tools/dependencies/check-LICENSE.sh | 7 ++-
114 files changed, 221 insertions(+), 253 deletions(-)
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/ConfigOption.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/ConfigOption.scala
index 4fefa0d8b..0ec3c84b9 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/ConfigOption.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/ConfigOption.scala
@@ -17,7 +17,7 @@
package org.apache.streampark.common.conf
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import java.util.Properties
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/FlinkVersion.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/FlinkVersion.scala
index 6ba10923c..e4950e247 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/FlinkVersion.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/FlinkVersion.scala
@@ -18,17 +18,17 @@
package org.apache.streampark.common.conf
import org.apache.streampark.common.util.{CommandUtils, Logger}
+import org.apache.streampark.common.util.Implicits._
import java.io.File
-import java.net.{URL => NetURL}
+import java.net.URL
import java.util.function.Consumer
import java.util.regex.Pattern
-import scala.collection.convert.ImplicitConversions._
import scala.collection.mutable
/** @param flinkHome actual flink home that must be a readable local path */
-class FlinkVersion(val flinkHome: String) extends java.io.Serializable with
Logger {
+class FlinkVersion(val flinkHome: String) extends Serializable with Logger {
private[this] lazy val FLINK_VER_PATTERN =
Pattern.compile("^(\\d+\\.\\d+)(\\.)?.*$")
@@ -63,7 +63,7 @@ class FlinkVersion(val flinkHome: String) extends
java.io.Serializable with Logg
lib
}
- lazy val flinkLibs: List[NetURL] =
flinkLib.listFiles().map(_.toURI.toURL).toList
+ lazy val flinkLibs: List[URL] =
flinkLib.listFiles().map(_.toURI.toURL).toList
lazy val version: String = {
val cmd = List(
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/InternalConfigHolder.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/InternalConfigHolder.scala
index 7816ecd20..3d19d6f12 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/InternalConfigHolder.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/InternalConfigHolder.scala
@@ -19,15 +19,13 @@ package org.apache.streampark.common.conf
import org.apache.streampark.common.Constant
import org.apache.streampark.common.util.{Logger, SystemPropertyUtils}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import javax.annotation.{Nonnull, Nullable}
import java.util
import java.util.concurrent.ConcurrentHashMap
-import scala.collection.convert.ImplicitConversions._
-
/**
* Thread-safe configuration storage containers. All configurations will be
automatically
* initialized from the spring configuration items of the same name.
@@ -123,7 +121,7 @@ object InternalConfigHolder extends Logger {
/** Get keys of all registered ConfigOption. */
@Nonnull
- def keys(): util.Set[String] = {
+ def keys(): JavaSet[String] = {
val map = new util.HashMap[String, InternalOption](confOptions.size())
map.putAll(confOptions)
map.keySet()
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/SparkVersion.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/SparkVersion.scala
index 725d69463..c6ba76888 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/SparkVersion.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/SparkVersion.scala
@@ -18,35 +18,37 @@
package org.apache.streampark.common.conf
import org.apache.streampark.common.util.{CommandUtils, Logger}
-
-import org.apache.commons.lang3.StringUtils
+import org.apache.streampark.common.util.Implicits._
import java.io.File
-import java.net.{URL => NetURL}
-import java.util.concurrent.atomic.{AtomicBoolean, AtomicReference}
+import java.net.URL
import java.util.function.Consumer
import java.util.regex.Pattern
-import scala.collection.convert.ImplicitConversions._
import scala.collection.mutable
/** @param sparkHome actual spark home that must be a readable local path */
-class SparkVersion(val sparkHome: String) extends java.io.Serializable with
Logger {
+class SparkVersion(val sparkHome: String) extends Serializable with Logger {
private[this] lazy val SPARK_VER_PATTERN =
Pattern.compile("^(\\d+\\.\\d+)(\\.)?.*$")
private[this] lazy val SPARK_VERSION_PATTERN = Pattern.compile("(version)
(\\d+\\.\\d+\\.\\d+)")
- private[this] lazy val SPARK_SCALA_VERSION_PATTERN =
- Pattern.compile("^spark-core_(.*)-[0-9].*.jar$")
+ private[this] lazy val SPARK_SCALA_VERSION_PATTERN =
Pattern.compile("^spark-core_(.*)-[0-9].*.jar$")
- lazy val scalaVersion: String = {
- val matcher = SPARK_SCALA_VERSION_PATTERN.matcher(sparkCoreJar.getName)
- if (matcher.matches()) {
- matcher.group(1);
- } else {
- "2.12"
+ lazy val scalaVersion: String =
SPARK_SCALA_VERSION_PATTERN.matcher(sparkCoreJar.getName).group(1)
+
+ lazy val sparkCoreJar: File = {
+ val distJar =
sparkLib.listFiles().filter(_.getName.matches("spark-core.*\\.jar"))
+ distJar match {
+ case x if x.isEmpty =>
+ throw new IllegalArgumentException(s"[StreamPark] can no found
spark-core jar in $sparkLib")
+ case x if x.length > 1 =>
+ throw new IllegalArgumentException(
+ s"[StreamPark] found multiple spark-core jar in $sparkLib")
+ case _ =>
}
+ distJar.head
}
def checkVersion(throwException: Boolean = true): Boolean = {
@@ -54,7 +56,7 @@ class SparkVersion(val sparkHome: String) extends
java.io.Serializable with Logg
case Array(3, v, _) if v >= 1 && v <= 3 => true
case _ =>
if (throwException) {
- throw new UnsupportedOperationException(s"Unsupported flink version:
$version")
+ throw new UnsupportedOperationException(s"Unsupported spark version:
$version")
} else {
false
}
@@ -73,7 +75,7 @@ class SparkVersion(val sparkHome: String) extends
java.io.Serializable with Logg
lib
}
- lazy val sparkLibs: List[NetURL] =
sparkLib.listFiles().map(_.toURI.toURL).toList
+ lazy val sparkLibs: List[URL] =
sparkLib.listFiles().map(_.toURI.toURL).toList
lazy val majorVersion: String = {
if (version == null) {
@@ -86,9 +88,8 @@ class SparkVersion(val sparkHome: String) extends
java.io.Serializable with Logg
}
lazy val version: String = {
- val sparkVersion = new AtomicReference[String]
+ var sparkVersion: String = null
val cmd = List(s"$sparkHome/bin/spark-submit --version")
- val success = new AtomicBoolean(false)
val buffer = new mutable.StringBuilder
CommandUtils.execute(
sparkHome,
@@ -97,31 +98,18 @@ class SparkVersion(val sparkHome: String) extends
java.io.Serializable with Logg
override def accept(out: String): Unit = {
buffer.append(out).append("\n")
val matcher = SPARK_VERSION_PATTERN.matcher(out)
- if (matcher.find && StringUtils.isBlank(sparkVersion.get())) {
- success.set(true)
- sparkVersion.set(matcher.group(2))
+ if (matcher.find) {
+ sparkVersion = matcher.group(2)
}
}
})
+
logInfo(buffer.toString())
- if (!success.get()) {
+ if (sparkVersion == null) {
throw new IllegalStateException(s"[StreamPark] parse spark version
failed. $buffer")
}
buffer.clear()
- sparkVersion.get
- }
-
- lazy val sparkCoreJar: File = {
- val distJar =
sparkLib.listFiles().filter(_.getName.matches("spark-core.*\\.jar"))
- distJar match {
- case x if x.isEmpty =>
- throw new IllegalArgumentException(s"[StreamPark] can no found
spark-core jar in $sparkLib")
- case x if x.length > 1 =>
- throw new IllegalArgumentException(
- s"[StreamPark] found multiple spark-core jar in $sparkLib")
- case _ =>
- }
- distJar.head
+ sparkVersion
}
override def toString: String =
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/Workspace.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/Workspace.scala
index f002d8d28..cc9161f5e 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/Workspace.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/Workspace.scala
@@ -19,7 +19,7 @@ package org.apache.streampark.common.conf
import org.apache.streampark.common.enums.StorageType
import org.apache.streampark.common.util.{HdfsUtils, SystemPropertyUtils}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import java.net.URI
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/AssertUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/AssertUtils.scala
index 96cfda60e..00934cad7 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/AssertUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/AssertUtils.scala
@@ -17,14 +17,13 @@
package org.apache.streampark.common.util
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.common.util.Utils.isEmpty
import javax.annotation.Nullable
import java.util
-import scala.collection.convert.ImplicitConversions._
-
/** @since 2.2.0 */
object AssertUtils {
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/CURLBuilder.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/CURLBuilder.scala
index 9699935c4..c18ef1896 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/CURLBuilder.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/CURLBuilder.scala
@@ -17,9 +17,9 @@
package org.apache.streampark.common.util
-import java.util
+import org.apache.streampark.common.util.Implicits._
-import scala.collection.convert.ImplicitConversions._
+import java.util
class CURLBuilder(val url: String) {
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/CommandUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/CommandUtils.scala
index a86df3cd0..b5ccf677a 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/CommandUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/CommandUtils.scala
@@ -17,12 +17,13 @@
package org.apache.streampark.common.util
+import org.apache.streampark.common.util.Implicits._
+
import java.io._
import java.lang.{Iterable => JavaIterable}
import java.util.Scanner
import java.util.function.Consumer
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Failure, Success, Try}
object CommandUtils extends Logger {
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/ConfigUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/ConfigUtils.scala
index 617670392..28c182110 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/ConfigUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/ConfigUtils.scala
@@ -19,10 +19,10 @@ package org.apache.streampark.common.util
import org.apache.streampark.common.Constant
import org.apache.streampark.common.conf.ConfigKeys._
+import org.apache.streampark.common.util.Implicits._
-import java.util.{Map => JavaMap, Properties}
+import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
import scala.collection.immutable.{Map => ScalaMap}
import scala.util.Try
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/FileUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/FileUtils.scala
index 0f1996e30..fbeb6f412 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/FileUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/FileUtils.scala
@@ -17,7 +17,7 @@
package org.apache.streampark.common.util
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import java.io._
import java.net.URL
@@ -29,7 +29,6 @@ import java.util
import java.util.Scanner
import java.util.stream.Collectors
-import scala.collection.convert.ImplicitConversions._
import scala.collection.mutable
object FileUtils {
@@ -114,7 +113,7 @@ object FileUtils {
filename.drop(filename.lastIndexOf("."))
}
- def listFileAsURL(dirPath: String): util.List[URL] = {
+ def listFileAsURL(dirPath: String): JavaList[URL] = {
new File(dirPath) match {
case x if x.exists() && x.isDirectory =>
val files = x.listFiles()
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HBaseClient.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HBaseClient.scala
index d5f0b5fdf..fe02a2e1f 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HBaseClient.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HBaseClient.scala
@@ -18,6 +18,7 @@
package org.apache.streampark.common.util
import org.apache.streampark.common.conf.ConfigKeys._
+import org.apache.streampark.common.util.Implicits._
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.hbase.{HBaseConfiguration, TableName}
@@ -26,8 +27,6 @@ import org.apache.hadoop.security.UserGroupInformation
import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
-
class HBaseClient(func: () => Connection) extends Serializable {
lazy val connection: Connection = func()
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopConfigUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopConfigUtils.scala
index f88a09c8a..c33687e70 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopConfigUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopConfigUtils.scala
@@ -20,15 +20,14 @@ package org.apache.streampark.common.util
import org.apache.streampark.common.conf.{CommonConfig, InternalConfigHolder}
import org.apache.streampark.common.conf.ConfigKeys._
import org.apache.streampark.common.fs.LfsOperator
+import org.apache.streampark.common.util.Implicits._
import org.apache.commons.io.{FileUtils => ApacheFileUtils}
import java.io.File
import java.nio.charset.StandardCharsets
-import java.util.{Collections, Map => JavaMap, Optional}
+import java.util.{Collections, Optional}
-import scala.collection.JavaConverters._
-import scala.collection.convert.ImplicitConversions._
import scala.collection.immutable.ListMap
import scala.util.{Failure, Success, Try}
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopUtils.scala
index f659303ac..8ed2b807d 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopUtils.scala
@@ -19,6 +19,7 @@ package org.apache.streampark.common.util
import org.apache.streampark.common.conf.{CommonConfig, InternalConfigHolder}
import org.apache.streampark.common.conf.ConfigKeys._
+import org.apache.streampark.common.util.Implicits._
import org.apache.commons.collections.CollectionUtils
import org.apache.commons.lang3.StringUtils
@@ -39,7 +40,6 @@ import java.util
import java.util.{Timer, TimerTask}
import java.util.concurrent._
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Failure, Success, Try}
object HadoopUtils extends Logger {
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HostsUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HostsUtils.scala
index 08686a283..19aa757ca 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HostsUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HostsUtils.scala
@@ -17,12 +17,12 @@
package org.apache.streampark.common.util
+import org.apache.streampark.common.util.Implicits._
+
import io.netty.resolver.HostsFileParser
import java.net.InetAddress
-import java.util.{Map => JavaMap}
-import scala.collection.convert.ImplicitConversions._
import scala.collection.immutable.ListMap
object HostsUtils {
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HttpClientUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HttpClientUtils.scala
index c0aa5241c..4a228d238 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HttpClientUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HttpClientUtils.scala
@@ -17,6 +17,8 @@
package org.apache.streampark.common.util
+import org.apache.streampark.common.util.Implicits._
+
import org.apache.hc.client5.http.auth.{AuthSchemeFactory, AuthScope,
Credentials, StandardAuthScheme}
import org.apache.hc.client5.http.classic.methods.{HttpGet, HttpPost,
HttpUriRequestBase}
import org.apache.hc.client5.http.config.RequestConfig
@@ -32,9 +34,6 @@ import org.apache.hc.core5.net.URIBuilder
import java.nio.charset.{Charset, StandardCharsets}
import java.security.Principal
-import java.util.{List => JavaList, Map => JavaMap}
-
-import scala.collection.convert.ImplicitConversions._
object HttpClientUtils {
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/ImplicitsUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/Implicits.scala
similarity index 66%
rename from
streampark-common/src/main/scala/org/apache/streampark/common/util/ImplicitsUtils.scala
rename to
streampark-common/src/main/scala/org/apache/streampark/common/util/Implicits.scala
index 4dd147dc9..40db96ae5 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/ImplicitsUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/Implicits.scala
@@ -19,11 +19,40 @@ package org.apache.streampark.common.util
import org.apache.streampark.common.util.Utils.close
-import java.lang.{Boolean => JavaBool, Byte => JavaByte, Double => JavaDouble,
Float => JavaFloat, Integer => JavaInt, Long => JavaLong, Short => JavaShort}
+import java.lang.{Boolean => JavaBool, Double => JavaDouble, Float =>
JavaFloat, Integer => JavaInt, Long => JavaLong, Short => JavaShort}
+import scala.collection.convert.{DecorateAsJava, DecorateAsScala,
ToJavaImplicits, ToScalaImplicits}
import scala.language.implicitConversions
-object ImplicitsUtils {
+object Implicits extends ToScalaImplicits with ToJavaImplicits with
DecorateAsJava with DecorateAsScala {
+
+ type JavaMap[K, V] = java.util.Map[K, V]
+
+ type JavaHashMap[K, V] = java.util.HashMap[K, V]
+
+ type JavaLinkedMap[K, V] = java.util.LinkedHashMap[K, V]
+
+ type JavaCollection[T] = java.util.Collection[T]
+
+ type JavaList[T] = java.util.List[T]
+
+ type JavaArrayList[T] = java.util.ArrayList[T]
+
+ type JavaSet[T] = java.util.Set[T]
+
+ type JavaBool = java.lang.Boolean
+
+ type JavaByte = java.lang.Byte
+
+ type JavaDouble = java.lang.Double
+
+ type JavaFloat = java.lang.Float
+
+ type JavaInt = java.lang.Integer
+
+ type JavaLong = java.lang.Long
+
+ type JavaShort = java.lang.Short
implicit class AutoCloseImplicits[T <: AutoCloseable](autoCloseable: T) {
@@ -53,12 +82,9 @@ object ImplicitsUtils {
case c if c == classOf[JavaByte] => v.toByte.asInstanceOf[T]
case c if c == classOf[JavaInt] => JavaInt.valueOf(v).asInstanceOf[T]
case c if c == classOf[JavaLong] => JavaLong.valueOf(v).asInstanceOf[T]
- case c if c == classOf[JavaFloat] =>
- JavaFloat.valueOf(v).asInstanceOf[T]
- case c if c == classOf[JavaDouble] =>
- JavaDouble.valueOf(v).asInstanceOf[T]
- case c if c == classOf[JavaShort] =>
- JavaShort.valueOf(v).asInstanceOf[T]
+ case c if c == classOf[JavaFloat] =>
JavaFloat.valueOf(v).asInstanceOf[T]
+ case c if c == classOf[JavaDouble] =>
JavaDouble.valueOf(v).asInstanceOf[T]
+ case c if c == classOf[JavaShort] =>
JavaShort.valueOf(v).asInstanceOf[T]
case c if c == classOf[JavaBool] => JavaBool.valueOf(v).asInstanceOf[T]
case _ =>
throw new IllegalArgumentException(s"Unsupported type: $classType")
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/JdbcUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/JdbcUtils.scala
index aba8e35e9..95071fcd5 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/JdbcUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/JdbcUtils.scala
@@ -18,6 +18,7 @@
package org.apache.streampark.common.util
import org.apache.streampark.common.conf.ConfigKeys._
+import org.apache.streampark.common.util.Implicits._
import com.zaxxer.hikari.{HikariConfig, HikariDataSource}
@@ -26,7 +27,6 @@ import java.util.Properties
import java.util.concurrent.ConcurrentHashMap
import java.util.concurrent.locks.ReentrantLock
-import scala.collection.convert.ImplicitConversions._
import scala.collection.mutable
import scala.collection.mutable.ArrayBuffer
import scala.util.Try
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/MongoConfig.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/MongoConfig.scala
index 5c5094465..64d610490 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/MongoConfig.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/MongoConfig.scala
@@ -18,13 +18,12 @@
package org.apache.streampark.common.util
import org.apache.streampark.common.conf.ConfigKeys._
+import org.apache.streampark.common.util.Implicits._
import com.mongodb._
import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
-
object MongoConfig {
val client_uri = "client-uri"
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/PropertiesUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/PropertiesUtils.scala
index e95a7539a..c5153e986 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/PropertiesUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/PropertiesUtils.scala
@@ -17,6 +17,8 @@
package org.apache.streampark.common.util
+import org.apache.streampark.common.util.Implicits._
+
import com.typesafe.config.ConfigFactory
import org.apache.commons.lang3.StringUtils
import org.yaml.snakeyaml.Yaml
@@ -24,11 +26,10 @@ import org.yaml.snakeyaml.Yaml
import javax.annotation.Nonnull
import java.io._
-import java.util.{HashMap => JavaMap, LinkedHashMap => LinkedMap, Properties,
Scanner}
+import java.util.{Properties, Scanner}
import java.util.concurrent.atomic.AtomicInteger
import java.util.regex.Pattern
-import scala.collection.convert.ImplicitConversions._
import scala.collection.mutable
import scala.collection.mutable.{ArrayBuffer, Map => MutableMap}
@@ -59,7 +60,7 @@ object PropertiesUtils extends Logger {
prefix: String = "",
proper: MutableMap[String, String] = MutableMap[String, String]()):
Map[String, String] = {
v match {
- case map: LinkedMap[_, _] =>
+ case map: JavaLinkedMap[_, _] =>
map
.flatMap(x => {
prefix match {
@@ -203,31 +204,31 @@ object PropertiesUtils extends Logger {
}
def fromYamlTextAsJava(text: String): JavaMap[String, String] =
- new JavaMap[String, String](fromYamlText(text))
+ new JavaHashMap[String, String](fromYamlText(text))
def fromHoconTextAsJava(text: String): JavaMap[String, String] =
- new JavaMap[String, String](fromHoconText(text))
+ new JavaHashMap[String, String](fromHoconText(text))
def fromPropertiesTextAsJava(text: String): JavaMap[String, String] =
- new JavaMap[String, String](fromPropertiesText(text))
+ new JavaHashMap[String, String](fromPropertiesText(text))
def fromYamlFileAsJava(filename: String): JavaMap[String, String] =
- new JavaMap[String, String](fromYamlFile(filename))
+ new JavaHashMap[String, String](fromYamlFile(filename))
def fromHoconFileAsJava(filename: String): JavaMap[String, String] =
- new JavaMap[String, String](fromHoconFile(filename))
+ new JavaHashMap[String, String](fromHoconFile(filename))
def fromPropertiesFileAsJava(filename: String): JavaMap[String, String] =
- new JavaMap[String, String](fromPropertiesFile(filename))
+ new JavaHashMap[String, String](fromPropertiesFile(filename))
def fromYamlFileAsJava(inputStream: InputStream): JavaMap[String, String] =
- new JavaMap[String, String](fromYamlFile(inputStream))
+ new JavaHashMap[String, String](fromYamlFile(inputStream))
def fromHoconFileAsJava(inputStream: InputStream): JavaMap[String, String] =
- new JavaMap[String, String](fromHoconFile(inputStream))
+ new JavaHashMap[String, String](fromHoconFile(inputStream))
def fromPropertiesFileAsJava(inputStream: InputStream): JavaMap[String,
String] =
- new JavaMap[String, String](fromPropertiesFile(inputStream))
+ new JavaHashMap[String, String](fromPropertiesFile(inputStream))
/**
* @param file
@@ -242,7 +243,7 @@ object PropertiesUtils extends Logger {
def loadFlinkConfYaml(yaml: String): JavaMap[String, String] = {
require(yaml != null && yaml.nonEmpty, "[StreamPark] loadFlinkConfYaml:
yaml must not be null")
- val flinkConf = new JavaMap[String, String]()
+ val flinkConf = new JavaHashMap[String, String]()
val scanner: Scanner = new Scanner(yaml)
val lineNo: AtomicInteger = new AtomicInteger(0)
while (scanner.hasNextLine) {
@@ -370,13 +371,13 @@ object PropertiesUtils extends Logger {
}
@Nonnull def extractDynamicPropertiesAsJava(properties: String):
JavaMap[String, String] =
- new JavaMap[String, String](extractDynamicProperties(properties))
+ new JavaHashMap[String, String](extractDynamicProperties(properties))
@Nonnull def extractMultipleArgumentsAsJava(
args: Array[String]): JavaMap[String, JavaMap[String, String]] = {
val map =
- extractMultipleArguments(args).map(c => c._1 -> new JavaMap[String,
String](c._2))
- new JavaMap[String, JavaMap[String, String]](map)
+ extractMultipleArguments(args).map(c => c._1 -> new JavaHashMap[String,
String](c._2))
+ new JavaHashMap[String, JavaMap[String, String]](map)
}
}
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisClient.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisClient.scala
index b2ae6dcc8..49e02273b 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisClient.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisClient.scala
@@ -18,6 +18,7 @@
package org.apache.streampark.common.util
import org.apache.streampark.common.Constant
+import org.apache.streampark.common.util.Implicits._
import redis.clients.jedis._
import redis.clients.jedis.exceptions.JedisConnectionException
@@ -26,7 +27,6 @@ import java.util.concurrent.ConcurrentHashMap
import scala.annotation.meta.getter
import scala.annotation.tailrec
-import scala.collection.convert.ImplicitConversions._
import scala.util.Random
object RedisClient extends Logger {
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisEndpoint.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisEndpoint.scala
index e64079199..1edaa771f 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisEndpoint.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisEndpoint.scala
@@ -18,6 +18,7 @@
package org.apache.streampark.common.util
import org.apache.streampark.common.conf.ConfigKeys._
+import org.apache.streampark.common.util.Implicits._
import redis.clients.jedis.{Jedis, Protocol}
import redis.clients.jedis.util.{JedisClusterCRC16, JedisURIHelper,
SafeEncoder}
@@ -25,8 +26,6 @@ import redis.clients.jedis.util.{JedisClusterCRC16,
JedisURIHelper, SafeEncoder}
import java.net.URI
import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
-
/**
* RedisEndpoint represents a redis connection endpoint info: host, port, auth
password db number,
* and timeout
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisUtils.scala
index a5362158f..31e62c8a1 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisUtils.scala
@@ -17,10 +17,10 @@
package org.apache.streampark.common.util
+import org.apache.streampark.common.util.Implicits._
+
import redis.clients.jedis.{Jedis, JedisCluster, Pipeline, ScanParams}
-import scala.collection.JavaConverters._
-import scala.collection.convert.ImplicitConversions._
import scala.collection.immutable
import scala.util.{Failure, Success, Try}
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/SqlConvertUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/SqlConvertUtils.scala
index e07c3f17b..74d0408b2 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/SqlConvertUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/SqlConvertUtils.scala
@@ -17,12 +17,13 @@
package org.apache.streampark.common.util
+import org.apache.streampark.common.util.Implicits._
+
import java.util
import java.util.Scanner
import java.util.regex.Pattern
import scala.annotation.tailrec
-import scala.collection.convert.ImplicitConversions._
object SqlConvertUtils extends Logger {
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/Utils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/Utils.scala
index 9cfdd2ac4..e0613bc57 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/Utils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/Utils.scala
@@ -17,19 +17,19 @@
package org.apache.streampark.common.util
-import org.apache.streampark.common.util.ImplicitsUtils.AutoCloseImplicits
+import org.apache.streampark.common.util.Implicits._
+import org.apache.streampark.common.util.Implicits.AutoCloseImplicits
import org.apache.commons.lang3.StringUtils
import java.io._
import java.net.{HttpURLConnection, URL}
import java.time.{Duration, LocalDateTime}
-import java.util.{jar, Collection => JavaCollection, Map => JavaMap,
Properties, UUID}
+import java.util.{jar, Properties, UUID}
import java.util.concurrent.locks.LockSupport
import java.util.jar.{JarFile, JarInputStream}
import scala.annotation.tailrec
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Failure, Success, Try}
object Utils extends Logger {
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/YarnUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/YarnUtils.scala
index f7c6c8a63..9fdbbde9a 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/YarnUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/YarnUtils.scala
@@ -19,6 +19,7 @@ package org.apache.streampark.common.util
import org.apache.streampark.common.Constant
import org.apache.streampark.common.conf.{CommonConfig, InternalConfigHolder}
+import org.apache.streampark.common.util.Implicits._
import org.apache.commons.lang3.StringUtils
import org.apache.hadoop.fs.CommonConfigurationKeys
@@ -34,10 +35,8 @@ import java.io.IOException
import java.net.InetAddress
import java.security.PrivilegedExceptionAction
import java.util
-import java.util.{HashMap => JavaHashMap, List => JavaList}
import java.util.concurrent.TimeUnit
-import scala.collection.convert.ImplicitConversions._
import scala.collection.mutable
import scala.collection.mutable.ArrayBuffer
import scala.util.{Failure, Success, Try}
diff --git
a/streampark-common/src/main/scala/org/apache/streampark/common/util/ZooKeeperUtils.scala
b/streampark-common/src/main/scala/org/apache/streampark/common/util/ZooKeeperUtils.scala
index b0762e72c..4ebcdb43d 100644
---
a/streampark-common/src/main/scala/org/apache/streampark/common/util/ZooKeeperUtils.scala
+++
b/streampark-common/src/main/scala/org/apache/streampark/common/util/ZooKeeperUtils.scala
@@ -17,6 +17,8 @@
package org.apache.streampark.common.util
+import org.apache.streampark.common.util.Implicits._
+
import org.apache.curator.RetryPolicy
import org.apache.curator.framework.{CuratorFramework, CuratorFrameworkFactory}
import org.apache.curator.retry.RetryNTimes
@@ -24,7 +26,6 @@ import org.apache.zookeeper.CreateMode
import java.nio.charset.StandardCharsets
-import scala.collection.convert.ImplicitConversions._
import scala.collection.mutable
object ZooKeeperUtils {
diff --git
a/streampark-common/src/test/scala/org/apache/streampark/common/util/ImplicitsUtilsTest.scala
b/streampark-common/src/test/scala/org/apache/streampark/common/util/ImplicitsTest.scala
similarity index 94%
rename from
streampark-common/src/test/scala/org/apache/streampark/common/util/ImplicitsUtilsTest.scala
rename to
streampark-common/src/test/scala/org/apache/streampark/common/util/ImplicitsTest.scala
index 7b77e927c..7a36475fa 100644
---
a/streampark-common/src/test/scala/org/apache/streampark/common/util/ImplicitsUtilsTest.scala
+++
b/streampark-common/src/test/scala/org/apache/streampark/common/util/ImplicitsTest.scala
@@ -17,11 +17,11 @@
package org.apache.streampark.common.util
-import org.apache.streampark.common.util.ImplicitsUtils.{AutoCloseImplicits,
StringImplicits}
+import org.apache.streampark.common.util.Implicits.{AutoCloseImplicits,
StringImplicits}
import org.scalatest.funsuite.AnyFunSuite
-class ImplicitsUtilsTest extends AnyFunSuite {
+class ImplicitsTest extends AnyFunSuite {
test(
"AutoCloseImplicits.autoClose should close the resource after execution
and handle exceptions") {
class MockResource extends AutoCloseable {
diff --git
a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/runner/QuickStartRunner.java
b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/runner/QuickStartRunner.java
index 3fef15890..6396801c9 100644
---
a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/runner/QuickStartRunner.java
+++
b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/runner/QuickStartRunner.java
@@ -37,7 +37,6 @@ import org.springframework.boot.ApplicationRunner;
import org.springframework.core.annotation.Order;
import org.springframework.stereotype.Component;
-import java.util.HashMap;
import java.util.Map;
@Order
@@ -64,7 +63,8 @@ public class QuickStartRunner implements ApplicationRunner {
@Override
public void run(ApplicationArguments args) throws Exception {
- Map<String, HashMap<String, String>> map =
PropertiesUtils.extractMultipleArgumentsAsJava(args.getSourceArgs());
+
+ Map<String, Map<String, String>> map =
PropertiesUtils.extractMultipleArgumentsAsJava(args.getSourceArgs());
Map<String, String> quickstart = map.get("quickstart");
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/CancelRequest.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/CancelRequest.scala
index 2ac83eb73..8ee9577b6 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/CancelRequest.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/CancelRequest.scala
@@ -20,11 +20,10 @@ package org.apache.streampark.flink.client.bean
import org.apache.streampark.common.Constant
import org.apache.streampark.common.conf.FlinkVersion
import org.apache.streampark.common.enums.FlinkExecutionMode
+import org.apache.streampark.common.util.Implicits.JavaMap
import javax.annotation.Nullable
-import java.util.{Map => JavaMap}
-
case class CancelRequest(
id: Long,
flinkVersion: FlinkVersion,
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequest.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequest.scala
index 5e6cf0181..2e6ca27f7 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequest.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequest.scala
@@ -19,6 +19,7 @@ package org.apache.streampark.flink.client.bean
import org.apache.streampark.common.conf.{FlinkVersion, Workspace}
import org.apache.streampark.common.enums.{FlinkExecutionMode,
FlinkK8sRestExposedType}
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.util.FlinkUtils
import org.apache.commons.io.FileUtils
@@ -27,7 +28,6 @@ import
org.apache.flink.kubernetes.configuration.KubernetesConfigOptions
import javax.annotation.Nullable
import java.io.File
-import java.util.{Map => JavaMap}
case class DeployRequest(
flinkVersion: FlinkVersion,
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequestTrait.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequestTrait.scala
index 3a4ba8332..7dd3f8092 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequestTrait.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequestTrait.scala
@@ -19,11 +19,10 @@ package org.apache.streampark.flink.client.bean
import org.apache.streampark.common.conf.FlinkVersion
import org.apache.streampark.common.enums.FlinkExecutionMode
+import org.apache.streampark.common.util.Implicits.JavaMap
import javax.annotation.Nullable
-import java.util.{Map => JavaMap}
-
trait DeployRequestTrait {
val flinkVersion: FlinkVersion
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SavepointRequestTrait.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SavepointRequestTrait.scala
index 526d37163..22c60b62f 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SavepointRequestTrait.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SavepointRequestTrait.scala
@@ -20,11 +20,10 @@ package org.apache.streampark.flink.client.bean
import org.apache.streampark.common.Constant
import org.apache.streampark.common.conf.FlinkVersion
import org.apache.streampark.common.enums.FlinkExecutionMode
+import org.apache.streampark.common.util.Implicits._
import javax.annotation.Nullable
-import java.util.{Map => JavaMap}
-
trait SavepointRequestTrait {
val flinkVersion: FlinkVersion
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/ShutDownRequest.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/ShutDownRequest.scala
index cae66c518..89cf2199e 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/ShutDownRequest.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/ShutDownRequest.scala
@@ -19,11 +19,10 @@ package org.apache.streampark.flink.client.bean
import org.apache.streampark.common.conf.FlinkVersion
import org.apache.streampark.common.enums.FlinkExecutionMode
+import org.apache.streampark.common.util.Implicits.JavaMap
import javax.annotation.Nullable
-import java.util.{Map => JavaMap}
-
case class ShutDownRequest(
flinkVersion: FlinkVersion,
executionMode: FlinkExecutionMode,
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitRequest.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitRequest.scala
index cc1e795c0..1b5184ce3 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitRequest.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitRequest.scala
@@ -22,6 +22,7 @@ import org.apache.streampark.common.conf.{FlinkVersion,
Workspace}
import org.apache.streampark.common.conf.ConfigKeys._
import org.apache.streampark.common.enums._
import org.apache.streampark.common.util._
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.packer.pipeline.{BuildResult,
ShadedBuildResponse}
import org.apache.streampark.flink.util.FlinkUtils
import org.apache.streampark.shaded.com.fasterxml.jackson.databind.ObjectMapper
@@ -34,9 +35,7 @@ import javax.annotation.Nullable
import java.io.File
import java.net.URL
-import java.util.{Map => JavaMap}
-import scala.collection.convert.ImplicitConversions._
import scala.util.Try
case class SubmitRequest(
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitResponse.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitResponse.scala
index d4d8f037f..e2466a204 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitResponse.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitResponse.scala
@@ -17,9 +17,9 @@
package org.apache.streampark.flink.client.bean
-import javax.annotation.Nullable
+import org.apache.streampark.common.util.Implicits.JavaMap
-import java.util.{Map => JavaMap}
+import javax.annotation.Nullable
case class SubmitResponse(
clusterId: String,
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/TriggerSavepointRequest.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/TriggerSavepointRequest.scala
index fbc81a1f3..cb8024c2e 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/TriggerSavepointRequest.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/TriggerSavepointRequest.scala
@@ -20,11 +20,10 @@ package org.apache.streampark.flink.client.bean
import org.apache.streampark.common.Constant
import org.apache.streampark.common.conf.FlinkVersion
import org.apache.streampark.common.enums.FlinkExecutionMode
+import org.apache.streampark.common.util.Implicits.JavaMap
import javax.annotation.Nullable
-import java.util.{Map => JavaMap}
-
/** Trigger savepoint request. */
case class TriggerSavepointRequest(
id: Long,
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeApplicationClient.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeApplicationClient.scala
index 7de527637..0de23ca82 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeApplicationClient.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeApplicationClient.scala
@@ -30,8 +30,6 @@ import
org.apache.flink.kubernetes.configuration.KubernetesConfigOptions
/**
* Kubernetes native application mode submit.
- * @deprecated
- * Please use [[KubernetesApplicationClientV2]] instead.
*/
object KubernetesNativeApplicationClient extends KubernetesNativeClientTrait {
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeSessionClient.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeSessionClient.scala
index 746134c3c..79dca5c78 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeSessionClient.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeSessionClient.scala
@@ -19,6 +19,7 @@ package org.apache.streampark.flink.client.impl
import org.apache.streampark.common.enums.FlinkExecutionMode
import org.apache.streampark.common.util.{Logger, Utils}
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.client.`trait`.KubernetesNativeClientTrait
import org.apache.streampark.flink.client.bean._
import org.apache.streampark.flink.client.tool.FlinkSessionSubmitHelper
@@ -27,7 +28,6 @@ import
org.apache.streampark.flink.kubernetes.KubernetesRetriever
import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode
import org.apache.streampark.flink.kubernetes.model.ClusterKey
-import io.fabric8.kubernetes.api.model.{Config => _}
import org.apache.commons.lang3.StringUtils
import org.apache.flink.client.program.ClusterClient
import org.apache.flink.configuration._
@@ -38,8 +38,6 @@ import
org.apache.flink.kubernetes.kubeclient.{FlinkKubeClient, FlinkKubeClientF
import java.io.File
-import scala.collection.convert.ImplicitConversions._
-
/** Kubernetes native session mode submit. */
object KubernetesNativeSessionClient extends KubernetesNativeClientTrait with
Logger {
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnApplicationClient.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnApplicationClient.scala
index 8e7fd90db..8231857fc 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnApplicationClient.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnApplicationClient.scala
@@ -22,6 +22,7 @@ import org.apache.streampark.common.conf.Workspace
import org.apache.streampark.common.enums.FlinkDevelopmentMode
import org.apache.streampark.common.fs.FsOperator
import org.apache.streampark.common.util.{AssertUtils, FileUtils, HdfsUtils}
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.client.`trait`.YarnClientTrait
import org.apache.streampark.flink.client.bean._
import org.apache.streampark.flink.packer.pipeline.ShadedBuildResponse
@@ -40,7 +41,6 @@ import java.security.PrivilegedAction
import java.util
import java.util.Collections
-import scala.collection.JavaConverters._
import scala.collection.mutable.ListBuffer
/** yarn application mode submit */
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnPerJobClient.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnPerJobClient.scala
index abbf89126..557bb9967 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnPerJobClient.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnPerJobClient.scala
@@ -17,6 +17,7 @@
package org.apache.streampark.flink.client.impl
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.client.`trait`.YarnClientTrait
import org.apache.streampark.flink.client.bean._
import org.apache.streampark.flink.util.FlinkUtils
@@ -33,8 +34,6 @@ import org.apache.hadoop.yarn.api.records.ApplicationId
import java.io.File
import java.lang.{Boolean => JavaBool}
-import scala.collection.convert.ImplicitConversions._
-
/** yarn PerJob mode submit */
object YarnPerJobClient extends YarnClientTrait {
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnSessionClient.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnSessionClient.scala
index 14af82cc7..a5af099cd 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnSessionClient.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnSessionClient.scala
@@ -18,6 +18,7 @@
package org.apache.streampark.flink.client.impl
import org.apache.streampark.common.util.{AssertUtils, Utils}
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.client.`trait`.YarnClientTrait
import org.apache.streampark.flink.client.bean._
@@ -36,8 +37,6 @@ import org.apache.hadoop.yarn.util.ConverterUtils
import java.util
-import scala.collection.convert.ImplicitConversions._
-
/** Submit Job to YARN Session Cluster */
object YarnSessionClient extends YarnClientTrait {
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/FlinkClientTrait.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/FlinkClientTrait.scala
index da9187426..7a8891962 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/FlinkClientTrait.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/FlinkClientTrait.scala
@@ -23,6 +23,7 @@ import org.apache.streampark.common.conf.Workspace
import org.apache.streampark.common.enums._
import org.apache.streampark.common.fs.FsOperator
import org.apache.streampark.common.util._
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.client.bean._
import org.apache.streampark.flink.core.FlinkClusterClient
import org.apache.streampark.flink.core.conf.FlinkRunOption
@@ -42,9 +43,8 @@ import org.apache.flink.runtime.jobgraph.{JobGraph,
SavepointConfigOptions}
import org.apache.flink.util.Preconditions.checkNotNull
import java.io.File
-import java.util.{Collections, List => JavaList, Map => JavaMap}
+import java.util.Collections
-import scala.collection.convert.ImplicitConversions._
import scala.collection.mutable
import scala.collection.mutable.ArrayBuffer
import scala.util.{Failure, Success, Try}
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/YarnClientTrait.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/YarnClientTrait.scala
index e1d8904be..853bc378b 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/YarnClientTrait.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/YarnClientTrait.scala
@@ -18,7 +18,7 @@
package org.apache.streampark.flink.client.`trait`
import org.apache.streampark.common.util.{AssertUtils, ExceptionUtils}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.client.bean._
import org.apache.flink.api.common.JobID
diff --git
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/test/scala/org/apache/streampark/flink/client/test/YarnPerJobTestCase.scala
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/test/scala/org/apache/streampark/flink/client/test/YarnPerJobTestCase.scala
index cbfb3bd85..db540f678 100644
---
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/test/scala/org/apache/streampark/flink/client/test/YarnPerJobTestCase.scala
+++
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/test/scala/org/apache/streampark/flink/client/test/YarnPerJobTestCase.scala
@@ -17,6 +17,7 @@
package org.apache.streampark.flink.client.test
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.common.util.Logger
import org.apache.streampark.flink.client.bean
import org.apache.streampark.flink.core.conf.FlinkRunOption
@@ -41,9 +42,6 @@ import java.io.File
import java.lang.reflect.Method
import java.util
-import scala.collection.JavaConverters._
-import scala.collection.convert.ImplicitConversions._
-
/** perJob to submit jobs programmatically, */
object YarnPerJobTestCase extends Logger {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/conf/ThresholdConf.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/conf/ThresholdConf.scala
index 839847028..bf0a0bd53 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/conf/ThresholdConf.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/conf/ThresholdConf.scala
@@ -22,8 +22,7 @@ import
org.apache.streampark.flink.connector.conf.FailoverStorageType.{Console,
import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
case class ThresholdConf(prefixStr: String, parameters: Properties) {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverChecker.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverChecker.scala
index 7a05b99c2..85b67f9f1 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverChecker.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverChecker.scala
@@ -21,7 +21,7 @@ import org.apache.streampark.common.util.{Logger, ThreadUtils}
import java.util.concurrent.{Executors, ScheduledExecutorService,
ThreadFactory, TimeUnit}
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
import scala.collection.mutable.ListBuffer
case class FailoverChecker(delayTime: Long) extends AutoCloseable with Logger {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverWriter.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverWriter.scala
index 1ef414dde..390a90125 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverWriter.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverWriter.scala
@@ -26,7 +26,7 @@ import org.apache.kafka.clients.producer.{Callback,
KafkaProducer, ProducerRecor
import java.util._
import java.util.concurrent.locks.ReentrantLock
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
class FailoverWriter(failoverStorage: FailoverStorageType, properties:
Properties)
extends AutoCloseable
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/SinkBuffer.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/SinkBuffer.scala
index 8cb9899c0..27b763456 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/SinkBuffer.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/SinkBuffer.scala
@@ -23,7 +23,7 @@ import java.util
import java.util.Collections
import java.util.concurrent.CopyOnWriteArrayList
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
case class SinkBuffer(writer: SinkWriter, flushInterval: Long, bufferSize: Int)
extends AutoCloseable
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/test/scala/org/apache/streampark/flink/connector/failover/SinkRequestTest.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/test/scala/org/apache/streampark/flink/connector/failover/SinkRequestTest.scala
index a67917565..634a8cc61 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/test/scala/org/apache/streampark/flink/connector/failover/SinkRequestTest.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/test/scala/org/apache/streampark/flink/connector/failover/SinkRequestTest.scala
@@ -19,7 +19,7 @@ package org.apache.streampark.flink.connector.failover
import org.junit.jupiter.api.{Assertions, Test}
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
class SinkRequestTest {
@Test
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseHttpConfig.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseHttpConfig.scala
index 041dbe5ac..83b92516a 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseHttpConfig.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseHttpConfig.scala
@@ -23,7 +23,7 @@ import org.apache.streampark.common.Constant
import java.util.{Base64, Properties}
import java.util.concurrent.ThreadLocalRandom
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
/**
* Flink sink for Clickhouse database. Powered by Async Http Client.
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseSinkConfigOption.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseSinkConfigOption.scala
index b9a8df8d5..4b054fd0c 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseSinkConfigOption.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseSinkConfigOption.scala
@@ -25,7 +25,7 @@ import
org.asynchttpclient.config.AsyncHttpClientConfigDefaults
import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
object ClickHouseSinkConfigOption {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkFunction.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkFunction.scala
index 707b14737..c13db6b3d 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkFunction.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkFunction.scala
@@ -34,7 +34,7 @@ import java.util
import java.util.Properties
import java.util.concurrent.atomic.AtomicLong
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
import scala.util.Try
class ClickHouseSinkFunction[T](apiType: ApiType = ApiType.scala, config:
Properties)
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkWriter.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkWriter.scala
index f2b147901..2b7c21995 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkWriter.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkWriter.scala
@@ -26,7 +26,7 @@ import org.asynchttpclient.{AsyncHttpClient,
DefaultAsyncHttpClientConfig, Dsl}
import java.util.concurrent._
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
import scala.collection.mutable.ListBuffer
case class ClickHouseSinkWriter(clickHouseConfig: ClickHouseHttpConfig)
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseWriterTask.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseWriterTask.scala
index 65c4fdf9d..3aaeabbaf 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseWriterTask.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseWriterTask.scala
@@ -26,7 +26,7 @@ import org.asynchttpclient.{AsyncHttpClient,
ListenableFuture, Request, Response
import java.util.concurrent.{BlockingQueue, ExecutorService, TimeUnit}
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
import scala.util.Try
case class ClickHouseWriterTask(
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisConfig.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisConfig.scala
index 5407886e8..2046a6817 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisConfig.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisConfig.scala
@@ -20,7 +20,7 @@ import org.apache.streampark.common.Constant
import org.apache.streampark.common.conf.ConfigKeys
import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
object DorisConfig {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisSinkConfigOption.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisSinkConfigOption.scala
index 89a86d66c..91149d6a6 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisSinkConfigOption.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisSinkConfigOption.scala
@@ -22,7 +22,7 @@ import org.apache.streampark.common.util.ConfigUtils
import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
object DorisSinkConfigOption {
val DORIS_SINK_PREFIX = "doris.sink"
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/conf/ESSinkConfigOption.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/conf/ESSinkConfigOption.scala
index c7d7bd091..b64645cea 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/conf/ESSinkConfigOption.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/conf/ESSinkConfigOption.scala
@@ -19,11 +19,10 @@ package
org.apache.streampark.flink.connector.elasticsearch5.conf
import org.apache.streampark.common.conf.ConfigOption
import org.apache.streampark.common.util.ConfigUtils
+import org.apache.streampark.common.util.Implicits._
import java.net.InetSocketAddress
-import java.util.{Map => JavaMap, Properties}
-
-import scala.collection.convert.ImplicitConversions._
+import java.util.Properties
object ESSinkConfigOption {
val ES_SINK_PREFIX = "es.sink"
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/sink/ES5Sink.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/sink/ES5Sink.scala
index c1e2ba742..6f999cbc0 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/sink/ES5Sink.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/sink/ES5Sink.scala
@@ -31,10 +31,10 @@ import
org.apache.flink.streaming.connectors.elasticsearch.util.RetryRejectedExe
import org.apache.flink.streaming.connectors.elasticsearch5.ElasticsearchSink
import org.elasticsearch.action.ActionRequest
-import java.util.{Map => JavaMap, Properties}
+import java.util.Properties
import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
object ES5Sink {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/conf/ES6SinkConfigOption.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/conf/ES6SinkConfigOption.scala
index ea6ff263c..3e3629d24 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/conf/ES6SinkConfigOption.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/conf/ES6SinkConfigOption.scala
@@ -23,9 +23,9 @@ import org.apache.streampark.common.util.ConfigUtils
import org.apache.http.HttpHost
import java.net.InetSocketAddress
-import java.util.{Map => JavaMap, Properties}
+import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
object ESSinkConfigOption {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/sink/ES6Sink.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/sink/ES6Sink.scala
index 5eacc0bf2..ff1067f93 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/sink/ES6Sink.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/sink/ES6Sink.scala
@@ -36,7 +36,7 @@ import org.elasticsearch.action.ActionRequest
import java.util.Properties
import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
object ES6Sink {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/conf/ES7SinkConfigOption.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/conf/ES7SinkConfigOption.scala
index 2feddef30..60f8b44a3 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/conf/ES7SinkConfigOption.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/conf/ES7SinkConfigOption.scala
@@ -23,9 +23,9 @@ import org.apache.streampark.common.util.ConfigUtils
import org.apache.http.HttpHost
import java.net.InetSocketAddress
-import java.util.{Map => JavaMap, Properties}
+import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
object ES7SinkConfigOption {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/sink/ES7Sink.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/sink/ES7Sink.scala
index 981772188..2399baad4 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/sink/ES7Sink.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/sink/ES7Sink.scala
@@ -38,7 +38,7 @@ import org.elasticsearch.action.ActionRequest
import java.util.Properties
import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
object ES7Sink {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSinkFunction.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSinkFunction.scala
index 072fc8afc..e136607da 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSinkFunction.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSinkFunction.scala
@@ -33,7 +33,7 @@ import java.util.Properties
import java.util.concurrent.{Executors, ScheduledExecutorService, TimeUnit}
import java.util.concurrent.atomic.{AtomicBoolean, AtomicLong}
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
import scala.collection.mutable.ArrayBuffer
class HBaseSinkFunction[T](apiType: ApiType = ApiType.scala, tabName: String,
prop: Properties)
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSourceFunction.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSourceFunction.scala
index 5ea77453f..d91c168b9 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSourceFunction.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSourceFunction.scala
@@ -37,7 +37,7 @@ import org.apache.hadoop.hbase.client.{Result, Table}
import java.lang
import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
import scala.util.{Success, Try}
class HBaseSourceFunction[R: TypeInformation](apiType: ApiType =
ApiType.scala, prop: Properties)
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/request/HBaseRequest.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/request/HBaseRequest.scala
index d82bfe14b..03d875996 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/request/HBaseRequest.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/request/HBaseRequest.scala
@@ -31,7 +31,7 @@ import java.util.concurrent.{CompletableFuture, Executors,
ExecutorService, Time
import java.util.function.{Consumer, Supplier}
import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
object HBaseRequest {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpSinkWriter.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpSinkWriter.scala
index a7b2f3998..d3ee49947 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpSinkWriter.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpSinkWriter.scala
@@ -25,7 +25,7 @@ import org.asynchttpclient.{AsyncHttpClient, Dsl}
import java.util.concurrent._
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
import scala.collection.mutable.ListBuffer
case class HttpSinkWriter(thresholdConf: ThresholdConf, header: Map[String,
String])
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpWriterTask.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpWriterTask.scala
index 2dda8cc05..09dd4ad9c 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpWriterTask.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpWriterTask.scala
@@ -28,7 +28,7 @@ import org.asynchttpclient.{AsyncHttpClient,
ListenableFuture, Request, Response
import java.util
import java.util.concurrent.{BlockingQueue, ExecutorService, TimeUnit}
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
import scala.util.Try
case class HttpWriterTask(
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/sink/HttpSink.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/sink/HttpSink.scala
index 058bda317..475f6e4ab 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/sink/HttpSink.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/sink/HttpSink.scala
@@ -29,7 +29,7 @@ import org.apache.http.client.methods._
import java.util.Properties
import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
object HttpSink {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-influx/src/main/scala/org/apache/streampark/flink/connector/influx/function/InfluxFunction.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-influx/src/main/scala/org/apache/streampark/flink/connector/influx/function/InfluxFunction.scala
index c3565c9a2..24a73e08d 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-influx/src/main/scala/org/apache/streampark/flink/connector/influx/function/InfluxFunction.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-influx/src/main/scala/org/apache/streampark/flink/connector/influx/function/InfluxFunction.scala
@@ -28,10 +28,10 @@ import
org.apache.flink.streaming.api.functions.sink.SinkFunction.Context
import org.influxdb.{InfluxDB, InfluxDBFactory}
import org.influxdb.dto.Point
-import java.util.{Map => JavaMap, Properties}
+import java.util.Properties
import java.util.concurrent.TimeUnit
-import scala.collection.convert.ImplicitConversions._;
+import org.apache.streampark.common.util.Implicits._;
class InfluxFunction[T](config: Properties)(implicit endpoint: InfluxEntity[T])
extends RichSinkFunction[T]
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-jdbc/src/main/scala/org/apache/streampark/flink/connector/jdbc/internal/JdbcSourceFunction.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-jdbc/src/main/scala/org/apache/streampark/flink/connector/jdbc/internal/JdbcSourceFunction.scala
index fc7b1b601..1ea839031 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-jdbc/src/main/scala/org/apache/streampark/flink/connector/jdbc/internal/JdbcSourceFunction.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-jdbc/src/main/scala/org/apache/streampark/flink/connector/jdbc/internal/JdbcSourceFunction.scala
@@ -33,9 +33,7 @@ import
org.apache.flink.streaming.api.functions.source.SourceFunction.SourceCont
import java.lang
import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
import scala.collection.Map
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Success, Try}
class JdbcSourceFunction[R: TypeInformation](apiType: ApiType = ApiType.scala,
jdbc: Properties)
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-kafka/src/main/scala/org/apache/streampark/flink/connector/kafka/source/KafkaSource.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-kafka/src/main/scala/org/apache/streampark/flink/connector/kafka/source/KafkaSource.scala
index aea660284..a5e0feb46 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-kafka/src/main/scala/org/apache/streampark/flink/connector/kafka/source/KafkaSource.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-kafka/src/main/scala/org/apache/streampark/flink/connector/kafka/source/KafkaSource.scala
@@ -35,7 +35,7 @@ import java.util.Properties
import java.util.regex.Pattern
import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
import scala.util.{Failure, Success, Try}
object KafkaSource {
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-mongo/src/main/scala/org/apache/streampark/flink/connector/mongo/internal/MongoSourceFunction.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-mongo/src/main/scala/org/apache/streampark/flink/connector/mongo/internal/MongoSourceFunction.scala
index 0e86fbae1..91ee78b03 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-mongo/src/main/scala/org/apache/streampark/flink/connector/mongo/internal/MongoSourceFunction.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-mongo/src/main/scala/org/apache/streampark/flink/connector/mongo/internal/MongoSourceFunction.scala
@@ -38,7 +38,7 @@ import org.bson.Document
import java.lang
import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
import scala.util.{Success, Try}
class MongoSourceFunction[R: TypeInformation](
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/conf/RedisSinkConfigOption.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/conf/RedisSinkConfigOption.scala
index 1bcfa4530..f41168515 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/conf/RedisSinkConfigOption.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/conf/RedisSinkConfigOption.scala
@@ -22,7 +22,7 @@ import org.apache.streampark.common.util.ConfigUtils
import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
object RedisSinkConfigOption {
val REDIS_SINK_PREFIX = "redis.sink"
diff --git
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/sink/RedisSink.scala
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/sink/RedisSink.scala
index 9adb40a43..a7f10fe56 100644
---
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/sink/RedisSink.scala
+++
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/sink/RedisSink.scala
@@ -36,7 +36,7 @@ import java.util
import java.util.Properties
import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
import scala.util.Try
object RedisSink {
diff --git
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/FlinkK8sWatchController.scala
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/FlinkK8sWatchController.scala
index 1faf0c69d..f96e9e5f6 100644
---
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/FlinkK8sWatchController.scala
+++
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/FlinkK8sWatchController.scala
@@ -18,6 +18,7 @@
package org.apache.streampark.flink.kubernetes
import org.apache.streampark.common.util.{Logger, Utils}
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.kubernetes.model._
import com.github.benmanes.caffeine.cache.{Cache, Caffeine}
@@ -25,8 +26,6 @@ import com.github.benmanes.caffeine.cache.{Cache, Caffeine}
import java.util.Objects
import java.util.concurrent.TimeUnit
-import scala.collection.convert.ImplicitConversions._
-
/** Tracking info cache pool on flink kubernetes mode. */
class FlinkK8sWatchController extends Logger with AutoCloseable {
diff --git
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/KubernetesRetriever.scala
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/KubernetesRetriever.scala
index 3e43db40e..b1d1784f8 100644
---
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/KubernetesRetriever.scala
+++
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/KubernetesRetriever.scala
@@ -17,7 +17,7 @@
package org.apache.streampark.flink.kubernetes
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.common.util.Logger
import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode
import org.apache.streampark.flink.kubernetes.ingress.IngressController
@@ -34,7 +34,6 @@ import org.apache.hc.core5.util.Timeout
import javax.annotation.Nullable
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Failure, Success, Try}
object KubernetesRetriever extends Logger {
diff --git
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/PodTemplateParser.scala
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/PodTemplateParser.scala
index 2cfa3c705..da021d1fb 100644
---
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/PodTemplateParser.scala
+++
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/PodTemplateParser.scala
@@ -17,14 +17,14 @@
package org.apache.streampark.flink.kubernetes
+import org.apache.streampark.common.util.Implicits._
+
import org.apache.commons.collections.CollectionUtils
import org.apache.commons.lang3.StringUtils
import org.yaml.snakeyaml.Yaml
import java.util
-import java.util.{List => JavaList, Map => JavaMap}
-import scala.collection.convert.ImplicitConversions._
import scala.util.Try
import scala.util.control.Breaks.{break, breakable}
diff --git
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/helper/KubernetesDeploymentHelper.scala
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/helper/KubernetesDeploymentHelper.scala
index ed0dc7e1c..25b25b871 100644
---
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/helper/KubernetesDeploymentHelper.scala
+++
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/helper/KubernetesDeploymentHelper.scala
@@ -18,7 +18,7 @@
package org.apache.streampark.flink.kubernetes.helper
import org.apache.streampark.common.util.{Logger, SystemPropertyUtils}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.kubernetes.KubernetesRetriever
import com.google.common.base.Charsets
@@ -28,7 +28,6 @@ import io.fabric8.kubernetes.client.DefaultKubernetesClient
import java.io.File
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Success, Try}
object KubernetesDeploymentHelper extends Logger {
diff --git
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressController.scala
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressController.scala
index 5bc4b2650..955fb33e4 100644
---
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressController.scala
+++
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressController.scala
@@ -17,7 +17,7 @@
package org.apache.streampark.flink.kubernetes.ingress
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.common.util.Logger
import io.fabric8.kubernetes.client.DefaultKubernetesClient
diff --git
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1.scala
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1.scala
index bea4e627e..064af32b7 100644
---
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1.scala
+++
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1.scala
@@ -17,13 +17,12 @@
package org.apache.streampark.flink.kubernetes.ingress
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import io.fabric8.kubernetes.api.model.networking.v1.IngressBuilder
import io.fabric8.kubernetes.client.DefaultKubernetesClient
import org.apache.flink.client.program.ClusterClient
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Failure, Success, Try}
class IngressStrategyV1 extends IngressStrategy {
diff --git
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1beta1.scala
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1beta1.scala
index 62c18ca2f..f24f3dbb3 100644
---
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1beta1.scala
+++
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1beta1.scala
@@ -17,7 +17,7 @@
package org.apache.streampark.flink.kubernetes.ingress
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import io.fabric8.kubernetes.api.model.IntOrString
import io.fabric8.kubernetes.api.model.networking.v1beta1.IngressBuilder
@@ -25,7 +25,6 @@ import io.fabric8.kubernetes.client.DefaultKubernetesClient
import org.apache.commons.lang3.StringUtils
import org.apache.flink.client.program.ClusterClient
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Failure, Success, Try}
class IngressStrategyV1beta1 extends IngressStrategy {
diff --git
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/watcher/FlinkJobStatusWatcher.scala
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/watcher/FlinkJobStatusWatcher.scala
index 6e5e03bd4..22446cec2 100644
---
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/watcher/FlinkJobStatusWatcher.scala
+++
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/watcher/FlinkJobStatusWatcher.scala
@@ -17,6 +17,7 @@
package org.apache.streampark.flink.kubernetes.watcher
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.common.util.Logger
import org.apache.streampark.flink.kubernetes._
import org.apache.streampark.flink.kubernetes.enums.{FlinkJobState,
FlinkK8sExecuteMode}
@@ -42,7 +43,6 @@ import java.io.File
import java.nio.charset.StandardCharsets
import java.util.concurrent.{ScheduledFuture, TimeUnit}
-import scala.collection.convert.ImplicitConversions._
import scala.concurrent.{Await, ExecutionContext,
ExecutionContextExecutorService, Future}
import scala.concurrent.duration.DurationLong
import scala.language.postfixOps
diff --git
a/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/FlinkRestJsonTest.scala
b/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/FlinkRestJsonTest.scala
index c614e0a2c..74528d875 100644
---
a/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/FlinkRestJsonTest.scala
+++
b/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/FlinkRestJsonTest.scala
@@ -17,6 +17,7 @@
package org.apache.streampark.flink.kubernetes
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.kubernetes.helper.KubernetesDeploymentHelper
import org.apache.streampark.flink.kubernetes.watcher.{Checkpoint,
FlinkRestJmConfigItem, FlinkRestOverview, JobDetails}
@@ -33,7 +34,6 @@ import org.junit.jupiter.api.Test
import java.io.File
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Failure, Success, Try}
// scalastyle:off println
diff --git
a/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/PodTemplateParserTest.scala
b/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/PodTemplateParserTest.scala
index ba8f0ae0b..ca37bbba6 100644
---
a/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/PodTemplateParserTest.scala
+++
b/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/PodTemplateParserTest.scala
@@ -17,11 +17,11 @@
package org.apache.streampark.flink.kubernetes
+import org.apache.streampark.common.util.Implicits._
+
import org.junit.jupiter.api.Assertions.assertEquals
import org.junit.jupiter.api.Test
-import scala.collection.convert.ImplicitConversions._
-
// scalastyle:off println
class PodTemplateParserTest {
diff --git
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/DependencyInfo.scala
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/DependencyInfo.scala
index 9384fe9c4..bc035bdd8 100644
---
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/DependencyInfo.scala
+++
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/DependencyInfo.scala
@@ -17,9 +17,7 @@
package org.apache.streampark.flink.packer.maven
-import java.util.{List => JavaList}
-
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
/**
* @param mavenArts
diff --git
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/MavenTool.scala
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/MavenTool.scala
index b640ad976..b0f199648 100644
---
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/MavenTool.scala
+++
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/MavenTool.scala
@@ -21,6 +21,7 @@ import org.apache.streampark.common.Constant
import org.apache.streampark.common.conf.{InternalConfigHolder, Workspace}
import org.apache.streampark.common.conf.CommonConfig.{MAVEN_AUTH_PASSWORD,
MAVEN_AUTH_USER, MAVEN_REMOTE_URL}
import org.apache.streampark.common.util.{AssertUtils, Logger, Utils}
+import org.apache.streampark.common.util.Implicits._
import com.google.common.collect.Lists
import org.apache.maven.plugins.shade.{DefaultShader, ShadeRequest}
@@ -44,9 +45,7 @@ import javax.annotation.{Nonnull, Nullable}
import java.io.File
import java.util
-import java.util.{List => JavaList, Set => JavaSet}
-import scala.collection.convert.ImplicitConversions._
import scala.collection.mutable.ArrayBuffer
import scala.util.Try
diff --git
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildPipeline.scala
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildPipeline.scala
index b711d3ed0..25062f94b 100644
---
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildPipeline.scala
+++
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildPipeline.scala
@@ -18,11 +18,11 @@
package org.apache.streampark.flink.packer.pipeline
import org.apache.streampark.common.util.{Logger, ThreadUtils}
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.packer.pipeline.BuildPipeline.executor
import java.util.concurrent.{Callable, LinkedBlockingQueue,
ThreadPoolExecutor, TimeUnit}
-import scala.collection.convert.ImplicitConversions._
import scala.collection.mutable
import scala.concurrent.{ExecutionContext, ExecutionContextExecutorService}
import scala.util.{Failure, Success, Try}
diff --git
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/DockerResolveSnapshot.scala
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/DockerResolveSnapshot.scala
index 14cd76af0..61fa36bc5 100644
---
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/DockerResolveSnapshot.scala
+++
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/DockerResolveSnapshot.scala
@@ -17,14 +17,11 @@
package org.apache.streampark.flink.packer.pipeline
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.common.util.Utils
import com.fasterxml.jackson.annotation.JsonIgnoreProperties
-import java.util.{List => JavaList}
-
-import scala.collection.convert.ImplicitConversions._
-
/** Snapshot for docker resolved progress */
@JsonIgnoreProperties(ignoreUnknown = true)
case class DockerResolvedSnapshot(
diff --git
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/PipelineSnapshot.scala
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/PipelineSnapshot.scala
index ea84594ca..2bf492034 100644
---
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/PipelineSnapshot.scala
+++
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/PipelineSnapshot.scala
@@ -17,12 +17,10 @@
package org.apache.streampark.flink.packer.pipeline
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.common.util.Utils
import java.lang.{Long => JavaLong}
-import java.util.{Map => JavaMap}
-
-import scala.collection.JavaConverters._
/**
* Snapshot for a BuildPipeline instance. see
diff --git
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkRemoteBuildPipeline.scala
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkRemoteBuildPipeline.scala
index b0d9b90c6..c7d16e44d 100644
---
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkRemoteBuildPipeline.scala
+++
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkRemoteBuildPipeline.scala
@@ -19,13 +19,12 @@ package org.apache.streampark.flink.packer.pipeline.impl
import org.apache.streampark.common.enums.FlinkDevelopmentMode
import org.apache.streampark.common.fs.{FsOperator, LfsOperator}
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.packer.maven.MavenTool
import org.apache.streampark.flink.packer.pipeline._
import java.io.File
-import scala.collection.convert.ImplicitConversions._
-
/** Building pipeline for flink standalone session mode */
class FlinkRemoteBuildPipeline(request: FlinkRemotePerJobBuildRequest) extends
BuildPipeline {
diff --git
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkYarnApplicationBuildPipeline.scala
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkYarnApplicationBuildPipeline.scala
index 93124d22d..66d5fcff3 100644
---
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkYarnApplicationBuildPipeline.scala
+++
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkYarnApplicationBuildPipeline.scala
@@ -20,7 +20,7 @@ package org.apache.streampark.flink.packer.pipeline.impl
import org.apache.streampark.common.conf.Workspace
import org.apache.streampark.common.enums.FlinkDevelopmentMode
import org.apache.streampark.common.fs.{FsOperator, HdfsOperator, LfsOperator}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.packer.maven.MavenTool
import org.apache.streampark.flink.packer.pipeline._
@@ -28,8 +28,6 @@ import org.apache.commons.codec.digest.DigestUtils
import java.io.{File, FileInputStream, IOException}
-import scala.collection.convert.ImplicitConversions._
-
/** Building pipeline for flink yarn application mode */
class FlinkYarnApplicationBuildPipeline(request:
FlinkYarnApplicationBuildRequest)
extends BuildPipeline {
diff --git
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/SparkYarnApplicationBuildPipeline.scala
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/SparkYarnApplicationBuildPipeline.scala
index 3cac30ada..9b91d1dcb 100644
---
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/SparkYarnApplicationBuildPipeline.scala
+++
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/SparkYarnApplicationBuildPipeline.scala
@@ -20,7 +20,7 @@ package org.apache.streampark.flink.packer.pipeline.impl
import org.apache.streampark.common.conf.Workspace
import org.apache.streampark.common.enums.FlinkDevelopmentMode
import org.apache.streampark.common.fs.{FsOperator, HdfsOperator, LfsOperator}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.packer.maven.MavenTool
import org.apache.streampark.flink.packer.pipeline._
@@ -28,8 +28,6 @@ import org.apache.commons.codec.digest.DigestUtils
import java.io.{File, FileInputStream, IOException}
-import scala.collection.convert.ImplicitConversions._
-
/** Building pipeline for spark yarn application mode */
class SparkYarnApplicationBuildPipeline(request:
SparkYarnApplicationBuildRequest)
extends BuildPipeline {
diff --git
a/streampark-flink/streampark-flink-packer/src/test/scala/org/apache/streampark/flink/packer/MavenToolSpec.scala
b/streampark-flink/streampark-flink-packer/src/test/scala/org/apache/streampark/flink/packer/MavenToolSpec.scala
index 821577630..67735d9cf 100644
---
a/streampark-flink/streampark-flink-packer/src/test/scala/org/apache/streampark/flink/packer/MavenToolSpec.scala
+++
b/streampark-flink/streampark-flink-packer/src/test/scala/org/apache/streampark/flink/packer/MavenToolSpec.scala
@@ -17,6 +17,7 @@
package org.apache.streampark.flink.packer
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.packer.maven.{Artifact, DependencyInfo,
MavenTool}
import org.apache.commons.io.FileUtils
@@ -27,8 +28,6 @@ import org.scalatest.wordspec.AnyWordSpec
import java.io.File
import java.util.jar.JarFile
-import scala.collection.convert.ImplicitConversions._
-
class MavenToolSpec extends AnyWordSpec with BeforeAndAfterAll with Matchers {
val outputDir = "MavenToolSpec-output/"
diff --git
a/streampark-flink/streampark-flink-proxy/src/main/scala/org/apache/streampark/flink/proxy/FlinkShimsProxy.scala
b/streampark-flink/streampark-flink-proxy/src/main/scala/org/apache/streampark/flink/proxy/FlinkShimsProxy.scala
index 5dd25ec5f..34ccb9f59 100644
---
a/streampark-flink/streampark-flink-proxy/src/main/scala/org/apache/streampark/flink/proxy/FlinkShimsProxy.scala
+++
b/streampark-flink/streampark-flink-proxy/src/main/scala/org/apache/streampark/flink/proxy/FlinkShimsProxy.scala
@@ -20,7 +20,7 @@ package org.apache.streampark.flink.proxy
import org.apache.streampark.common.Constant
import org.apache.streampark.common.conf.{ConfigKeys, FlinkVersion}
import org.apache.streampark.common.util.{ChildFirstClassLoader,
ClassLoaderObjectInputStream, ClassLoaderUtils, Logger}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import java.io.{ByteArrayInputStream, ByteArrayOutputStream, File,
ObjectOutputStream}
import java.net.URL
diff --git
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamTableTrait.scala
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamTableTrait.scala
index 4fc75d616..9b3a804c3 100644
---
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamTableTrait.scala
+++
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamTableTrait.scala
@@ -17,6 +17,7 @@
package org.apache.streampark.flink.core
+import org.apache.streampark.common.util.Implicits.JavaList
import org.apache.streampark.common.util.Utils
import org.apache.streampark.flink.core.EnhancerImplicit._
@@ -47,7 +48,7 @@ import org.apache.flink.table.module.Module
import org.apache.flink.table.types.AbstractDataType
import org.apache.flink.util.SplittableIterator
-import java.util.{List => JavaList, Optional}
+import java.util.Optional
/**
* Integration api of stream and table
diff --git
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamingInitializer.scala
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamingInitializer.scala
index 2c8fa5f86..7ce7a85a0 100644
---
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamingInitializer.scala
+++
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamingInitializer.scala
@@ -21,6 +21,7 @@ import org.apache.streampark.common.conf.ConfigKeys._
import org.apache.streampark.common.enums.ApiType
import org.apache.streampark.common.enums.ApiType.ApiType
import org.apache.streampark.common.util._
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.core.conf.FlinkConfiguration
import collection.{mutable, Map}
@@ -32,8 +33,6 @@ import org.apache.flink.table.api.TableConfig
import java.io.File
-import scala.collection.convert.ImplicitConversions._
-
private[flink] object FlinkStreamingInitializer {
def initialize(args: Array[String], config: (StreamExecutionEnvironment,
ParameterTool) => Unit): (ParameterTool, StreamExecutionEnvironment) = {
diff --git
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkTableInitializer.scala
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkTableInitializer.scala
index 71ac0a523..7cb463ed7 100644
---
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkTableInitializer.scala
+++
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkTableInitializer.scala
@@ -21,6 +21,7 @@ import org.apache.streampark.common.conf.ConfigKeys._
import org.apache.streampark.common.enums.{ApiType, PlannerType}
import org.apache.streampark.common.enums.ApiType.ApiType
import org.apache.streampark.common.util.{DeflaterUtils, PropertiesUtils}
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.core.EnhancerImplicit._
import org.apache.streampark.flink.core.conf.FlinkConfiguration
@@ -33,7 +34,6 @@ import
org.apache.flink.table.api.bridge.scala.StreamTableEnvironment
import java.io.File
import scala.collection.{mutable, Map}
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Failure, Success, Try}
private[flink] object FlinkTableInitializer {
diff --git
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/conf/FlinkRunOption.scala
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/conf/FlinkRunOption.scala
index e0c30cb06..6272a018d 100644
---
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/conf/FlinkRunOption.scala
+++
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/conf/FlinkRunOption.scala
@@ -17,11 +17,12 @@
package org.apache.streampark.flink.core.conf
+import org.apache.streampark.common.util.Implicits._
+
import org.apache.commons.cli.{CommandLine, DefaultParser, Option, Options}
import java.lang.{Boolean => JavaBoolean}
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Failure, Success, Try}
/** Applies to all optional parameters under flink run */
diff --git
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
index f5542df84..b2c56a264 100644
---
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
+++
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
@@ -17,6 +17,8 @@
package org.apache.streampark.flink.core
+import org.apache.streampark.common.util.Implicits.JavaList
+
import org.apache.flink.api.java.utils.ParameterTool
import org.apache.flink.streaming.api.scala.{DataStream,
StreamExecutionEnvironment}
import org.apache.flink.table.api._
@@ -27,8 +29,6 @@ import org.apache.flink.table.resource.ResourceUri
import org.apache.flink.table.types.AbstractDataType
import org.apache.flink.types.Row
-import java.util.{List => JavaList}
-
class StreamTableContext(
override val parameter: ParameterTool,
private val streamEnv: StreamExecutionEnvironment,
diff --git
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
index 6b9d0f03b..e7fe52592 100644
---
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
+++
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
@@ -17,13 +17,13 @@
package org.apache.streampark.flink.core
+import org.apache.streampark.common.util.Implicits.JavaList
+
import org.apache.flink.api.java.utils.ParameterTool
import org.apache.flink.table.api._
import org.apache.flink.table.module.ModuleEntry
import org.apache.flink.table.resource.ResourceUri
-import java.util.{List => JavaList}
-
class TableContext(override val parameter: ParameterTool, private val
tableEnv: TableEnvironment)
extends FlinkTableTrait(parameter, tableEnv) {
diff --git
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
index f5db09054..8d655abec 100644
---
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
+++
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
@@ -17,6 +17,8 @@
package org.apache.streampark.flink.core
+import org.apache.streampark.common.util.Implicits.JavaList
+
import org.apache.flink.api.java.utils.ParameterTool
import org.apache.flink.streaming.api.scala.{DataStream,
StreamExecutionEnvironment}
import org.apache.flink.table.api._
@@ -28,8 +30,6 @@ import org.apache.flink.table.resource.ResourceUri
import org.apache.flink.table.types.AbstractDataType
import org.apache.flink.types.Row
-import java.util.{List => JavaList}
-
class StreamTableContext(
override val parameter: ParameterTool,
private val streamEnv: StreamExecutionEnvironment,
diff --git
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
index c1b348ae8..9d295b6b7 100644
---
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
+++
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
@@ -17,14 +17,14 @@
package org.apache.streampark.flink.core
+import org.apache.streampark.common.util.Implicits.JavaList
+
import org.apache.flink.api.java.utils.ParameterTool
import org.apache.flink.table.api._
import org.apache.flink.table.catalog.CatalogDescriptor
import org.apache.flink.table.module.ModuleEntry
import org.apache.flink.table.resource.ResourceUri
-import java.util.{List => JavaList}
-
class TableContext(override val parameter: ParameterTool, private val
tableEnv: TableEnvironment)
extends FlinkTableTrait(parameter, tableEnv) {
diff --git
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
index f5db09054..8d655abec 100644
---
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
+++
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
@@ -17,6 +17,8 @@
package org.apache.streampark.flink.core
+import org.apache.streampark.common.util.Implicits.JavaList
+
import org.apache.flink.api.java.utils.ParameterTool
import org.apache.flink.streaming.api.scala.{DataStream,
StreamExecutionEnvironment}
import org.apache.flink.table.api._
@@ -28,8 +30,6 @@ import org.apache.flink.table.resource.ResourceUri
import org.apache.flink.table.types.AbstractDataType
import org.apache.flink.types.Row
-import java.util.{List => JavaList}
-
class StreamTableContext(
override val parameter: ParameterTool,
private val streamEnv: StreamExecutionEnvironment,
diff --git
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
index c1b348ae8..9d295b6b7 100644
---
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
+++
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
@@ -17,14 +17,14 @@
package org.apache.streampark.flink.core
+import org.apache.streampark.common.util.Implicits.JavaList
+
import org.apache.flink.api.java.utils.ParameterTool
import org.apache.flink.table.api._
import org.apache.flink.table.catalog.CatalogDescriptor
import org.apache.flink.table.module.ModuleEntry
import org.apache.flink.table.resource.ResourceUri
-import java.util.{List => JavaList}
-
class TableContext(override val parameter: ParameterTool, private val
tableEnv: TableEnvironment)
extends FlinkTableTrait(parameter, tableEnv) {
diff --git
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/StopRequest.scala
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/StopRequest.scala
index 2f8b5fcbb..1adcc862e 100644
---
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/StopRequest.scala
+++
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/StopRequest.scala
@@ -19,11 +19,10 @@ package org.apache.streampark.spark.client.bean
import org.apache.streampark.common.conf.SparkVersion
import org.apache.streampark.common.enums.SparkExecutionMode
+import org.apache.streampark.common.util.Implicits.JavaMap
import javax.annotation.Nullable
-import java.util.{Map => JavaMap}
-
case class StopRequest(
id: Long,
sparkVersion: SparkVersion,
diff --git
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitRequest.scala
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitRequest.scala
index cf71c99b9..4079f47e4 100644
---
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitRequest.scala
+++
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitRequest.scala
@@ -22,6 +22,7 @@ import org.apache.streampark.common.conf.{SparkVersion,
Workspace}
import org.apache.streampark.common.conf.ConfigKeys._
import org.apache.streampark.common.enums._
import org.apache.streampark.common.util.{DeflaterUtils, HdfsUtils,
PropertiesUtils}
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.flink.packer.pipeline.{BuildResult,
ShadedBuildResponse}
import com.fasterxml.jackson.databind.ObjectMapper
@@ -31,9 +32,6 @@ import javax.annotation.Nullable
import java.io.{File, IOException}
import java.nio.file.Files
-import java.util.{Map => JavaMap}
-
-import scala.collection.convert.ImplicitConversions._
case class SubmitRequest(
sparkVersion: SparkVersion,
diff --git
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitResponse.scala
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitResponse.scala
index b97b1f146..da5720a2d 100644
---
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitResponse.scala
+++
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitResponse.scala
@@ -17,7 +17,7 @@
package org.apache.streampark.spark.client.bean
-import java.util.{Map => JavaMap}
+import org.apache.streampark.common.util.Implicits.JavaMap
case class SubmitResponse(
clusterId: String,
diff --git
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/proxy/SparkShimsProxy.scala
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/proxy/SparkShimsProxy.scala
index 97e749e2b..873a5b531 100644
---
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/proxy/SparkShimsProxy.scala
+++
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/proxy/SparkShimsProxy.scala
@@ -20,7 +20,7 @@ package org.apache.streampark.spark.client.proxy
import org.apache.streampark.common.Constant
import org.apache.streampark.common.conf.{ConfigKeys, SparkVersion}
import org.apache.streampark.common.util.{ChildFirstClassLoader,
ClassLoaderObjectInputStream, ClassLoaderUtils, Logger}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
import java.io.{ByteArrayInputStream, ByteArrayOutputStream, File,
ObjectOutputStream}
import java.net.URL
diff --git
a/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/impl/YarnClient.scala
b/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/impl/YarnClient.scala
index 8248a04cb..19e6beaf7 100644
---
a/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/impl/YarnClient.scala
+++
b/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/impl/YarnClient.scala
@@ -19,6 +19,7 @@ package org.apache.streampark.spark.client.impl
import org.apache.streampark.common.enums.SparkExecutionMode
import org.apache.streampark.common.util.HadoopUtils
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.spark.client.`trait`.SparkClientTrait
import org.apache.streampark.spark.client.bean._
@@ -27,7 +28,6 @@ import org.apache.spark.launcher.{SparkAppHandle,
SparkLauncher}
import java.util.concurrent.CountDownLatch
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Failure, Success, Try}
/** yarn application mode submit */
diff --git
a/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/trait/SparkClientTrait.scala
b/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/trait/SparkClientTrait.scala
index bfefa8892..2ef438dc0 100644
---
a/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/trait/SparkClientTrait.scala
+++
b/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/trait/SparkClientTrait.scala
@@ -18,9 +18,9 @@
package org.apache.streampark.spark.client.`trait`
import org.apache.streampark.common.util._
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.spark.client.bean._
-import scala.collection.convert.ImplicitConversions._
import scala.util.{Failure, Success, Try}
trait SparkClientTrait extends Logger {
diff --git
a/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/HBaseOffset.scala
b/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/HBaseOffset.scala
index a3451cbe8..de7ca0a25 100644
---
a/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/HBaseOffset.scala
+++
b/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/HBaseOffset.scala
@@ -18,6 +18,7 @@
package org.apache.streampark.spark.connector.kafka.offset
import org.apache.streampark.common.util.HBaseClient
+import org.apache.streampark.common.util.Implicits._
import org.apache.hadoop.hbase._
import org.apache.hadoop.hbase.client.{Delete, Put, Scan, Table}
@@ -28,7 +29,6 @@ import org.apache.spark.SparkConf
import java.util
-import scala.collection.convert.ImplicitConversions._
import scala.collection.mutable
private[kafka] class HBaseOffset(val sparkConf: SparkConf) extends Offset {
diff --git
a/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/Offset.scala
b/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/Offset.scala
index 6b6048038..eb8f87b57 100644
---
a/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/Offset.scala
+++
b/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/Offset.scala
@@ -17,6 +17,7 @@
package org.apache.streampark.spark.connector.kafka.offset
+import org.apache.streampark.common.util.Implicits._
import org.apache.streampark.common.util.Logger
import org.apache.kafka.common.TopicPartition
@@ -24,7 +25,6 @@ import org.apache.spark.SparkConf
import java.util.Properties
-import scala.collection.JavaConverters._
import scala.language.implicitConversions
import scala.util.Try
diff --git
a/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/RedisOffset.scala
b/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/RedisOffset.scala
index 077cfba31..c512fc06e 100644
---
a/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/RedisOffset.scala
+++
b/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/RedisOffset.scala
@@ -18,12 +18,12 @@
package org.apache.streampark.spark.connector.kafka.offset
import org.apache.streampark.common.util.{RedisEndpoint, RedisUtils}
+import org.apache.streampark.common.util.Implicits._
import org.apache.kafka.common.TopicPartition
import org.apache.spark.SparkConf
import redis.clients.jedis.Protocol
-import scala.collection.convert.ImplicitConversions._
import scala.util.Try
/** Redis Offset Manager */
diff --git a/tools/dependencies/check-LICENSE.sh
b/tools/dependencies/check-LICENSE.sh
index 58ca164db..f56aebd1d 100755
--- a/tools/dependencies/check-LICENSE.sh
+++ b/tools/dependencies/check-LICENSE.sh
@@ -19,9 +19,8 @@
check_path=dist-license-check
-if [ -d "$check_path" ];then
-rm -rf $check_path
-fi
+[[ -d "$check_path" ]] && rm -rf $check_path
+
mkdir $check_path || true
tar -zxf dist/apache-streampark*-bin.tar.gz --strip=1 -C $check_path
@@ -45,7 +44,7 @@ echo '=== Third party dependencies: ' && grep -vf
$check_path/self-modules.txt $
diff -w -B -U0 <(sort < tools/dependencies/known-dependencies.txt) <(sort <
$check_path/third-party-dependencies.txt)
-if [ $? -ne 0 ]; then
+if [[ $? -ne 0 ]]; then
echo "Third-party dependencies are not all known, please add the license to
LICENSE file and add the dependency to
tools/dependencies/known-dependencies.txt"
exit 1
fi