This is an automated email from the ASF dual-hosted git repository.

benjobs pushed a commit to branch dev
in repository https://gitbox.apache.org/repos/asf/incubator-streampark.git


The following commit(s) were added to refs/heads/dev by this push:
     new 9998ca604 [Improve] scala implicits some improvements (#3902)
9998ca604 is described below

commit 9998ca6042eea0dc3af178bf7794a6f76b2880e4
Author: benjobs <[email protected]>
AuthorDate: Sun Jul 21 00:43:15 2024 +0800

    [Improve] scala implicits some improvements (#3902)
    
    * [Improve] scala implicits some improvements
    * [Improve] check-LICENSE.sh minor improvement
    * [Improve] FlinkVersion and SparkVersion improvements
---
 .../streampark/common/conf/ConfigOption.scala      |  2 +-
 .../streampark/common/conf/FlinkVersion.scala      |  8 +--
 .../common/conf/InternalConfigHolder.scala         |  6 +--
 .../streampark/common/conf/SparkVersion.scala      | 60 +++++++++-------------
 .../apache/streampark/common/conf/Workspace.scala  |  2 +-
 .../streampark/common/util/AssertUtils.scala       |  3 +-
 .../streampark/common/util/CURLBuilder.scala       |  4 +-
 .../streampark/common/util/CommandUtils.scala      |  3 +-
 .../streampark/common/util/ConfigUtils.scala       |  4 +-
 .../apache/streampark/common/util/FileUtils.scala  |  5 +-
 .../streampark/common/util/HBaseClient.scala       |  3 +-
 .../streampark/common/util/HadoopConfigUtils.scala |  5 +-
 .../streampark/common/util/HadoopUtils.scala       |  2 +-
 .../apache/streampark/common/util/HostsUtils.scala |  4 +-
 .../streampark/common/util/HttpClientUtils.scala   |  5 +-
 .../util/{ImplicitsUtils.scala => Implicits.scala} | 42 ++++++++++++---
 .../apache/streampark/common/util/JdbcUtils.scala  |  2 +-
 .../streampark/common/util/MongoConfig.scala       |  3 +-
 .../streampark/common/util/PropertiesUtils.scala   | 33 ++++++------
 .../streampark/common/util/RedisClient.scala       |  2 +-
 .../streampark/common/util/RedisEndpoint.scala     |  3 +-
 .../apache/streampark/common/util/RedisUtils.scala |  4 +-
 .../streampark/common/util/SqlConvertUtils.scala   |  3 +-
 .../org/apache/streampark/common/util/Utils.scala  |  6 +--
 .../apache/streampark/common/util/YarnUtils.scala  |  3 +-
 .../streampark/common/util/ZooKeeperUtils.scala    |  3 +-
 ...mplicitsUtilsTest.scala => ImplicitsTest.scala} |  4 +-
 .../console/core/runner/QuickStartRunner.java      |  4 +-
 .../flink/client/bean/CancelRequest.scala          |  3 +-
 .../flink/client/bean/DeployRequest.scala          |  2 +-
 .../flink/client/bean/DeployRequestTrait.scala     |  3 +-
 .../flink/client/bean/SavepointRequestTrait.scala  |  3 +-
 .../flink/client/bean/ShutDownRequest.scala        |  3 +-
 .../flink/client/bean/SubmitRequest.scala          |  3 +-
 .../flink/client/bean/SubmitResponse.scala         |  4 +-
 .../client/bean/TriggerSavepointRequest.scala      |  3 +-
 .../impl/KubernetesNativeApplicationClient.scala   |  2 -
 .../impl/KubernetesNativeSessionClient.scala       |  4 +-
 .../flink/client/impl/YarnApplicationClient.scala  |  2 +-
 .../flink/client/impl/YarnPerJobClient.scala       |  3 +-
 .../flink/client/impl/YarnSessionClient.scala      |  3 +-
 .../flink/client/trait/FlinkClientTrait.scala      |  4 +-
 .../flink/client/trait/YarnClientTrait.scala       |  2 +-
 .../flink/client/test/YarnPerJobTestCase.scala     |  4 +-
 .../flink/connector/conf/ThresholdConf.scala       |  3 +-
 .../flink/connector/failover/FailoverChecker.scala |  2 +-
 .../flink/connector/failover/FailoverWriter.scala  |  2 +-
 .../flink/connector/failover/SinkBuffer.scala      |  2 +-
 .../flink/connector/failover/SinkRequestTest.scala |  2 +-
 .../clickhouse/conf/ClickHouseHttpConfig.scala     |  2 +-
 .../conf/ClickHouseSinkConfigOption.scala          |  2 +-
 .../internal/ClickHouseSinkFunction.scala          |  2 +-
 .../clickhouse/internal/ClickHouseSinkWriter.scala |  2 +-
 .../clickhouse/internal/ClickHouseWriterTask.scala |  2 +-
 .../connector/doris/conf/DorisConfig.scala         |  2 +-
 .../doris/conf/DorisSinkConfigOption.scala         |  2 +-
 .../elasticsearch5/conf/ESSinkConfigOption.scala   |  5 +-
 .../connector/elasticsearch5/sink/ES5Sink.scala    |  4 +-
 .../elasticsearch6/conf/ES6SinkConfigOption.scala  |  4 +-
 .../connector/elasticsearch6/sink/ES6Sink.scala    |  2 +-
 .../elasticsearch7/conf/ES7SinkConfigOption.scala  |  4 +-
 .../connector/elasticsearch7/sink/ES7Sink.scala    |  2 +-
 .../hbase/internal/HBaseSinkFunction.scala         |  2 +-
 .../hbase/internal/HBaseSourceFunction.scala       |  2 +-
 .../connector/hbase/request/HBaseRequest.scala     |  2 +-
 .../connector/http/internal/HttpSinkWriter.scala   |  2 +-
 .../connector/http/internal/HttpWriterTask.scala   |  2 +-
 .../flink/connector/http/sink/HttpSink.scala       |  2 +-
 .../connector/influx/function/InfluxFunction.scala |  4 +-
 .../jdbc/internal/JdbcSourceFunction.scala         |  2 -
 .../flink/connector/kafka/source/KafkaSource.scala |  2 +-
 .../mongo/internal/MongoSourceFunction.scala       |  2 +-
 .../redis/conf/RedisSinkConfigOption.scala         |  2 +-
 .../flink/connector/redis/sink/RedisSink.scala     |  2 +-
 .../flink/kubernetes/FlinkK8sWatchController.scala |  3 +-
 .../flink/kubernetes/KubernetesRetriever.scala     |  3 +-
 .../flink/kubernetes/PodTemplateParser.scala       |  4 +-
 .../helper/KubernetesDeploymentHelper.scala        |  3 +-
 .../kubernetes/ingress/IngressController.scala     |  2 +-
 .../kubernetes/ingress/IngressStrategyV1.scala     |  3 +-
 .../ingress/IngressStrategyV1beta1.scala           |  3 +-
 .../kubernetes/watcher/FlinkJobStatusWatcher.scala |  2 +-
 .../flink/kubernetes/FlinkRestJsonTest.scala       |  2 +-
 .../flink/kubernetes/PodTemplateParserTest.scala   |  4 +-
 .../flink/packer/maven/DependencyInfo.scala        |  4 +-
 .../streampark/flink/packer/maven/MavenTool.scala  |  3 +-
 .../flink/packer/pipeline/BuildPipeline.scala      |  2 +-
 .../packer/pipeline/DockerResolveSnapshot.scala    |  5 +-
 .../flink/packer/pipeline/PipelineSnapshot.scala   |  4 +-
 .../pipeline/impl/FlinkRemoteBuildPipeline.scala   |  3 +-
 .../impl/FlinkYarnApplicationBuildPipeline.scala   |  4 +-
 .../impl/SparkYarnApplicationBuildPipeline.scala   |  4 +-
 .../streampark/flink/packer/MavenToolSpec.scala    |  3 +-
 .../streampark/flink/proxy/FlinkShimsProxy.scala   |  2 +-
 .../flink/core/FlinkStreamTableTrait.scala         |  3 +-
 .../flink/core/FlinkStreamingInitializer.scala     |  3 +-
 .../flink/core/FlinkTableInitializer.scala         |  2 +-
 .../flink/core/conf/FlinkRunOption.scala           |  3 +-
 .../streampark/flink/core/StreamTableContext.scala |  4 +-
 .../streampark/flink/core/TableContext.scala       |  4 +-
 .../streampark/flink/core/StreamTableContext.scala |  4 +-
 .../streampark/flink/core/TableContext.scala       |  4 +-
 .../streampark/flink/core/StreamTableContext.scala |  4 +-
 .../streampark/flink/core/TableContext.scala       |  4 +-
 .../streampark/spark/client/bean/StopRequest.scala |  3 +-
 .../spark/client/bean/SubmitRequest.scala          |  4 +-
 .../spark/client/bean/SubmitResponse.scala         |  2 +-
 .../spark/client/proxy/SparkShimsProxy.scala       |  2 +-
 .../streampark/spark/client/impl/YarnClient.scala  |  2 +-
 .../spark/client/trait/SparkClientTrait.scala      |  2 +-
 .../spark/connector/kafka/offset/HBaseOffset.scala |  2 +-
 .../spark/connector/kafka/offset/Offset.scala      |  2 +-
 .../spark/connector/kafka/offset/RedisOffset.scala |  2 +-
 tools/dependencies/check-LICENSE.sh                |  7 ++-
 114 files changed, 221 insertions(+), 253 deletions(-)

diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/ConfigOption.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/ConfigOption.scala
index 4fefa0d8b..0ec3c84b9 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/ConfigOption.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/ConfigOption.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.common.conf
 
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 
 import java.util.Properties
 
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/FlinkVersion.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/FlinkVersion.scala
index 6ba10923c..e4950e247 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/FlinkVersion.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/FlinkVersion.scala
@@ -18,17 +18,17 @@
 package org.apache.streampark.common.conf
 
 import org.apache.streampark.common.util.{CommandUtils, Logger}
+import org.apache.streampark.common.util.Implicits._
 
 import java.io.File
-import java.net.{URL => NetURL}
+import java.net.URL
 import java.util.function.Consumer
 import java.util.regex.Pattern
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.mutable
 
 /** @param flinkHome actual flink home that must be a readable local path */
-class FlinkVersion(val flinkHome: String) extends java.io.Serializable with 
Logger {
+class FlinkVersion(val flinkHome: String) extends Serializable with Logger {
 
   private[this] lazy val FLINK_VER_PATTERN = 
Pattern.compile("^(\\d+\\.\\d+)(\\.)?.*$")
 
@@ -63,7 +63,7 @@ class FlinkVersion(val flinkHome: String) extends 
java.io.Serializable with Logg
     lib
   }
 
-  lazy val flinkLibs: List[NetURL] = 
flinkLib.listFiles().map(_.toURI.toURL).toList
+  lazy val flinkLibs: List[URL] = 
flinkLib.listFiles().map(_.toURI.toURL).toList
 
   lazy val version: String = {
     val cmd = List(
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/InternalConfigHolder.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/InternalConfigHolder.scala
index 7816ecd20..3d19d6f12 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/InternalConfigHolder.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/InternalConfigHolder.scala
@@ -19,15 +19,13 @@ package org.apache.streampark.common.conf
 
 import org.apache.streampark.common.Constant
 import org.apache.streampark.common.util.{Logger, SystemPropertyUtils}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 
 import javax.annotation.{Nonnull, Nullable}
 
 import java.util
 import java.util.concurrent.ConcurrentHashMap
 
-import scala.collection.convert.ImplicitConversions._
-
 /**
  * Thread-safe configuration storage containers. All configurations will be 
automatically
  * initialized from the spring configuration items of the same name.
@@ -123,7 +121,7 @@ object InternalConfigHolder extends Logger {
 
   /** Get keys of all registered ConfigOption. */
   @Nonnull
-  def keys(): util.Set[String] = {
+  def keys(): JavaSet[String] = {
     val map = new util.HashMap[String, InternalOption](confOptions.size())
     map.putAll(confOptions)
     map.keySet()
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/SparkVersion.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/SparkVersion.scala
index 725d69463..c6ba76888 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/SparkVersion.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/SparkVersion.scala
@@ -18,35 +18,37 @@
 package org.apache.streampark.common.conf
 
 import org.apache.streampark.common.util.{CommandUtils, Logger}
-
-import org.apache.commons.lang3.StringUtils
+import org.apache.streampark.common.util.Implicits._
 
 import java.io.File
-import java.net.{URL => NetURL}
-import java.util.concurrent.atomic.{AtomicBoolean, AtomicReference}
+import java.net.URL
 import java.util.function.Consumer
 import java.util.regex.Pattern
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.mutable
 
 /** @param sparkHome actual spark home that must be a readable local path */
-class SparkVersion(val sparkHome: String) extends java.io.Serializable with 
Logger {
+class SparkVersion(val sparkHome: String) extends Serializable with Logger {
 
   private[this] lazy val SPARK_VER_PATTERN = 
Pattern.compile("^(\\d+\\.\\d+)(\\.)?.*$")
 
   private[this] lazy val SPARK_VERSION_PATTERN = Pattern.compile("(version) 
(\\d+\\.\\d+\\.\\d+)")
 
-  private[this] lazy val SPARK_SCALA_VERSION_PATTERN =
-    Pattern.compile("^spark-core_(.*)-[0-9].*.jar$")
+  private[this] lazy val SPARK_SCALA_VERSION_PATTERN = 
Pattern.compile("^spark-core_(.*)-[0-9].*.jar$")
 
-  lazy val scalaVersion: String = {
-    val matcher = SPARK_SCALA_VERSION_PATTERN.matcher(sparkCoreJar.getName)
-    if (matcher.matches()) {
-      matcher.group(1);
-    } else {
-      "2.12"
+  lazy val scalaVersion: String = 
SPARK_SCALA_VERSION_PATTERN.matcher(sparkCoreJar.getName).group(1)
+
+  lazy val sparkCoreJar: File = {
+    val distJar = 
sparkLib.listFiles().filter(_.getName.matches("spark-core.*\\.jar"))
+    distJar match {
+      case x if x.isEmpty =>
+        throw new IllegalArgumentException(s"[StreamPark] can no found 
spark-core jar in $sparkLib")
+      case x if x.length > 1 =>
+        throw new IllegalArgumentException(
+          s"[StreamPark] found multiple spark-core jar in $sparkLib")
+      case _ =>
     }
+    distJar.head
   }
 
   def checkVersion(throwException: Boolean = true): Boolean = {
@@ -54,7 +56,7 @@ class SparkVersion(val sparkHome: String) extends 
java.io.Serializable with Logg
       case Array(3, v, _) if v >= 1 && v <= 3 => true
       case _ =>
         if (throwException) {
-          throw new UnsupportedOperationException(s"Unsupported flink version: 
$version")
+          throw new UnsupportedOperationException(s"Unsupported spark version: 
$version")
         } else {
           false
         }
@@ -73,7 +75,7 @@ class SparkVersion(val sparkHome: String) extends 
java.io.Serializable with Logg
     lib
   }
 
-  lazy val sparkLibs: List[NetURL] = 
sparkLib.listFiles().map(_.toURI.toURL).toList
+  lazy val sparkLibs: List[URL] = 
sparkLib.listFiles().map(_.toURI.toURL).toList
 
   lazy val majorVersion: String = {
     if (version == null) {
@@ -86,9 +88,8 @@ class SparkVersion(val sparkHome: String) extends 
java.io.Serializable with Logg
   }
 
   lazy val version: String = {
-    val sparkVersion = new AtomicReference[String]
+    var sparkVersion: String = null
     val cmd = List(s"$sparkHome/bin/spark-submit --version")
-    val success = new AtomicBoolean(false)
     val buffer = new mutable.StringBuilder
     CommandUtils.execute(
       sparkHome,
@@ -97,31 +98,18 @@ class SparkVersion(val sparkHome: String) extends 
java.io.Serializable with Logg
         override def accept(out: String): Unit = {
           buffer.append(out).append("\n")
           val matcher = SPARK_VERSION_PATTERN.matcher(out)
-          if (matcher.find && StringUtils.isBlank(sparkVersion.get())) {
-            success.set(true)
-            sparkVersion.set(matcher.group(2))
+          if (matcher.find) {
+            sparkVersion = matcher.group(2)
           }
         }
       })
+
     logInfo(buffer.toString())
-    if (!success.get()) {
+    if (sparkVersion == null) {
       throw new IllegalStateException(s"[StreamPark] parse spark version 
failed. $buffer")
     }
     buffer.clear()
-    sparkVersion.get
-  }
-
-  lazy val sparkCoreJar: File = {
-    val distJar = 
sparkLib.listFiles().filter(_.getName.matches("spark-core.*\\.jar"))
-    distJar match {
-      case x if x.isEmpty =>
-        throw new IllegalArgumentException(s"[StreamPark] can no found 
spark-core jar in $sparkLib")
-      case x if x.length > 1 =>
-        throw new IllegalArgumentException(
-          s"[StreamPark] found multiple spark-core jar in $sparkLib")
-      case _ =>
-    }
-    distJar.head
+    sparkVersion
   }
 
   override def toString: String =
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/Workspace.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/Workspace.scala
index f002d8d28..cc9161f5e 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/conf/Workspace.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/conf/Workspace.scala
@@ -19,7 +19,7 @@ package org.apache.streampark.common.conf
 
 import org.apache.streampark.common.enums.StorageType
 import org.apache.streampark.common.util.{HdfsUtils, SystemPropertyUtils}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 
 import java.net.URI
 
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/AssertUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/AssertUtils.scala
index 96cfda60e..00934cad7 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/AssertUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/AssertUtils.scala
@@ -17,14 +17,13 @@
 
 package org.apache.streampark.common.util
 
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.common.util.Utils.isEmpty
 
 import javax.annotation.Nullable
 
 import java.util
 
-import scala.collection.convert.ImplicitConversions._
-
 /** @since 2.2.0 */
 object AssertUtils {
 
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/CURLBuilder.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/CURLBuilder.scala
index 9699935c4..c18ef1896 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/CURLBuilder.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/CURLBuilder.scala
@@ -17,9 +17,9 @@
 
 package org.apache.streampark.common.util
 
-import java.util
+import org.apache.streampark.common.util.Implicits._
 
-import scala.collection.convert.ImplicitConversions._
+import java.util
 
 class CURLBuilder(val url: String) {
 
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/CommandUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/CommandUtils.scala
index a86df3cd0..b5ccf677a 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/CommandUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/CommandUtils.scala
@@ -17,12 +17,13 @@
 
 package org.apache.streampark.common.util
 
+import org.apache.streampark.common.util.Implicits._
+
 import java.io._
 import java.lang.{Iterable => JavaIterable}
 import java.util.Scanner
 import java.util.function.Consumer
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Failure, Success, Try}
 
 object CommandUtils extends Logger {
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/ConfigUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/ConfigUtils.scala
index 617670392..28c182110 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/ConfigUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/ConfigUtils.scala
@@ -19,10 +19,10 @@ package org.apache.streampark.common.util
 
 import org.apache.streampark.common.Constant
 import org.apache.streampark.common.conf.ConfigKeys._
+import org.apache.streampark.common.util.Implicits._
 
-import java.util.{Map => JavaMap, Properties}
+import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.immutable.{Map => ScalaMap}
 import scala.util.Try
 
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/FileUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/FileUtils.scala
index 0f1996e30..fbeb6f412 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/FileUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/FileUtils.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.common.util
 
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 
 import java.io._
 import java.net.URL
@@ -29,7 +29,6 @@ import java.util
 import java.util.Scanner
 import java.util.stream.Collectors
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.mutable
 
 object FileUtils {
@@ -114,7 +113,7 @@ object FileUtils {
     filename.drop(filename.lastIndexOf("."))
   }
 
-  def listFileAsURL(dirPath: String): util.List[URL] = {
+  def listFileAsURL(dirPath: String): JavaList[URL] = {
     new File(dirPath) match {
       case x if x.exists() && x.isDirectory =>
         val files = x.listFiles()
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HBaseClient.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HBaseClient.scala
index d5f0b5fdf..fe02a2e1f 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HBaseClient.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HBaseClient.scala
@@ -18,6 +18,7 @@
 package org.apache.streampark.common.util
 
 import org.apache.streampark.common.conf.ConfigKeys._
+import org.apache.streampark.common.util.Implicits._
 
 import org.apache.hadoop.conf.Configuration
 import org.apache.hadoop.hbase.{HBaseConfiguration, TableName}
@@ -26,8 +27,6 @@ import org.apache.hadoop.security.UserGroupInformation
 
 import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
-
 class HBaseClient(func: () => Connection) extends Serializable {
   lazy val connection: Connection = func()
 
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopConfigUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopConfigUtils.scala
index f88a09c8a..c33687e70 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopConfigUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopConfigUtils.scala
@@ -20,15 +20,14 @@ package org.apache.streampark.common.util
 import org.apache.streampark.common.conf.{CommonConfig, InternalConfigHolder}
 import org.apache.streampark.common.conf.ConfigKeys._
 import org.apache.streampark.common.fs.LfsOperator
+import org.apache.streampark.common.util.Implicits._
 
 import org.apache.commons.io.{FileUtils => ApacheFileUtils}
 
 import java.io.File
 import java.nio.charset.StandardCharsets
-import java.util.{Collections, Map => JavaMap, Optional}
+import java.util.{Collections, Optional}
 
-import scala.collection.JavaConverters._
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.immutable.ListMap
 import scala.util.{Failure, Success, Try}
 
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopUtils.scala
index f659303ac..8ed2b807d 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HadoopUtils.scala
@@ -19,6 +19,7 @@ package org.apache.streampark.common.util
 
 import org.apache.streampark.common.conf.{CommonConfig, InternalConfigHolder}
 import org.apache.streampark.common.conf.ConfigKeys._
+import org.apache.streampark.common.util.Implicits._
 
 import org.apache.commons.collections.CollectionUtils
 import org.apache.commons.lang3.StringUtils
@@ -39,7 +40,6 @@ import java.util
 import java.util.{Timer, TimerTask}
 import java.util.concurrent._
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Failure, Success, Try}
 
 object HadoopUtils extends Logger {
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HostsUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HostsUtils.scala
index 08686a283..19aa757ca 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HostsUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HostsUtils.scala
@@ -17,12 +17,12 @@
 
 package org.apache.streampark.common.util
 
+import org.apache.streampark.common.util.Implicits._
+
 import io.netty.resolver.HostsFileParser
 
 import java.net.InetAddress
-import java.util.{Map => JavaMap}
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.immutable.ListMap
 
 object HostsUtils {
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HttpClientUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HttpClientUtils.scala
index c0aa5241c..4a228d238 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/HttpClientUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/HttpClientUtils.scala
@@ -17,6 +17,8 @@
 
 package org.apache.streampark.common.util
 
+import org.apache.streampark.common.util.Implicits._
+
 import org.apache.hc.client5.http.auth.{AuthSchemeFactory, AuthScope, 
Credentials, StandardAuthScheme}
 import org.apache.hc.client5.http.classic.methods.{HttpGet, HttpPost, 
HttpUriRequestBase}
 import org.apache.hc.client5.http.config.RequestConfig
@@ -32,9 +34,6 @@ import org.apache.hc.core5.net.URIBuilder
 
 import java.nio.charset.{Charset, StandardCharsets}
 import java.security.Principal
-import java.util.{List => JavaList, Map => JavaMap}
-
-import scala.collection.convert.ImplicitConversions._
 
 object HttpClientUtils {
 
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/ImplicitsUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/Implicits.scala
similarity index 66%
rename from 
streampark-common/src/main/scala/org/apache/streampark/common/util/ImplicitsUtils.scala
rename to 
streampark-common/src/main/scala/org/apache/streampark/common/util/Implicits.scala
index 4dd147dc9..40db96ae5 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/ImplicitsUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/Implicits.scala
@@ -19,11 +19,40 @@ package org.apache.streampark.common.util
 
 import org.apache.streampark.common.util.Utils.close
 
-import java.lang.{Boolean => JavaBool, Byte => JavaByte, Double => JavaDouble, 
Float => JavaFloat, Integer => JavaInt, Long => JavaLong, Short => JavaShort}
+import java.lang.{Boolean => JavaBool, Double => JavaDouble, Float => 
JavaFloat, Integer => JavaInt, Long => JavaLong, Short => JavaShort}
 
+import scala.collection.convert.{DecorateAsJava, DecorateAsScala, 
ToJavaImplicits, ToScalaImplicits}
 import scala.language.implicitConversions
 
-object ImplicitsUtils {
+object Implicits extends ToScalaImplicits with ToJavaImplicits with 
DecorateAsJava with DecorateAsScala {
+
+  type JavaMap[K, V] = java.util.Map[K, V]
+
+  type JavaHashMap[K, V] = java.util.HashMap[K, V]
+
+  type JavaLinkedMap[K, V] = java.util.LinkedHashMap[K, V]
+
+  type JavaCollection[T] = java.util.Collection[T]
+
+  type JavaList[T] = java.util.List[T]
+
+  type JavaArrayList[T] = java.util.ArrayList[T]
+
+  type JavaSet[T] = java.util.Set[T]
+
+  type JavaBool = java.lang.Boolean
+
+  type JavaByte = java.lang.Byte
+
+  type JavaDouble = java.lang.Double
+
+  type JavaFloat = java.lang.Float
+
+  type JavaInt = java.lang.Integer
+
+  type JavaLong = java.lang.Long
+
+  type JavaShort = java.lang.Short
 
   implicit class AutoCloseImplicits[T <: AutoCloseable](autoCloseable: T) {
 
@@ -53,12 +82,9 @@ object ImplicitsUtils {
         case c if c == classOf[JavaByte] => v.toByte.asInstanceOf[T]
         case c if c == classOf[JavaInt] => JavaInt.valueOf(v).asInstanceOf[T]
         case c if c == classOf[JavaLong] => JavaLong.valueOf(v).asInstanceOf[T]
-        case c if c == classOf[JavaFloat] =>
-          JavaFloat.valueOf(v).asInstanceOf[T]
-        case c if c == classOf[JavaDouble] =>
-          JavaDouble.valueOf(v).asInstanceOf[T]
-        case c if c == classOf[JavaShort] =>
-          JavaShort.valueOf(v).asInstanceOf[T]
+        case c if c == classOf[JavaFloat] => 
JavaFloat.valueOf(v).asInstanceOf[T]
+        case c if c == classOf[JavaDouble] => 
JavaDouble.valueOf(v).asInstanceOf[T]
+        case c if c == classOf[JavaShort] => 
JavaShort.valueOf(v).asInstanceOf[T]
         case c if c == classOf[JavaBool] => JavaBool.valueOf(v).asInstanceOf[T]
         case _ =>
           throw new IllegalArgumentException(s"Unsupported type: $classType")
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/JdbcUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/JdbcUtils.scala
index aba8e35e9..95071fcd5 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/JdbcUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/JdbcUtils.scala
@@ -18,6 +18,7 @@
 package org.apache.streampark.common.util
 
 import org.apache.streampark.common.conf.ConfigKeys._
+import org.apache.streampark.common.util.Implicits._
 
 import com.zaxxer.hikari.{HikariConfig, HikariDataSource}
 
@@ -26,7 +27,6 @@ import java.util.Properties
 import java.util.concurrent.ConcurrentHashMap
 import java.util.concurrent.locks.ReentrantLock
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.mutable
 import scala.collection.mutable.ArrayBuffer
 import scala.util.Try
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/MongoConfig.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/MongoConfig.scala
index 5c5094465..64d610490 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/MongoConfig.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/MongoConfig.scala
@@ -18,13 +18,12 @@
 package org.apache.streampark.common.util
 
 import org.apache.streampark.common.conf.ConfigKeys._
+import org.apache.streampark.common.util.Implicits._
 
 import com.mongodb._
 
 import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
-
 object MongoConfig {
 
   val client_uri = "client-uri"
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/PropertiesUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/PropertiesUtils.scala
index e95a7539a..c5153e986 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/PropertiesUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/PropertiesUtils.scala
@@ -17,6 +17,8 @@
 
 package org.apache.streampark.common.util
 
+import org.apache.streampark.common.util.Implicits._
+
 import com.typesafe.config.ConfigFactory
 import org.apache.commons.lang3.StringUtils
 import org.yaml.snakeyaml.Yaml
@@ -24,11 +26,10 @@ import org.yaml.snakeyaml.Yaml
 import javax.annotation.Nonnull
 
 import java.io._
-import java.util.{HashMap => JavaMap, LinkedHashMap => LinkedMap, Properties, 
Scanner}
+import java.util.{Properties, Scanner}
 import java.util.concurrent.atomic.AtomicInteger
 import java.util.regex.Pattern
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.mutable
 import scala.collection.mutable.{ArrayBuffer, Map => MutableMap}
 
@@ -59,7 +60,7 @@ object PropertiesUtils extends Logger {
       prefix: String = "",
       proper: MutableMap[String, String] = MutableMap[String, String]()): 
Map[String, String] = {
     v match {
-      case map: LinkedMap[_, _] =>
+      case map: JavaLinkedMap[_, _] =>
         map
           .flatMap(x => {
             prefix match {
@@ -203,31 +204,31 @@ object PropertiesUtils extends Logger {
   }
 
   def fromYamlTextAsJava(text: String): JavaMap[String, String] =
-    new JavaMap[String, String](fromYamlText(text))
+    new JavaHashMap[String, String](fromYamlText(text))
 
   def fromHoconTextAsJava(text: String): JavaMap[String, String] =
-    new JavaMap[String, String](fromHoconText(text))
+    new JavaHashMap[String, String](fromHoconText(text))
 
   def fromPropertiesTextAsJava(text: String): JavaMap[String, String] =
-    new JavaMap[String, String](fromPropertiesText(text))
+    new JavaHashMap[String, String](fromPropertiesText(text))
 
   def fromYamlFileAsJava(filename: String): JavaMap[String, String] =
-    new JavaMap[String, String](fromYamlFile(filename))
+    new JavaHashMap[String, String](fromYamlFile(filename))
 
   def fromHoconFileAsJava(filename: String): JavaMap[String, String] =
-    new JavaMap[String, String](fromHoconFile(filename))
+    new JavaHashMap[String, String](fromHoconFile(filename))
 
   def fromPropertiesFileAsJava(filename: String): JavaMap[String, String] =
-    new JavaMap[String, String](fromPropertiesFile(filename))
+    new JavaHashMap[String, String](fromPropertiesFile(filename))
 
   def fromYamlFileAsJava(inputStream: InputStream): JavaMap[String, String] =
-    new JavaMap[String, String](fromYamlFile(inputStream))
+    new JavaHashMap[String, String](fromYamlFile(inputStream))
 
   def fromHoconFileAsJava(inputStream: InputStream): JavaMap[String, String] =
-    new JavaMap[String, String](fromHoconFile(inputStream))
+    new JavaHashMap[String, String](fromHoconFile(inputStream))
 
   def fromPropertiesFileAsJava(inputStream: InputStream): JavaMap[String, 
String] =
-    new JavaMap[String, String](fromPropertiesFile(inputStream))
+    new JavaHashMap[String, String](fromPropertiesFile(inputStream))
 
   /**
    * @param file
@@ -242,7 +243,7 @@ object PropertiesUtils extends Logger {
 
   def loadFlinkConfYaml(yaml: String): JavaMap[String, String] = {
     require(yaml != null && yaml.nonEmpty, "[StreamPark] loadFlinkConfYaml: 
yaml must not be null")
-    val flinkConf = new JavaMap[String, String]()
+    val flinkConf = new JavaHashMap[String, String]()
     val scanner: Scanner = new Scanner(yaml)
     val lineNo: AtomicInteger = new AtomicInteger(0)
     while (scanner.hasNextLine) {
@@ -370,13 +371,13 @@ object PropertiesUtils extends Logger {
   }
 
   @Nonnull def extractDynamicPropertiesAsJava(properties: String): 
JavaMap[String, String] =
-    new JavaMap[String, String](extractDynamicProperties(properties))
+    new JavaHashMap[String, String](extractDynamicProperties(properties))
 
   @Nonnull def extractMultipleArgumentsAsJava(
       args: Array[String]): JavaMap[String, JavaMap[String, String]] = {
     val map =
-      extractMultipleArguments(args).map(c => c._1 -> new JavaMap[String, 
String](c._2))
-    new JavaMap[String, JavaMap[String, String]](map)
+      extractMultipleArguments(args).map(c => c._1 -> new JavaHashMap[String, 
String](c._2))
+    new JavaHashMap[String, JavaMap[String, String]](map)
   }
 
 }
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisClient.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisClient.scala
index b2ae6dcc8..49e02273b 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisClient.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisClient.scala
@@ -18,6 +18,7 @@
 package org.apache.streampark.common.util
 
 import org.apache.streampark.common.Constant
+import org.apache.streampark.common.util.Implicits._
 
 import redis.clients.jedis._
 import redis.clients.jedis.exceptions.JedisConnectionException
@@ -26,7 +27,6 @@ import java.util.concurrent.ConcurrentHashMap
 
 import scala.annotation.meta.getter
 import scala.annotation.tailrec
-import scala.collection.convert.ImplicitConversions._
 import scala.util.Random
 
 object RedisClient extends Logger {
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisEndpoint.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisEndpoint.scala
index e64079199..1edaa771f 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisEndpoint.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisEndpoint.scala
@@ -18,6 +18,7 @@
 package org.apache.streampark.common.util
 
 import org.apache.streampark.common.conf.ConfigKeys._
+import org.apache.streampark.common.util.Implicits._
 
 import redis.clients.jedis.{Jedis, Protocol}
 import redis.clients.jedis.util.{JedisClusterCRC16, JedisURIHelper, 
SafeEncoder}
@@ -25,8 +26,6 @@ import redis.clients.jedis.util.{JedisClusterCRC16, 
JedisURIHelper, SafeEncoder}
 import java.net.URI
 import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
-
 /**
  * RedisEndpoint represents a redis connection endpoint info: host, port, auth 
password db number,
  * and timeout
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisUtils.scala
index a5362158f..31e62c8a1 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/RedisUtils.scala
@@ -17,10 +17,10 @@
 
 package org.apache.streampark.common.util
 
+import org.apache.streampark.common.util.Implicits._
+
 import redis.clients.jedis.{Jedis, JedisCluster, Pipeline, ScanParams}
 
-import scala.collection.JavaConverters._
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.immutable
 import scala.util.{Failure, Success, Try}
 
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/SqlConvertUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/SqlConvertUtils.scala
index e07c3f17b..74d0408b2 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/SqlConvertUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/SqlConvertUtils.scala
@@ -17,12 +17,13 @@
 
 package org.apache.streampark.common.util
 
+import org.apache.streampark.common.util.Implicits._
+
 import java.util
 import java.util.Scanner
 import java.util.regex.Pattern
 
 import scala.annotation.tailrec
-import scala.collection.convert.ImplicitConversions._
 
 object SqlConvertUtils extends Logger {
 
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/Utils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/Utils.scala
index 9cfdd2ac4..e0613bc57 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/Utils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/Utils.scala
@@ -17,19 +17,19 @@
 
 package org.apache.streampark.common.util
 
-import org.apache.streampark.common.util.ImplicitsUtils.AutoCloseImplicits
+import org.apache.streampark.common.util.Implicits._
+import org.apache.streampark.common.util.Implicits.AutoCloseImplicits
 
 import org.apache.commons.lang3.StringUtils
 
 import java.io._
 import java.net.{HttpURLConnection, URL}
 import java.time.{Duration, LocalDateTime}
-import java.util.{jar, Collection => JavaCollection, Map => JavaMap, 
Properties, UUID}
+import java.util.{jar, Properties, UUID}
 import java.util.concurrent.locks.LockSupport
 import java.util.jar.{JarFile, JarInputStream}
 
 import scala.annotation.tailrec
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Failure, Success, Try}
 
 object Utils extends Logger {
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/YarnUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/YarnUtils.scala
index f7c6c8a63..9fdbbde9a 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/YarnUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/YarnUtils.scala
@@ -19,6 +19,7 @@ package org.apache.streampark.common.util
 
 import org.apache.streampark.common.Constant
 import org.apache.streampark.common.conf.{CommonConfig, InternalConfigHolder}
+import org.apache.streampark.common.util.Implicits._
 
 import org.apache.commons.lang3.StringUtils
 import org.apache.hadoop.fs.CommonConfigurationKeys
@@ -34,10 +35,8 @@ import java.io.IOException
 import java.net.InetAddress
 import java.security.PrivilegedExceptionAction
 import java.util
-import java.util.{HashMap => JavaHashMap, List => JavaList}
 import java.util.concurrent.TimeUnit
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.mutable
 import scala.collection.mutable.ArrayBuffer
 import scala.util.{Failure, Success, Try}
diff --git 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/ZooKeeperUtils.scala
 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/ZooKeeperUtils.scala
index b0762e72c..4ebcdb43d 100644
--- 
a/streampark-common/src/main/scala/org/apache/streampark/common/util/ZooKeeperUtils.scala
+++ 
b/streampark-common/src/main/scala/org/apache/streampark/common/util/ZooKeeperUtils.scala
@@ -17,6 +17,8 @@
 
 package org.apache.streampark.common.util
 
+import org.apache.streampark.common.util.Implicits._
+
 import org.apache.curator.RetryPolicy
 import org.apache.curator.framework.{CuratorFramework, CuratorFrameworkFactory}
 import org.apache.curator.retry.RetryNTimes
@@ -24,7 +26,6 @@ import org.apache.zookeeper.CreateMode
 
 import java.nio.charset.StandardCharsets
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.mutable
 
 object ZooKeeperUtils {
diff --git 
a/streampark-common/src/test/scala/org/apache/streampark/common/util/ImplicitsUtilsTest.scala
 
b/streampark-common/src/test/scala/org/apache/streampark/common/util/ImplicitsTest.scala
similarity index 94%
rename from 
streampark-common/src/test/scala/org/apache/streampark/common/util/ImplicitsUtilsTest.scala
rename to 
streampark-common/src/test/scala/org/apache/streampark/common/util/ImplicitsTest.scala
index 7b77e927c..7a36475fa 100644
--- 
a/streampark-common/src/test/scala/org/apache/streampark/common/util/ImplicitsUtilsTest.scala
+++ 
b/streampark-common/src/test/scala/org/apache/streampark/common/util/ImplicitsTest.scala
@@ -17,11 +17,11 @@
 
 package org.apache.streampark.common.util
 
-import org.apache.streampark.common.util.ImplicitsUtils.{AutoCloseImplicits, 
StringImplicits}
+import org.apache.streampark.common.util.Implicits.{AutoCloseImplicits, 
StringImplicits}
 
 import org.scalatest.funsuite.AnyFunSuite
 
-class ImplicitsUtilsTest extends AnyFunSuite {
+class ImplicitsTest extends AnyFunSuite {
   test(
     "AutoCloseImplicits.autoClose should close the resource after execution 
and handle exceptions") {
     class MockResource extends AutoCloseable {
diff --git 
a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/runner/QuickStartRunner.java
 
b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/runner/QuickStartRunner.java
index 3fef15890..6396801c9 100644
--- 
a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/runner/QuickStartRunner.java
+++ 
b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/runner/QuickStartRunner.java
@@ -37,7 +37,6 @@ import org.springframework.boot.ApplicationRunner;
 import org.springframework.core.annotation.Order;
 import org.springframework.stereotype.Component;
 
-import java.util.HashMap;
 import java.util.Map;
 
 @Order
@@ -64,7 +63,8 @@ public class QuickStartRunner implements ApplicationRunner {
 
     @Override
     public void run(ApplicationArguments args) throws Exception {
-        Map<String, HashMap<String, String>> map = 
PropertiesUtils.extractMultipleArgumentsAsJava(args.getSourceArgs());
+
+        Map<String, Map<String, String>> map = 
PropertiesUtils.extractMultipleArgumentsAsJava(args.getSourceArgs());
 
         Map<String, String> quickstart = map.get("quickstart");
 
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/CancelRequest.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/CancelRequest.scala
index 2ac83eb73..8ee9577b6 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/CancelRequest.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/CancelRequest.scala
@@ -20,11 +20,10 @@ package org.apache.streampark.flink.client.bean
 import org.apache.streampark.common.Constant
 import org.apache.streampark.common.conf.FlinkVersion
 import org.apache.streampark.common.enums.FlinkExecutionMode
+import org.apache.streampark.common.util.Implicits.JavaMap
 
 import javax.annotation.Nullable
 
-import java.util.{Map => JavaMap}
-
 case class CancelRequest(
     id: Long,
     flinkVersion: FlinkVersion,
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequest.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequest.scala
index 5e6cf0181..2e6ca27f7 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequest.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequest.scala
@@ -19,6 +19,7 @@ package org.apache.streampark.flink.client.bean
 
 import org.apache.streampark.common.conf.{FlinkVersion, Workspace}
 import org.apache.streampark.common.enums.{FlinkExecutionMode, 
FlinkK8sRestExposedType}
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.util.FlinkUtils
 
 import org.apache.commons.io.FileUtils
@@ -27,7 +28,6 @@ import 
org.apache.flink.kubernetes.configuration.KubernetesConfigOptions
 import javax.annotation.Nullable
 
 import java.io.File
-import java.util.{Map => JavaMap}
 
 case class DeployRequest(
     flinkVersion: FlinkVersion,
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequestTrait.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequestTrait.scala
index 3a4ba8332..7dd3f8092 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequestTrait.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequestTrait.scala
@@ -19,11 +19,10 @@ package org.apache.streampark.flink.client.bean
 
 import org.apache.streampark.common.conf.FlinkVersion
 import org.apache.streampark.common.enums.FlinkExecutionMode
+import org.apache.streampark.common.util.Implicits.JavaMap
 
 import javax.annotation.Nullable
 
-import java.util.{Map => JavaMap}
-
 trait DeployRequestTrait {
 
   val flinkVersion: FlinkVersion
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SavepointRequestTrait.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SavepointRequestTrait.scala
index 526d37163..22c60b62f 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SavepointRequestTrait.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SavepointRequestTrait.scala
@@ -20,11 +20,10 @@ package org.apache.streampark.flink.client.bean
 import org.apache.streampark.common.Constant
 import org.apache.streampark.common.conf.FlinkVersion
 import org.apache.streampark.common.enums.FlinkExecutionMode
+import org.apache.streampark.common.util.Implicits._
 
 import javax.annotation.Nullable
 
-import java.util.{Map => JavaMap}
-
 trait SavepointRequestTrait {
 
   val flinkVersion: FlinkVersion
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/ShutDownRequest.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/ShutDownRequest.scala
index cae66c518..89cf2199e 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/ShutDownRequest.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/ShutDownRequest.scala
@@ -19,11 +19,10 @@ package org.apache.streampark.flink.client.bean
 
 import org.apache.streampark.common.conf.FlinkVersion
 import org.apache.streampark.common.enums.FlinkExecutionMode
+import org.apache.streampark.common.util.Implicits.JavaMap
 
 import javax.annotation.Nullable
 
-import java.util.{Map => JavaMap}
-
 case class ShutDownRequest(
     flinkVersion: FlinkVersion,
     executionMode: FlinkExecutionMode,
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitRequest.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitRequest.scala
index cc1e795c0..1b5184ce3 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitRequest.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitRequest.scala
@@ -22,6 +22,7 @@ import org.apache.streampark.common.conf.{FlinkVersion, 
Workspace}
 import org.apache.streampark.common.conf.ConfigKeys._
 import org.apache.streampark.common.enums._
 import org.apache.streampark.common.util._
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.packer.pipeline.{BuildResult, 
ShadedBuildResponse}
 import org.apache.streampark.flink.util.FlinkUtils
 import org.apache.streampark.shaded.com.fasterxml.jackson.databind.ObjectMapper
@@ -34,9 +35,7 @@ import javax.annotation.Nullable
 
 import java.io.File
 import java.net.URL
-import java.util.{Map => JavaMap}
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.Try
 
 case class SubmitRequest(
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitResponse.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitResponse.scala
index d4d8f037f..e2466a204 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitResponse.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitResponse.scala
@@ -17,9 +17,9 @@
 
 package org.apache.streampark.flink.client.bean
 
-import javax.annotation.Nullable
+import org.apache.streampark.common.util.Implicits.JavaMap
 
-import java.util.{Map => JavaMap}
+import javax.annotation.Nullable
 
 case class SubmitResponse(
     clusterId: String,
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/TriggerSavepointRequest.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/TriggerSavepointRequest.scala
index fbc81a1f3..cb8024c2e 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/TriggerSavepointRequest.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/TriggerSavepointRequest.scala
@@ -20,11 +20,10 @@ package org.apache.streampark.flink.client.bean
 import org.apache.streampark.common.Constant
 import org.apache.streampark.common.conf.FlinkVersion
 import org.apache.streampark.common.enums.FlinkExecutionMode
+import org.apache.streampark.common.util.Implicits.JavaMap
 
 import javax.annotation.Nullable
 
-import java.util.{Map => JavaMap}
-
 /** Trigger savepoint request. */
 case class TriggerSavepointRequest(
     id: Long,
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeApplicationClient.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeApplicationClient.scala
index 7de527637..0de23ca82 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeApplicationClient.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeApplicationClient.scala
@@ -30,8 +30,6 @@ import 
org.apache.flink.kubernetes.configuration.KubernetesConfigOptions
 
 /**
  * Kubernetes native application mode submit.
- * @deprecated
- *   Please use [[KubernetesApplicationClientV2]] instead.
  */
 object KubernetesNativeApplicationClient extends KubernetesNativeClientTrait {
 
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeSessionClient.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeSessionClient.scala
index 746134c3c..79dca5c78 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeSessionClient.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeSessionClient.scala
@@ -19,6 +19,7 @@ package org.apache.streampark.flink.client.impl
 
 import org.apache.streampark.common.enums.FlinkExecutionMode
 import org.apache.streampark.common.util.{Logger, Utils}
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.client.`trait`.KubernetesNativeClientTrait
 import org.apache.streampark.flink.client.bean._
 import org.apache.streampark.flink.client.tool.FlinkSessionSubmitHelper
@@ -27,7 +28,6 @@ import 
org.apache.streampark.flink.kubernetes.KubernetesRetriever
 import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode
 import org.apache.streampark.flink.kubernetes.model.ClusterKey
 
-import io.fabric8.kubernetes.api.model.{Config => _}
 import org.apache.commons.lang3.StringUtils
 import org.apache.flink.client.program.ClusterClient
 import org.apache.flink.configuration._
@@ -38,8 +38,6 @@ import 
org.apache.flink.kubernetes.kubeclient.{FlinkKubeClient, FlinkKubeClientF
 
 import java.io.File
 
-import scala.collection.convert.ImplicitConversions._
-
 /** Kubernetes native session mode submit. */
 object KubernetesNativeSessionClient extends KubernetesNativeClientTrait with 
Logger {
 
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnApplicationClient.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnApplicationClient.scala
index 8e7fd90db..8231857fc 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnApplicationClient.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnApplicationClient.scala
@@ -22,6 +22,7 @@ import org.apache.streampark.common.conf.Workspace
 import org.apache.streampark.common.enums.FlinkDevelopmentMode
 import org.apache.streampark.common.fs.FsOperator
 import org.apache.streampark.common.util.{AssertUtils, FileUtils, HdfsUtils}
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.client.`trait`.YarnClientTrait
 import org.apache.streampark.flink.client.bean._
 import org.apache.streampark.flink.packer.pipeline.ShadedBuildResponse
@@ -40,7 +41,6 @@ import java.security.PrivilegedAction
 import java.util
 import java.util.Collections
 
-import scala.collection.JavaConverters._
 import scala.collection.mutable.ListBuffer
 
 /** yarn application mode submit */
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnPerJobClient.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnPerJobClient.scala
index abbf89126..557bb9967 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnPerJobClient.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnPerJobClient.scala
@@ -17,6 +17,7 @@
 
 package org.apache.streampark.flink.client.impl
 
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.client.`trait`.YarnClientTrait
 import org.apache.streampark.flink.client.bean._
 import org.apache.streampark.flink.util.FlinkUtils
@@ -33,8 +34,6 @@ import org.apache.hadoop.yarn.api.records.ApplicationId
 import java.io.File
 import java.lang.{Boolean => JavaBool}
 
-import scala.collection.convert.ImplicitConversions._
-
 /** yarn PerJob mode submit */
 object YarnPerJobClient extends YarnClientTrait {
 
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnSessionClient.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnSessionClient.scala
index 14af82cc7..a5af099cd 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnSessionClient.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnSessionClient.scala
@@ -18,6 +18,7 @@
 package org.apache.streampark.flink.client.impl
 
 import org.apache.streampark.common.util.{AssertUtils, Utils}
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.client.`trait`.YarnClientTrait
 import org.apache.streampark.flink.client.bean._
 
@@ -36,8 +37,6 @@ import org.apache.hadoop.yarn.util.ConverterUtils
 
 import java.util
 
-import scala.collection.convert.ImplicitConversions._
-
 /** Submit Job to YARN Session Cluster */
 object YarnSessionClient extends YarnClientTrait {
 
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/FlinkClientTrait.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/FlinkClientTrait.scala
index da9187426..7a8891962 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/FlinkClientTrait.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/FlinkClientTrait.scala
@@ -23,6 +23,7 @@ import org.apache.streampark.common.conf.Workspace
 import org.apache.streampark.common.enums._
 import org.apache.streampark.common.fs.FsOperator
 import org.apache.streampark.common.util._
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.client.bean._
 import org.apache.streampark.flink.core.FlinkClusterClient
 import org.apache.streampark.flink.core.conf.FlinkRunOption
@@ -42,9 +43,8 @@ import org.apache.flink.runtime.jobgraph.{JobGraph, 
SavepointConfigOptions}
 import org.apache.flink.util.Preconditions.checkNotNull
 
 import java.io.File
-import java.util.{Collections, List => JavaList, Map => JavaMap}
+import java.util.Collections
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.mutable
 import scala.collection.mutable.ArrayBuffer
 import scala.util.{Failure, Success, Try}
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/YarnClientTrait.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/YarnClientTrait.scala
index e1d8904be..853bc378b 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/YarnClientTrait.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/YarnClientTrait.scala
@@ -18,7 +18,7 @@
 package org.apache.streampark.flink.client.`trait`
 
 import org.apache.streampark.common.util.{AssertUtils, ExceptionUtils}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.client.bean._
 
 import org.apache.flink.api.common.JobID
diff --git 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/test/scala/org/apache/streampark/flink/client/test/YarnPerJobTestCase.scala
 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/test/scala/org/apache/streampark/flink/client/test/YarnPerJobTestCase.scala
index cbfb3bd85..db540f678 100644
--- 
a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/test/scala/org/apache/streampark/flink/client/test/YarnPerJobTestCase.scala
+++ 
b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/test/scala/org/apache/streampark/flink/client/test/YarnPerJobTestCase.scala
@@ -17,6 +17,7 @@
 
 package org.apache.streampark.flink.client.test
 
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.common.util.Logger
 import org.apache.streampark.flink.client.bean
 import org.apache.streampark.flink.core.conf.FlinkRunOption
@@ -41,9 +42,6 @@ import java.io.File
 import java.lang.reflect.Method
 import java.util
 
-import scala.collection.JavaConverters._
-import scala.collection.convert.ImplicitConversions._
-
 /** perJob to submit jobs programmatically, */
 object YarnPerJobTestCase extends Logger {
 
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/conf/ThresholdConf.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/conf/ThresholdConf.scala
index 839847028..bf0a0bd53 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/conf/ThresholdConf.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/conf/ThresholdConf.scala
@@ -22,8 +22,7 @@ import 
org.apache.streampark.flink.connector.conf.FailoverStorageType.{Console,
 
 import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 case class ThresholdConf(prefixStr: String, parameters: Properties) {
 
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverChecker.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverChecker.scala
index 7a05b99c2..85b67f9f1 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverChecker.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverChecker.scala
@@ -21,7 +21,7 @@ import org.apache.streampark.common.util.{Logger, ThreadUtils}
 
 import java.util.concurrent.{Executors, ScheduledExecutorService, 
ThreadFactory, TimeUnit}
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 import scala.collection.mutable.ListBuffer
 
 case class FailoverChecker(delayTime: Long) extends AutoCloseable with Logger {
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverWriter.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverWriter.scala
index 1ef414dde..390a90125 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverWriter.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/FailoverWriter.scala
@@ -26,7 +26,7 @@ import org.apache.kafka.clients.producer.{Callback, 
KafkaProducer, ProducerRecor
 import java.util._
 import java.util.concurrent.locks.ReentrantLock
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 class FailoverWriter(failoverStorage: FailoverStorageType, properties: 
Properties)
   extends AutoCloseable
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/SinkBuffer.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/SinkBuffer.scala
index 8cb9899c0..27b763456 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/SinkBuffer.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/main/scala/org/apache/streampark/flink/connector/failover/SinkBuffer.scala
@@ -23,7 +23,7 @@ import java.util
 import java.util.Collections
 import java.util.concurrent.CopyOnWriteArrayList
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 case class SinkBuffer(writer: SinkWriter, flushInterval: Long, bufferSize: Int)
   extends AutoCloseable
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/test/scala/org/apache/streampark/flink/connector/failover/SinkRequestTest.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/test/scala/org/apache/streampark/flink/connector/failover/SinkRequestTest.scala
index a67917565..634a8cc61 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/test/scala/org/apache/streampark/flink/connector/failover/SinkRequestTest.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-base/src/test/scala/org/apache/streampark/flink/connector/failover/SinkRequestTest.scala
@@ -19,7 +19,7 @@ package org.apache.streampark.flink.connector.failover
 
 import org.junit.jupiter.api.{Assertions, Test}
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 class SinkRequestTest {
   @Test
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseHttpConfig.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseHttpConfig.scala
index 041dbe5ac..83b92516a 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseHttpConfig.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseHttpConfig.scala
@@ -23,7 +23,7 @@ import org.apache.streampark.common.Constant
 import java.util.{Base64, Properties}
 import java.util.concurrent.ThreadLocalRandom
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 /**
  * Flink sink for Clickhouse database. Powered by Async Http Client.
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseSinkConfigOption.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseSinkConfigOption.scala
index b9a8df8d5..4b054fd0c 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseSinkConfigOption.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/conf/ClickHouseSinkConfigOption.scala
@@ -25,7 +25,7 @@ import 
org.asynchttpclient.config.AsyncHttpClientConfigDefaults
 
 import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 object ClickHouseSinkConfigOption {
 
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkFunction.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkFunction.scala
index 707b14737..c13db6b3d 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkFunction.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkFunction.scala
@@ -34,7 +34,7 @@ import java.util
 import java.util.Properties
 import java.util.concurrent.atomic.AtomicLong
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 import scala.util.Try
 
 class ClickHouseSinkFunction[T](apiType: ApiType = ApiType.scala, config: 
Properties)
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkWriter.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkWriter.scala
index f2b147901..2b7c21995 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkWriter.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseSinkWriter.scala
@@ -26,7 +26,7 @@ import org.asynchttpclient.{AsyncHttpClient, 
DefaultAsyncHttpClientConfig, Dsl}
 
 import java.util.concurrent._
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 import scala.collection.mutable.ListBuffer
 
 case class ClickHouseSinkWriter(clickHouseConfig: ClickHouseHttpConfig)
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseWriterTask.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseWriterTask.scala
index 65c4fdf9d..3aaeabbaf 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseWriterTask.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-clickhouse/src/main/scala/org/apache/streampark/flink/connector/clickhouse/internal/ClickHouseWriterTask.scala
@@ -26,7 +26,7 @@ import org.asynchttpclient.{AsyncHttpClient, 
ListenableFuture, Request, Response
 
 import java.util.concurrent.{BlockingQueue, ExecutorService, TimeUnit}
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 import scala.util.Try
 
 case class ClickHouseWriterTask(
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisConfig.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisConfig.scala
index 5407886e8..2046a6817 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisConfig.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisConfig.scala
@@ -20,7 +20,7 @@ import org.apache.streampark.common.Constant
 import org.apache.streampark.common.conf.ConfigKeys
 
 import java.util.Properties
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 object DorisConfig {
 
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisSinkConfigOption.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisSinkConfigOption.scala
index 89a86d66c..91149d6a6 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisSinkConfigOption.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-doris/src/main/scala/org/apache/streampark/connector/doris/conf/DorisSinkConfigOption.scala
@@ -22,7 +22,7 @@ import org.apache.streampark.common.util.ConfigUtils
 
 import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 object DorisSinkConfigOption {
   val DORIS_SINK_PREFIX = "doris.sink"
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/conf/ESSinkConfigOption.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/conf/ESSinkConfigOption.scala
index c7d7bd091..b64645cea 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/conf/ESSinkConfigOption.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/conf/ESSinkConfigOption.scala
@@ -19,11 +19,10 @@ package 
org.apache.streampark.flink.connector.elasticsearch5.conf
 
 import org.apache.streampark.common.conf.ConfigOption
 import org.apache.streampark.common.util.ConfigUtils
+import org.apache.streampark.common.util.Implicits._
 
 import java.net.InetSocketAddress
-import java.util.{Map => JavaMap, Properties}
-
-import scala.collection.convert.ImplicitConversions._
+import java.util.Properties
 
 object ESSinkConfigOption {
   val ES_SINK_PREFIX = "es.sink"
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/sink/ES5Sink.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/sink/ES5Sink.scala
index c1e2ba742..6f999cbc0 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/sink/ES5Sink.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch5/src/main/scala/org/apache/streampark/flink/connector/elasticsearch5/sink/ES5Sink.scala
@@ -31,10 +31,10 @@ import 
org.apache.flink.streaming.connectors.elasticsearch.util.RetryRejectedExe
 import org.apache.flink.streaming.connectors.elasticsearch5.ElasticsearchSink
 import org.elasticsearch.action.ActionRequest
 
-import java.util.{Map => JavaMap, Properties}
+import java.util.Properties
 
 import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 object ES5Sink {
 
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/conf/ES6SinkConfigOption.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/conf/ES6SinkConfigOption.scala
index ea6ff263c..3e3629d24 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/conf/ES6SinkConfigOption.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/conf/ES6SinkConfigOption.scala
@@ -23,9 +23,9 @@ import org.apache.streampark.common.util.ConfigUtils
 import org.apache.http.HttpHost
 
 import java.net.InetSocketAddress
-import java.util.{Map => JavaMap, Properties}
+import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 object ESSinkConfigOption {
 
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/sink/ES6Sink.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/sink/ES6Sink.scala
index 5eacc0bf2..ff1067f93 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/sink/ES6Sink.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch6/src/main/scala/org/apache/streampark/flink/connector/elasticsearch6/sink/ES6Sink.scala
@@ -36,7 +36,7 @@ import org.elasticsearch.action.ActionRequest
 import java.util.Properties
 
 import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 object ES6Sink {
 
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/conf/ES7SinkConfigOption.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/conf/ES7SinkConfigOption.scala
index 2feddef30..60f8b44a3 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/conf/ES7SinkConfigOption.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/conf/ES7SinkConfigOption.scala
@@ -23,9 +23,9 @@ import org.apache.streampark.common.util.ConfigUtils
 import org.apache.http.HttpHost
 
 import java.net.InetSocketAddress
-import java.util.{Map => JavaMap, Properties}
+import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 object ES7SinkConfigOption {
 
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/sink/ES7Sink.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/sink/ES7Sink.scala
index 981772188..2399baad4 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/sink/ES7Sink.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-elasticsearch/streampark-flink-connector-elasticsearch7/src/main/scala/org/apache/streampark/flink/connector/elasticsearch7/sink/ES7Sink.scala
@@ -38,7 +38,7 @@ import org.elasticsearch.action.ActionRequest
 import java.util.Properties
 
 import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 object ES7Sink {
 
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSinkFunction.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSinkFunction.scala
index 072fc8afc..e136607da 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSinkFunction.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSinkFunction.scala
@@ -33,7 +33,7 @@ import java.util.Properties
 import java.util.concurrent.{Executors, ScheduledExecutorService, TimeUnit}
 import java.util.concurrent.atomic.{AtomicBoolean, AtomicLong}
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 import scala.collection.mutable.ArrayBuffer
 
 class HBaseSinkFunction[T](apiType: ApiType = ApiType.scala, tabName: String, 
prop: Properties)
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSourceFunction.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSourceFunction.scala
index 5ea77453f..d91c168b9 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSourceFunction.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/internal/HBaseSourceFunction.scala
@@ -37,7 +37,7 @@ import org.apache.hadoop.hbase.client.{Result, Table}
 import java.lang
 import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 import scala.util.{Success, Try}
 
 class HBaseSourceFunction[R: TypeInformation](apiType: ApiType = 
ApiType.scala, prop: Properties)
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/request/HBaseRequest.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/request/HBaseRequest.scala
index d82bfe14b..03d875996 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/request/HBaseRequest.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-hbase/src/main/scala/org/apache/streampark/flink/connector/hbase/request/HBaseRequest.scala
@@ -31,7 +31,7 @@ import java.util.concurrent.{CompletableFuture, Executors, 
ExecutorService, Time
 import java.util.function.{Consumer, Supplier}
 
 import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 object HBaseRequest {
 
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpSinkWriter.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpSinkWriter.scala
index a7b2f3998..d3ee49947 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpSinkWriter.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpSinkWriter.scala
@@ -25,7 +25,7 @@ import org.asynchttpclient.{AsyncHttpClient, Dsl}
 
 import java.util.concurrent._
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 import scala.collection.mutable.ListBuffer
 
 case class HttpSinkWriter(thresholdConf: ThresholdConf, header: Map[String, 
String])
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpWriterTask.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpWriterTask.scala
index 2dda8cc05..09dd4ad9c 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpWriterTask.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/internal/HttpWriterTask.scala
@@ -28,7 +28,7 @@ import org.asynchttpclient.{AsyncHttpClient, 
ListenableFuture, Request, Response
 import java.util
 import java.util.concurrent.{BlockingQueue, ExecutorService, TimeUnit}
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 import scala.util.Try
 
 case class HttpWriterTask(
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/sink/HttpSink.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/sink/HttpSink.scala
index 058bda317..475f6e4ab 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/sink/HttpSink.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-http/src/main/scala/org/apache/streampark/flink/connector/http/sink/HttpSink.scala
@@ -29,7 +29,7 @@ import org.apache.http.client.methods._
 import java.util.Properties
 
 import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 object HttpSink {
 
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-influx/src/main/scala/org/apache/streampark/flink/connector/influx/function/InfluxFunction.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-influx/src/main/scala/org/apache/streampark/flink/connector/influx/function/InfluxFunction.scala
index c3565c9a2..24a73e08d 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-influx/src/main/scala/org/apache/streampark/flink/connector/influx/function/InfluxFunction.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-influx/src/main/scala/org/apache/streampark/flink/connector/influx/function/InfluxFunction.scala
@@ -28,10 +28,10 @@ import 
org.apache.flink.streaming.api.functions.sink.SinkFunction.Context
 import org.influxdb.{InfluxDB, InfluxDBFactory}
 import org.influxdb.dto.Point
 
-import java.util.{Map => JavaMap, Properties}
+import java.util.Properties
 import java.util.concurrent.TimeUnit
 
-import scala.collection.convert.ImplicitConversions._;
+import org.apache.streampark.common.util.Implicits._;
 
 class InfluxFunction[T](config: Properties)(implicit endpoint: InfluxEntity[T])
   extends RichSinkFunction[T]
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-jdbc/src/main/scala/org/apache/streampark/flink/connector/jdbc/internal/JdbcSourceFunction.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-jdbc/src/main/scala/org/apache/streampark/flink/connector/jdbc/internal/JdbcSourceFunction.scala
index fc7b1b601..1ea839031 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-jdbc/src/main/scala/org/apache/streampark/flink/connector/jdbc/internal/JdbcSourceFunction.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-jdbc/src/main/scala/org/apache/streampark/flink/connector/jdbc/internal/JdbcSourceFunction.scala
@@ -33,9 +33,7 @@ import 
org.apache.flink.streaming.api.functions.source.SourceFunction.SourceCont
 import java.lang
 import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.Map
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Success, Try}
 
 class JdbcSourceFunction[R: TypeInformation](apiType: ApiType = ApiType.scala, 
jdbc: Properties)
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-kafka/src/main/scala/org/apache/streampark/flink/connector/kafka/source/KafkaSource.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-kafka/src/main/scala/org/apache/streampark/flink/connector/kafka/source/KafkaSource.scala
index aea660284..a5e0feb46 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-kafka/src/main/scala/org/apache/streampark/flink/connector/kafka/source/KafkaSource.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-kafka/src/main/scala/org/apache/streampark/flink/connector/kafka/source/KafkaSource.scala
@@ -35,7 +35,7 @@ import java.util.Properties
 import java.util.regex.Pattern
 
 import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 import scala.util.{Failure, Success, Try}
 
 object KafkaSource {
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-mongo/src/main/scala/org/apache/streampark/flink/connector/mongo/internal/MongoSourceFunction.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-mongo/src/main/scala/org/apache/streampark/flink/connector/mongo/internal/MongoSourceFunction.scala
index 0e86fbae1..91ee78b03 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-mongo/src/main/scala/org/apache/streampark/flink/connector/mongo/internal/MongoSourceFunction.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-mongo/src/main/scala/org/apache/streampark/flink/connector/mongo/internal/MongoSourceFunction.scala
@@ -38,7 +38,7 @@ import org.bson.Document
 import java.lang
 import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 import scala.util.{Success, Try}
 
 class MongoSourceFunction[R: TypeInformation](
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/conf/RedisSinkConfigOption.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/conf/RedisSinkConfigOption.scala
index 1bcfa4530..f41168515 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/conf/RedisSinkConfigOption.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/conf/RedisSinkConfigOption.scala
@@ -22,7 +22,7 @@ import org.apache.streampark.common.util.ConfigUtils
 
 import java.util.Properties
 
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 object RedisSinkConfigOption {
   val REDIS_SINK_PREFIX = "redis.sink"
diff --git 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/sink/RedisSink.scala
 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/sink/RedisSink.scala
index 9adb40a43..a7f10fe56 100644
--- 
a/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/sink/RedisSink.scala
+++ 
b/streampark-flink/streampark-flink-connector/streampark-flink-connector-redis/src/main/scala/org/apache/streampark/flink/connector/redis/sink/RedisSink.scala
@@ -36,7 +36,7 @@ import java.util
 import java.util.Properties
 
 import scala.annotation.meta.param
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 import scala.util.Try
 
 object RedisSink {
diff --git 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/FlinkK8sWatchController.scala
 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/FlinkK8sWatchController.scala
index 1faf0c69d..f96e9e5f6 100644
--- 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/FlinkK8sWatchController.scala
+++ 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/FlinkK8sWatchController.scala
@@ -18,6 +18,7 @@
 package org.apache.streampark.flink.kubernetes
 
 import org.apache.streampark.common.util.{Logger, Utils}
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.kubernetes.model._
 
 import com.github.benmanes.caffeine.cache.{Cache, Caffeine}
@@ -25,8 +26,6 @@ import com.github.benmanes.caffeine.cache.{Cache, Caffeine}
 import java.util.Objects
 import java.util.concurrent.TimeUnit
 
-import scala.collection.convert.ImplicitConversions._
-
 /** Tracking info cache pool on flink kubernetes mode. */
 class FlinkK8sWatchController extends Logger with AutoCloseable {
 
diff --git 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/KubernetesRetriever.scala
 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/KubernetesRetriever.scala
index 3e43db40e..b1d1784f8 100644
--- 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/KubernetesRetriever.scala
+++ 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/KubernetesRetriever.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.flink.kubernetes
 
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.common.util.Logger
 import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode
 import org.apache.streampark.flink.kubernetes.ingress.IngressController
@@ -34,7 +34,6 @@ import org.apache.hc.core5.util.Timeout
 
 import javax.annotation.Nullable
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Failure, Success, Try}
 
 object KubernetesRetriever extends Logger {
diff --git 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/PodTemplateParser.scala
 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/PodTemplateParser.scala
index 2cfa3c705..da021d1fb 100644
--- 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/PodTemplateParser.scala
+++ 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/PodTemplateParser.scala
@@ -17,14 +17,14 @@
 
 package org.apache.streampark.flink.kubernetes
 
+import org.apache.streampark.common.util.Implicits._
+
 import org.apache.commons.collections.CollectionUtils
 import org.apache.commons.lang3.StringUtils
 import org.yaml.snakeyaml.Yaml
 
 import java.util
-import java.util.{List => JavaList, Map => JavaMap}
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.Try
 import scala.util.control.Breaks.{break, breakable}
 
diff --git 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/helper/KubernetesDeploymentHelper.scala
 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/helper/KubernetesDeploymentHelper.scala
index ed0dc7e1c..25b25b871 100644
--- 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/helper/KubernetesDeploymentHelper.scala
+++ 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/helper/KubernetesDeploymentHelper.scala
@@ -18,7 +18,7 @@
 package org.apache.streampark.flink.kubernetes.helper
 
 import org.apache.streampark.common.util.{Logger, SystemPropertyUtils}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.kubernetes.KubernetesRetriever
 
 import com.google.common.base.Charsets
@@ -28,7 +28,6 @@ import io.fabric8.kubernetes.client.DefaultKubernetesClient
 
 import java.io.File
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Success, Try}
 
 object KubernetesDeploymentHelper extends Logger {
diff --git 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressController.scala
 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressController.scala
index 5bc4b2650..955fb33e4 100644
--- 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressController.scala
+++ 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressController.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.flink.kubernetes.ingress
 
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.common.util.Logger
 
 import io.fabric8.kubernetes.client.DefaultKubernetesClient
diff --git 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1.scala
 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1.scala
index bea4e627e..064af32b7 100644
--- 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1.scala
+++ 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1.scala
@@ -17,13 +17,12 @@
 
 package org.apache.streampark.flink.kubernetes.ingress
 
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 
 import io.fabric8.kubernetes.api.model.networking.v1.IngressBuilder
 import io.fabric8.kubernetes.client.DefaultKubernetesClient
 import org.apache.flink.client.program.ClusterClient
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Failure, Success, Try}
 
 class IngressStrategyV1 extends IngressStrategy {
diff --git 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1beta1.scala
 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1beta1.scala
index 62c18ca2f..f24f3dbb3 100644
--- 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1beta1.scala
+++ 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/ingress/IngressStrategyV1beta1.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.flink.kubernetes.ingress
 
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 
 import io.fabric8.kubernetes.api.model.IntOrString
 import io.fabric8.kubernetes.api.model.networking.v1beta1.IngressBuilder
@@ -25,7 +25,6 @@ import io.fabric8.kubernetes.client.DefaultKubernetesClient
 import org.apache.commons.lang3.StringUtils
 import org.apache.flink.client.program.ClusterClient
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Failure, Success, Try}
 
 class IngressStrategyV1beta1 extends IngressStrategy {
diff --git 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/watcher/FlinkJobStatusWatcher.scala
 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/watcher/FlinkJobStatusWatcher.scala
index 6e5e03bd4..22446cec2 100644
--- 
a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/watcher/FlinkJobStatusWatcher.scala
+++ 
b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/watcher/FlinkJobStatusWatcher.scala
@@ -17,6 +17,7 @@
 
 package org.apache.streampark.flink.kubernetes.watcher
 
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.common.util.Logger
 import org.apache.streampark.flink.kubernetes._
 import org.apache.streampark.flink.kubernetes.enums.{FlinkJobState, 
FlinkK8sExecuteMode}
@@ -42,7 +43,6 @@ import java.io.File
 import java.nio.charset.StandardCharsets
 import java.util.concurrent.{ScheduledFuture, TimeUnit}
 
-import scala.collection.convert.ImplicitConversions._
 import scala.concurrent.{Await, ExecutionContext, 
ExecutionContextExecutorService, Future}
 import scala.concurrent.duration.DurationLong
 import scala.language.postfixOps
diff --git 
a/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/FlinkRestJsonTest.scala
 
b/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/FlinkRestJsonTest.scala
index c614e0a2c..74528d875 100644
--- 
a/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/FlinkRestJsonTest.scala
+++ 
b/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/FlinkRestJsonTest.scala
@@ -17,6 +17,7 @@
 
 package org.apache.streampark.flink.kubernetes
 
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.kubernetes.helper.KubernetesDeploymentHelper
 import org.apache.streampark.flink.kubernetes.watcher.{Checkpoint, 
FlinkRestJmConfigItem, FlinkRestOverview, JobDetails}
 
@@ -33,7 +34,6 @@ import org.junit.jupiter.api.Test
 
 import java.io.File
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Failure, Success, Try}
 
 // scalastyle:off println
diff --git 
a/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/PodTemplateParserTest.scala
 
b/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/PodTemplateParserTest.scala
index ba8f0ae0b..ca37bbba6 100644
--- 
a/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/PodTemplateParserTest.scala
+++ 
b/streampark-flink/streampark-flink-kubernetes/src/test/scala/org/apache/streampark/flink/kubernetes/PodTemplateParserTest.scala
@@ -17,11 +17,11 @@
 
 package org.apache.streampark.flink.kubernetes
 
+import org.apache.streampark.common.util.Implicits._
+
 import org.junit.jupiter.api.Assertions.assertEquals
 import org.junit.jupiter.api.Test
 
-import scala.collection.convert.ImplicitConversions._
-
 // scalastyle:off println
 class PodTemplateParserTest {
 
diff --git 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/DependencyInfo.scala
 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/DependencyInfo.scala
index 9384fe9c4..bc035bdd8 100644
--- 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/DependencyInfo.scala
+++ 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/DependencyInfo.scala
@@ -17,9 +17,7 @@
 
 package org.apache.streampark.flink.packer.maven
 
-import java.util.{List => JavaList}
-
-import scala.collection.convert.ImplicitConversions._
+import org.apache.streampark.common.util.Implicits._
 
 /**
  * @param mavenArts
diff --git 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/MavenTool.scala
 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/MavenTool.scala
index b640ad976..b0f199648 100644
--- 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/MavenTool.scala
+++ 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/maven/MavenTool.scala
@@ -21,6 +21,7 @@ import org.apache.streampark.common.Constant
 import org.apache.streampark.common.conf.{InternalConfigHolder, Workspace}
 import org.apache.streampark.common.conf.CommonConfig.{MAVEN_AUTH_PASSWORD, 
MAVEN_AUTH_USER, MAVEN_REMOTE_URL}
 import org.apache.streampark.common.util.{AssertUtils, Logger, Utils}
+import org.apache.streampark.common.util.Implicits._
 
 import com.google.common.collect.Lists
 import org.apache.maven.plugins.shade.{DefaultShader, ShadeRequest}
@@ -44,9 +45,7 @@ import javax.annotation.{Nonnull, Nullable}
 
 import java.io.File
 import java.util
-import java.util.{List => JavaList, Set => JavaSet}
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.mutable.ArrayBuffer
 import scala.util.Try
 
diff --git 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildPipeline.scala
 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildPipeline.scala
index b711d3ed0..25062f94b 100644
--- 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildPipeline.scala
+++ 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildPipeline.scala
@@ -18,11 +18,11 @@
 package org.apache.streampark.flink.packer.pipeline
 
 import org.apache.streampark.common.util.{Logger, ThreadUtils}
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.packer.pipeline.BuildPipeline.executor
 
 import java.util.concurrent.{Callable, LinkedBlockingQueue, 
ThreadPoolExecutor, TimeUnit}
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.mutable
 import scala.concurrent.{ExecutionContext, ExecutionContextExecutorService}
 import scala.util.{Failure, Success, Try}
diff --git 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/DockerResolveSnapshot.scala
 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/DockerResolveSnapshot.scala
index 14cd76af0..61fa36bc5 100644
--- 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/DockerResolveSnapshot.scala
+++ 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/DockerResolveSnapshot.scala
@@ -17,14 +17,11 @@
 
 package org.apache.streampark.flink.packer.pipeline
 
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.common.util.Utils
 
 import com.fasterxml.jackson.annotation.JsonIgnoreProperties
 
-import java.util.{List => JavaList}
-
-import scala.collection.convert.ImplicitConversions._
-
 /** Snapshot for docker resolved progress */
 @JsonIgnoreProperties(ignoreUnknown = true)
 case class DockerResolvedSnapshot(
diff --git 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/PipelineSnapshot.scala
 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/PipelineSnapshot.scala
index ea84594ca..2bf492034 100644
--- 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/PipelineSnapshot.scala
+++ 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/PipelineSnapshot.scala
@@ -17,12 +17,10 @@
 
 package org.apache.streampark.flink.packer.pipeline
 
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.common.util.Utils
 
 import java.lang.{Long => JavaLong}
-import java.util.{Map => JavaMap}
-
-import scala.collection.JavaConverters._
 
 /**
  * Snapshot for a BuildPipeline instance. see
diff --git 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkRemoteBuildPipeline.scala
 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkRemoteBuildPipeline.scala
index b0d9b90c6..c7d16e44d 100644
--- 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkRemoteBuildPipeline.scala
+++ 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkRemoteBuildPipeline.scala
@@ -19,13 +19,12 @@ package org.apache.streampark.flink.packer.pipeline.impl
 
 import org.apache.streampark.common.enums.FlinkDevelopmentMode
 import org.apache.streampark.common.fs.{FsOperator, LfsOperator}
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.packer.maven.MavenTool
 import org.apache.streampark.flink.packer.pipeline._
 
 import java.io.File
 
-import scala.collection.convert.ImplicitConversions._
-
 /** Building pipeline for flink standalone session mode */
 class FlinkRemoteBuildPipeline(request: FlinkRemotePerJobBuildRequest) extends 
BuildPipeline {
 
diff --git 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkYarnApplicationBuildPipeline.scala
 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkYarnApplicationBuildPipeline.scala
index 93124d22d..66d5fcff3 100644
--- 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkYarnApplicationBuildPipeline.scala
+++ 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkYarnApplicationBuildPipeline.scala
@@ -20,7 +20,7 @@ package org.apache.streampark.flink.packer.pipeline.impl
 import org.apache.streampark.common.conf.Workspace
 import org.apache.streampark.common.enums.FlinkDevelopmentMode
 import org.apache.streampark.common.fs.{FsOperator, HdfsOperator, LfsOperator}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.packer.maven.MavenTool
 import org.apache.streampark.flink.packer.pipeline._
 
@@ -28,8 +28,6 @@ import org.apache.commons.codec.digest.DigestUtils
 
 import java.io.{File, FileInputStream, IOException}
 
-import scala.collection.convert.ImplicitConversions._
-
 /** Building pipeline for flink yarn application mode */
 class FlinkYarnApplicationBuildPipeline(request: 
FlinkYarnApplicationBuildRequest)
   extends BuildPipeline {
diff --git 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/SparkYarnApplicationBuildPipeline.scala
 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/SparkYarnApplicationBuildPipeline.scala
index 3cac30ada..9b91d1dcb 100644
--- 
a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/SparkYarnApplicationBuildPipeline.scala
+++ 
b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/SparkYarnApplicationBuildPipeline.scala
@@ -20,7 +20,7 @@ package org.apache.streampark.flink.packer.pipeline.impl
 import org.apache.streampark.common.conf.Workspace
 import org.apache.streampark.common.enums.FlinkDevelopmentMode
 import org.apache.streampark.common.fs.{FsOperator, HdfsOperator, LfsOperator}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.packer.maven.MavenTool
 import org.apache.streampark.flink.packer.pipeline._
 
@@ -28,8 +28,6 @@ import org.apache.commons.codec.digest.DigestUtils
 
 import java.io.{File, FileInputStream, IOException}
 
-import scala.collection.convert.ImplicitConversions._
-
 /** Building pipeline for spark yarn application mode */
 class SparkYarnApplicationBuildPipeline(request: 
SparkYarnApplicationBuildRequest)
   extends BuildPipeline {
diff --git 
a/streampark-flink/streampark-flink-packer/src/test/scala/org/apache/streampark/flink/packer/MavenToolSpec.scala
 
b/streampark-flink/streampark-flink-packer/src/test/scala/org/apache/streampark/flink/packer/MavenToolSpec.scala
index 821577630..67735d9cf 100644
--- 
a/streampark-flink/streampark-flink-packer/src/test/scala/org/apache/streampark/flink/packer/MavenToolSpec.scala
+++ 
b/streampark-flink/streampark-flink-packer/src/test/scala/org/apache/streampark/flink/packer/MavenToolSpec.scala
@@ -17,6 +17,7 @@
 
 package org.apache.streampark.flink.packer
 
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.packer.maven.{Artifact, DependencyInfo, 
MavenTool}
 
 import org.apache.commons.io.FileUtils
@@ -27,8 +28,6 @@ import org.scalatest.wordspec.AnyWordSpec
 import java.io.File
 import java.util.jar.JarFile
 
-import scala.collection.convert.ImplicitConversions._
-
 class MavenToolSpec extends AnyWordSpec with BeforeAndAfterAll with Matchers {
 
   val outputDir = "MavenToolSpec-output/"
diff --git 
a/streampark-flink/streampark-flink-proxy/src/main/scala/org/apache/streampark/flink/proxy/FlinkShimsProxy.scala
 
b/streampark-flink/streampark-flink-proxy/src/main/scala/org/apache/streampark/flink/proxy/FlinkShimsProxy.scala
index 5dd25ec5f..34ccb9f59 100644
--- 
a/streampark-flink/streampark-flink-proxy/src/main/scala/org/apache/streampark/flink/proxy/FlinkShimsProxy.scala
+++ 
b/streampark-flink/streampark-flink-proxy/src/main/scala/org/apache/streampark/flink/proxy/FlinkShimsProxy.scala
@@ -20,7 +20,7 @@ package org.apache.streampark.flink.proxy
 import org.apache.streampark.common.Constant
 import org.apache.streampark.common.conf.{ConfigKeys, FlinkVersion}
 import org.apache.streampark.common.util.{ChildFirstClassLoader, 
ClassLoaderObjectInputStream, ClassLoaderUtils, Logger}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 
 import java.io.{ByteArrayInputStream, ByteArrayOutputStream, File, 
ObjectOutputStream}
 import java.net.URL
diff --git 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamTableTrait.scala
 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamTableTrait.scala
index 4fc75d616..9b3a804c3 100644
--- 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamTableTrait.scala
+++ 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamTableTrait.scala
@@ -17,6 +17,7 @@
 
 package org.apache.streampark.flink.core
 
+import org.apache.streampark.common.util.Implicits.JavaList
 import org.apache.streampark.common.util.Utils
 import org.apache.streampark.flink.core.EnhancerImplicit._
 
@@ -47,7 +48,7 @@ import org.apache.flink.table.module.Module
 import org.apache.flink.table.types.AbstractDataType
 import org.apache.flink.util.SplittableIterator
 
-import java.util.{List => JavaList, Optional}
+import java.util.Optional
 
 /**
  * Integration api of stream and table
diff --git 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamingInitializer.scala
 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamingInitializer.scala
index 2c8fa5f86..7ce7a85a0 100644
--- 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamingInitializer.scala
+++ 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamingInitializer.scala
@@ -21,6 +21,7 @@ import org.apache.streampark.common.conf.ConfigKeys._
 import org.apache.streampark.common.enums.ApiType
 import org.apache.streampark.common.enums.ApiType.ApiType
 import org.apache.streampark.common.util._
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.core.conf.FlinkConfiguration
 
 import collection.{mutable, Map}
@@ -32,8 +33,6 @@ import org.apache.flink.table.api.TableConfig
 
 import java.io.File
 
-import scala.collection.convert.ImplicitConversions._
-
 private[flink] object FlinkStreamingInitializer {
 
   def initialize(args: Array[String], config: (StreamExecutionEnvironment, 
ParameterTool) => Unit): (ParameterTool, StreamExecutionEnvironment) = {
diff --git 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkTableInitializer.scala
 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkTableInitializer.scala
index 71ac0a523..7cb463ed7 100644
--- 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkTableInitializer.scala
+++ 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkTableInitializer.scala
@@ -21,6 +21,7 @@ import org.apache.streampark.common.conf.ConfigKeys._
 import org.apache.streampark.common.enums.{ApiType, PlannerType}
 import org.apache.streampark.common.enums.ApiType.ApiType
 import org.apache.streampark.common.util.{DeflaterUtils, PropertiesUtils}
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.core.EnhancerImplicit._
 import org.apache.streampark.flink.core.conf.FlinkConfiguration
 
@@ -33,7 +34,6 @@ import 
org.apache.flink.table.api.bridge.scala.StreamTableEnvironment
 import java.io.File
 
 import scala.collection.{mutable, Map}
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Failure, Success, Try}
 
 private[flink] object FlinkTableInitializer {
diff --git 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/conf/FlinkRunOption.scala
 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/conf/FlinkRunOption.scala
index e0c30cb06..6272a018d 100644
--- 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/conf/FlinkRunOption.scala
+++ 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/conf/FlinkRunOption.scala
@@ -17,11 +17,12 @@
 
 package org.apache.streampark.flink.core.conf
 
+import org.apache.streampark.common.util.Implicits._
+
 import org.apache.commons.cli.{CommandLine, DefaultParser, Option, Options}
 
 import java.lang.{Boolean => JavaBoolean}
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Failure, Success, Try}
 
 /** Applies to all optional parameters under flink run */
diff --git 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
index f5542df84..b2c56a264 100644
--- 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
+++ 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
@@ -17,6 +17,8 @@
 
 package org.apache.streampark.flink.core
 
+import org.apache.streampark.common.util.Implicits.JavaList
+
 import org.apache.flink.api.java.utils.ParameterTool
 import org.apache.flink.streaming.api.scala.{DataStream, 
StreamExecutionEnvironment}
 import org.apache.flink.table.api._
@@ -27,8 +29,6 @@ import org.apache.flink.table.resource.ResourceUri
 import org.apache.flink.table.types.AbstractDataType
 import org.apache.flink.types.Row
 
-import java.util.{List => JavaList}
-
 class StreamTableContext(
     override val parameter: ParameterTool,
     private val streamEnv: StreamExecutionEnvironment,
diff --git 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
index 6b9d0f03b..e7fe52592 100644
--- 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
+++ 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.17/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
@@ -17,13 +17,13 @@
 
 package org.apache.streampark.flink.core
 
+import org.apache.streampark.common.util.Implicits.JavaList
+
 import org.apache.flink.api.java.utils.ParameterTool
 import org.apache.flink.table.api._
 import org.apache.flink.table.module.ModuleEntry
 import org.apache.flink.table.resource.ResourceUri
 
-import java.util.{List => JavaList}
-
 class TableContext(override val parameter: ParameterTool, private val 
tableEnv: TableEnvironment)
   extends FlinkTableTrait(parameter, tableEnv) {
 
diff --git 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
index f5db09054..8d655abec 100644
--- 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
+++ 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
@@ -17,6 +17,8 @@
 
 package org.apache.streampark.flink.core
 
+import org.apache.streampark.common.util.Implicits.JavaList
+
 import org.apache.flink.api.java.utils.ParameterTool
 import org.apache.flink.streaming.api.scala.{DataStream, 
StreamExecutionEnvironment}
 import org.apache.flink.table.api._
@@ -28,8 +30,6 @@ import org.apache.flink.table.resource.ResourceUri
 import org.apache.flink.table.types.AbstractDataType
 import org.apache.flink.types.Row
 
-import java.util.{List => JavaList}
-
 class StreamTableContext(
     override val parameter: ParameterTool,
     private val streamEnv: StreamExecutionEnvironment,
diff --git 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
index c1b348ae8..9d295b6b7 100644
--- 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
+++ 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.18/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
@@ -17,14 +17,14 @@
 
 package org.apache.streampark.flink.core
 
+import org.apache.streampark.common.util.Implicits.JavaList
+
 import org.apache.flink.api.java.utils.ParameterTool
 import org.apache.flink.table.api._
 import org.apache.flink.table.catalog.CatalogDescriptor
 import org.apache.flink.table.module.ModuleEntry
 import org.apache.flink.table.resource.ResourceUri
 
-import java.util.{List => JavaList}
-
 class TableContext(override val parameter: ParameterTool, private val 
tableEnv: TableEnvironment)
   extends FlinkTableTrait(parameter, tableEnv) {
 
diff --git 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
index f5db09054..8d655abec 100644
--- 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
+++ 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/StreamTableContext.scala
@@ -17,6 +17,8 @@
 
 package org.apache.streampark.flink.core
 
+import org.apache.streampark.common.util.Implicits.JavaList
+
 import org.apache.flink.api.java.utils.ParameterTool
 import org.apache.flink.streaming.api.scala.{DataStream, 
StreamExecutionEnvironment}
 import org.apache.flink.table.api._
@@ -28,8 +30,6 @@ import org.apache.flink.table.resource.ResourceUri
 import org.apache.flink.table.types.AbstractDataType
 import org.apache.flink.types.Row
 
-import java.util.{List => JavaList}
-
 class StreamTableContext(
     override val parameter: ParameterTool,
     private val streamEnv: StreamExecutionEnvironment,
diff --git 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
index c1b348ae8..9d295b6b7 100644
--- 
a/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
+++ 
b/streampark-flink/streampark-flink-shims/streampark-flink-shims_flink-1.19/src/main/scala/org/apache/streampark/flink/core/TableContext.scala
@@ -17,14 +17,14 @@
 
 package org.apache.streampark.flink.core
 
+import org.apache.streampark.common.util.Implicits.JavaList
+
 import org.apache.flink.api.java.utils.ParameterTool
 import org.apache.flink.table.api._
 import org.apache.flink.table.catalog.CatalogDescriptor
 import org.apache.flink.table.module.ModuleEntry
 import org.apache.flink.table.resource.ResourceUri
 
-import java.util.{List => JavaList}
-
 class TableContext(override val parameter: ParameterTool, private val 
tableEnv: TableEnvironment)
   extends FlinkTableTrait(parameter, tableEnv) {
 
diff --git 
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/StopRequest.scala
 
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/StopRequest.scala
index 2f8b5fcbb..1adcc862e 100644
--- 
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/StopRequest.scala
+++ 
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/StopRequest.scala
@@ -19,11 +19,10 @@ package org.apache.streampark.spark.client.bean
 
 import org.apache.streampark.common.conf.SparkVersion
 import org.apache.streampark.common.enums.SparkExecutionMode
+import org.apache.streampark.common.util.Implicits.JavaMap
 
 import javax.annotation.Nullable
 
-import java.util.{Map => JavaMap}
-
 case class StopRequest(
     id: Long,
     sparkVersion: SparkVersion,
diff --git 
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitRequest.scala
 
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitRequest.scala
index cf71c99b9..4079f47e4 100644
--- 
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitRequest.scala
+++ 
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitRequest.scala
@@ -22,6 +22,7 @@ import org.apache.streampark.common.conf.{SparkVersion, 
Workspace}
 import org.apache.streampark.common.conf.ConfigKeys._
 import org.apache.streampark.common.enums._
 import org.apache.streampark.common.util.{DeflaterUtils, HdfsUtils, 
PropertiesUtils}
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.flink.packer.pipeline.{BuildResult, 
ShadedBuildResponse}
 
 import com.fasterxml.jackson.databind.ObjectMapper
@@ -31,9 +32,6 @@ import javax.annotation.Nullable
 
 import java.io.{File, IOException}
 import java.nio.file.Files
-import java.util.{Map => JavaMap}
-
-import scala.collection.convert.ImplicitConversions._
 
 case class SubmitRequest(
     sparkVersion: SparkVersion,
diff --git 
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitResponse.scala
 
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitResponse.scala
index b97b1f146..da5720a2d 100644
--- 
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitResponse.scala
+++ 
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/bean/SubmitResponse.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.spark.client.bean
 
-import java.util.{Map => JavaMap}
+import org.apache.streampark.common.util.Implicits.JavaMap
 
 case class SubmitResponse(
     clusterId: String,
diff --git 
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/proxy/SparkShimsProxy.scala
 
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/proxy/SparkShimsProxy.scala
index 97e749e2b..873a5b531 100644
--- 
a/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/proxy/SparkShimsProxy.scala
+++ 
b/streampark-spark/streampark-spark-client/streampark-spark-client-api/src/main/scala/org/apache/streampark/spark/client/proxy/SparkShimsProxy.scala
@@ -20,7 +20,7 @@ package org.apache.streampark.spark.client.proxy
 import org.apache.streampark.common.Constant
 import org.apache.streampark.common.conf.{ConfigKeys, SparkVersion}
 import org.apache.streampark.common.util.{ChildFirstClassLoader, 
ClassLoaderObjectInputStream, ClassLoaderUtils, Logger}
-import org.apache.streampark.common.util.ImplicitsUtils._
+import org.apache.streampark.common.util.Implicits._
 
 import java.io.{ByteArrayInputStream, ByteArrayOutputStream, File, 
ObjectOutputStream}
 import java.net.URL
diff --git 
a/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/impl/YarnClient.scala
 
b/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/impl/YarnClient.scala
index 8248a04cb..19e6beaf7 100644
--- 
a/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/impl/YarnClient.scala
+++ 
b/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/impl/YarnClient.scala
@@ -19,6 +19,7 @@ package org.apache.streampark.spark.client.impl
 
 import org.apache.streampark.common.enums.SparkExecutionMode
 import org.apache.streampark.common.util.HadoopUtils
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.spark.client.`trait`.SparkClientTrait
 import org.apache.streampark.spark.client.bean._
 
@@ -27,7 +28,6 @@ import org.apache.spark.launcher.{SparkAppHandle, 
SparkLauncher}
 
 import java.util.concurrent.CountDownLatch
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Failure, Success, Try}
 
 /** yarn application mode submit */
diff --git 
a/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/trait/SparkClientTrait.scala
 
b/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/trait/SparkClientTrait.scala
index bfefa8892..2ef438dc0 100644
--- 
a/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/trait/SparkClientTrait.scala
+++ 
b/streampark-spark/streampark-spark-client/streampark-spark-client-core/src/main/scala/org/apache/streampark/spark/client/trait/SparkClientTrait.scala
@@ -18,9 +18,9 @@
 package org.apache.streampark.spark.client.`trait`
 
 import org.apache.streampark.common.util._
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.spark.client.bean._
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.{Failure, Success, Try}
 
 trait SparkClientTrait extends Logger {
diff --git 
a/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/HBaseOffset.scala
 
b/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/HBaseOffset.scala
index a3451cbe8..de7ca0a25 100644
--- 
a/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/HBaseOffset.scala
+++ 
b/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/HBaseOffset.scala
@@ -18,6 +18,7 @@
 package org.apache.streampark.spark.connector.kafka.offset
 
 import org.apache.streampark.common.util.HBaseClient
+import org.apache.streampark.common.util.Implicits._
 
 import org.apache.hadoop.hbase._
 import org.apache.hadoop.hbase.client.{Delete, Put, Scan, Table}
@@ -28,7 +29,6 @@ import org.apache.spark.SparkConf
 
 import java.util
 
-import scala.collection.convert.ImplicitConversions._
 import scala.collection.mutable
 
 private[kafka] class HBaseOffset(val sparkConf: SparkConf) extends Offset {
diff --git 
a/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/Offset.scala
 
b/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/Offset.scala
index 6b6048038..eb8f87b57 100644
--- 
a/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/Offset.scala
+++ 
b/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/Offset.scala
@@ -17,6 +17,7 @@
 
 package org.apache.streampark.spark.connector.kafka.offset
 
+import org.apache.streampark.common.util.Implicits._
 import org.apache.streampark.common.util.Logger
 
 import org.apache.kafka.common.TopicPartition
@@ -24,7 +25,6 @@ import org.apache.spark.SparkConf
 
 import java.util.Properties
 
-import scala.collection.JavaConverters._
 import scala.language.implicitConversions
 import scala.util.Try
 
diff --git 
a/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/RedisOffset.scala
 
b/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/RedisOffset.scala
index 077cfba31..c512fc06e 100644
--- 
a/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/RedisOffset.scala
+++ 
b/streampark-spark/streampark-spark-connector/streampark-spark-connector-kafka/src/main/scala/org/apache/streampark/spark/connector/kafka/offset/RedisOffset.scala
@@ -18,12 +18,12 @@
 package org.apache.streampark.spark.connector.kafka.offset
 
 import org.apache.streampark.common.util.{RedisEndpoint, RedisUtils}
+import org.apache.streampark.common.util.Implicits._
 
 import org.apache.kafka.common.TopicPartition
 import org.apache.spark.SparkConf
 import redis.clients.jedis.Protocol
 
-import scala.collection.convert.ImplicitConversions._
 import scala.util.Try
 
 /** Redis Offset Manager */
diff --git a/tools/dependencies/check-LICENSE.sh 
b/tools/dependencies/check-LICENSE.sh
index 58ca164db..f56aebd1d 100755
--- a/tools/dependencies/check-LICENSE.sh
+++ b/tools/dependencies/check-LICENSE.sh
@@ -19,9 +19,8 @@
 
 check_path=dist-license-check
 
-if [ -d "$check_path" ];then
-rm -rf $check_path
-fi
+[[ -d "$check_path" ]] && rm -rf $check_path
+
 mkdir $check_path || true
 
 tar -zxf dist/apache-streampark*-bin.tar.gz --strip=1 -C $check_path
@@ -45,7 +44,7 @@ echo '=== Third party dependencies: ' && grep -vf 
$check_path/self-modules.txt $
 
 diff -w -B -U0 <(sort < tools/dependencies/known-dependencies.txt) <(sort < 
$check_path/third-party-dependencies.txt)
 
-if [ $? -ne 0 ]; then
+if [[ $? -ne 0 ]]; then
   echo "Third-party dependencies are not all known, please add the license to 
LICENSE file and add the dependency to 
tools/dependencies/known-dependencies.txt"
   exit 1
 fi


Reply via email to