This is an automated email from the ASF dual-hosted git repository.

martijnvisser pushed a commit to branch v3.1
in repository https://gitbox.apache.org/repos/asf/flink-connector-jdbc.git


The following commit(s) were added to refs/heads/v3.1 by this push:
     new 19b61ba  [FLINK-31793][JDBC] remove dependency on flink-shaded guava 
for flink-connector-jdbc
19b61ba is described below

commit 19b61babb201f61c82b89202c45ce569bf038aa5
Author: Wencong Liu <[email protected]>
AuthorDate: Fri Apr 14 11:51:44 2023 +0800

    [FLINK-31793][JDBC] remove dependency on flink-shaded guava for 
flink-connector-jdbc
    
    (cherry picked from commit 7f5a8b7310671ae1f2a244e14de6cf233e1cf05b)
---
 .../flink/connector/jdbc/utils/JdbcTypeUtil.java   | 43 +++++++++++-----------
 .../jdbc/catalog/MySqlCatalogTestBase.java         | 11 +++---
 pom.xml                                            |  6 ---
 tools/maven/checkstyle.xml                         |  4 --
 4 files changed, 27 insertions(+), 37 deletions(-)

diff --git 
a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/utils/JdbcTypeUtil.java
 
b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/utils/JdbcTypeUtil.java
index 16a62f0..4c4c434 100644
--- 
a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/utils/JdbcTypeUtil.java
+++ 
b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/utils/JdbcTypeUtil.java
@@ -26,8 +26,6 @@ import org.apache.flink.api.common.typeinfo.TypeInformation;
 import org.apache.flink.api.java.typeutils.ObjectArrayTypeInfo;
 import org.apache.flink.table.types.logical.LogicalTypeRoot;
 
-import org.apache.flink.shaded.guava30.com.google.common.collect.ImmutableMap;
-
 import java.sql.Types;
 import java.util.Collections;
 import java.util.HashMap;
@@ -89,25 +87,28 @@ public class JdbcTypeUtil {
     }
 
     private static final Map<LogicalTypeRoot, Integer> LOGICAL_TYPE_MAPPING =
-            ImmutableMap.<LogicalTypeRoot, Integer>builder()
-                    .put(VARCHAR, Types.VARCHAR)
-                    .put(CHAR, Types.CHAR)
-                    .put(VARBINARY, Types.VARBINARY)
-                    .put(BOOLEAN, Types.BOOLEAN)
-                    .put(BINARY, Types.BINARY)
-                    .put(TINYINT, Types.TINYINT)
-                    .put(SMALLINT, Types.SMALLINT)
-                    .put(INTEGER, Types.INTEGER)
-                    .put(BIGINT, Types.BIGINT)
-                    .put(FLOAT, Types.REAL)
-                    .put(DOUBLE, Types.DOUBLE)
-                    .put(DATE, Types.DATE)
-                    .put(TIMESTAMP_WITHOUT_TIME_ZONE, Types.TIMESTAMP)
-                    .put(TIMESTAMP_WITH_TIME_ZONE, 
Types.TIMESTAMP_WITH_TIMEZONE)
-                    .put(TIME_WITHOUT_TIME_ZONE, Types.TIME)
-                    .put(DECIMAL, Types.DECIMAL)
-                    .put(ARRAY, Types.ARRAY)
-                    .build();
+            Collections.unmodifiableMap(
+                    new HashMap<LogicalTypeRoot, Integer>() {
+                        {
+                            put(VARCHAR, Types.VARCHAR);
+                            put(CHAR, Types.CHAR);
+                            put(VARBINARY, Types.VARBINARY);
+                            put(BOOLEAN, Types.BOOLEAN);
+                            put(BINARY, Types.BINARY);
+                            put(TINYINT, Types.TINYINT);
+                            put(SMALLINT, Types.SMALLINT);
+                            put(INTEGER, Types.INTEGER);
+                            put(BIGINT, Types.BIGINT);
+                            put(FLOAT, Types.REAL);
+                            put(DOUBLE, Types.DOUBLE);
+                            put(DATE, Types.DATE);
+                            put(TIMESTAMP_WITHOUT_TIME_ZONE, Types.TIMESTAMP);
+                            put(TIMESTAMP_WITH_TIME_ZONE, 
Types.TIMESTAMP_WITH_TIMEZONE);
+                            put(TIME_WITHOUT_TIME_ZONE, Types.TIME);
+                            put(DECIMAL, Types.DECIMAL);
+                            put(ARRAY, Types.ARRAY);
+                        }
+                    });
 
     private JdbcTypeUtil() {}
 
diff --git 
a/flink-connector-jdbc/src/test/java/org/apache/flink/connector/jdbc/catalog/MySqlCatalogTestBase.java
 
b/flink-connector-jdbc/src/test/java/org/apache/flink/connector/jdbc/catalog/MySqlCatalogTestBase.java
index e46b029..9920a7f 100644
--- 
a/flink-connector-jdbc/src/test/java/org/apache/flink/connector/jdbc/catalog/MySqlCatalogTestBase.java
+++ 
b/flink-connector-jdbc/src/test/java/org/apache/flink/connector/jdbc/catalog/MySqlCatalogTestBase.java
@@ -30,8 +30,6 @@ import org.apache.flink.types.Row;
 import org.apache.flink.types.RowKind;
 import org.apache.flink.util.CollectionUtil;
 
-import org.apache.flink.shaded.guava30.com.google.common.collect.Lists;
-
 import org.junit.jupiter.api.BeforeEach;
 import org.junit.jupiter.api.Test;
 import org.slf4j.Logger;
@@ -123,11 +121,11 @@ abstract class MySqlCatalogTestBase {
                     .column("col_time_p3", DataTypes.TIME(3))
                     .column("col_timestamp_p3", DataTypes.TIMESTAMP(3))
                     .column("col_varbinary", DataTypes.BYTES())
-                    .primaryKeyNamed("PRIMARY", Lists.newArrayList("pid"))
+                    .primaryKeyNamed("PRIMARY", 
Collections.singletonList("pid"))
                     .build();
 
     protected static final List<Row> TABLE_ROWS =
-            Lists.newArrayList(
+            Arrays.asList(
                     Row.ofKind(
                             RowKind.INSERT,
                             1L,
@@ -380,7 +378,7 @@ abstract class MySqlCatalogTestBase {
                                 .collect());
         assertThat(results)
                 .isEqualTo(
-                        Lists.newArrayList(
+                        Arrays.asList(
                                 Row.ofKind(RowKind.INSERT, 1L), 
Row.ofKind(RowKind.INSERT, 2L)));
     }
 
@@ -458,6 +456,7 @@ abstract class MySqlCatalogTestBase {
                                                 
TEST_TABLE_SINK_FROM_GROUPED_BY))
                                 .execute()
                                 .collect());
-        
assertThat(results).isEqualTo(Lists.newArrayList(Row.ofKind(RowKind.INSERT, 2L, 
-1L)));
+        assertThat(results)
+                
.isEqualTo(Collections.singletonList(Row.ofKind(RowKind.INSERT, 2L, -1L)));
     }
 }
diff --git a/pom.xml b/pom.xml
index 4c10d6b..5c493d2 100644
--- a/pom.xml
+++ b/pom.xml
@@ -48,7 +48,6 @@ under the License.
 
     <properties>
         <flink.version>1.17.0</flink.version>
-        <flink.shaded.version>16.1</flink.shaded.version>
 
         <jackson-bom.version>2.13.4.20221013</jackson-bom.version>
         <junit4.version>4.13.2</junit4.version>
@@ -66,11 +65,6 @@ under the License.
     </properties>
 
     <dependencies>
-        <dependency>
-            <groupId>org.apache.flink</groupId>
-            <artifactId>flink-shaded-force-shading</artifactId>
-            <version>${flink.shaded.version}</version>
-        </dependency>
 
         <!-- Root dependencies for all projects -->
 
diff --git a/tools/maven/checkstyle.xml b/tools/maven/checkstyle.xml
index 2048fd1..7d1f9fd 100644
--- a/tools/maven/checkstyle.xml
+++ b/tools/maven/checkstyle.xml
@@ -224,10 +224,6 @@ This file is based on the checkstyle file of Apache Beam.
                        <property name="illegalPkgs" value="io.netty"/>
                        <message key="import.illegal" value="{0}; Use 
flink-shaded-netty instead."/>
                </module>
-               <module name="IllegalImport">
-                       <property name="illegalPkgs" value="com.google.common"/>
-                       <message key="import.illegal" value="{0}; Use 
flink-shaded-guava instead."/>
-               </module>
 
                <module name="RedundantModifier">
                        <!-- Checks for redundant modifiers on various symbol 
definitions.

Reply via email to