This is an automated email from the ASF dual-hosted git repository.
danny0405 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/master by this push:
new 1e95f14b6c [HUDI-4644] Change default flink profile to 1.15.x (#6445)
1e95f14b6c is described below
commit 1e95f14b6ca144b2569b20b15840c29b00009b95
Author: Danny Chan <[email protected]>
AuthorDate: Sat Aug 20 17:32:26 2022 +0800
[HUDI-4644] Change default flink profile to 1.15.x (#6445)
---
hudi-examples/hudi-examples-flink/pom.xml | 7 ----
hudi-flink-datasource/hudi-flink/pom.xml | 5 ---
.../apache/hudi/sink/ITTestDataStreamWrite.java | 1 -
pom.xml | 39 ++++++----------------
4 files changed, 11 insertions(+), 41 deletions(-)
diff --git a/hudi-examples/hudi-examples-flink/pom.xml
b/hudi-examples/hudi-examples-flink/pom.xml
index a9092411d9..7038cd9d46 100644
--- a/hudi-examples/hudi-examples-flink/pom.xml
+++ b/hudi-examples/hudi-examples-flink/pom.xml
@@ -333,13 +333,6 @@
<scope>test</scope>
<type>test-jar</type>
</dependency>
- <dependency>
- <groupId>org.apache.flink</groupId>
- <artifactId>flink-json</artifactId>
- <version>${flink.version}</version>
- <scope>test</scope>
- <type>test-jar</type>
- </dependency>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-csv</artifactId>
diff --git a/hudi-flink-datasource/hudi-flink/pom.xml
b/hudi-flink-datasource/hudi-flink/pom.xml
index f6bd3a00ac..396b10b11a 100644
--- a/hudi-flink-datasource/hudi-flink/pom.xml
+++ b/hudi-flink-datasource/hudi-flink/pom.xml
@@ -239,11 +239,6 @@
<artifactId>jcommander</artifactId>
<scope>compile</scope>
</dependency>
- <dependency>
- <groupId>com.twitter</groupId>
- <artifactId>bijection-avro_${scala.binary.version}</artifactId>
- <version>0.9.7</version>
- </dependency>
<dependency>
<groupId>joda-time</groupId>
<artifactId>joda-time</artifactId>
diff --git
a/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/sink/ITTestDataStreamWrite.java
b/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/sink/ITTestDataStreamWrite.java
index aa420a433d..4862cda07a 100644
---
a/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/sink/ITTestDataStreamWrite.java
+++
b/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/sink/ITTestDataStreamWrite.java
@@ -246,7 +246,6 @@ public class ITTestDataStreamWrite extends TestLogger {
execEnv.addOperator(pipeline.getTransformation());
if (isMor) {
- Pipelines.clean(conf, pipeline);
Pipelines.compact(conf, pipeline);
}
diff --git a/pom.xml b/pom.xml
index ff1f1e26ea..6489e632b4 100644
--- a/pom.xml
+++ b/pom.xml
@@ -131,20 +131,20 @@
<flink1.15.version>1.15.1</flink1.15.version>
<flink1.14.version>1.14.5</flink1.14.version>
<flink1.13.version>1.13.6</flink1.13.version>
- <flink.version>${flink1.13.version}</flink.version>
- <hudi.flink.module>hudi-flink1.13.x</hudi.flink.module>
- <flink.bundle.version>1.13</flink.bundle.version>
+ <flink.version>${flink1.15.version}</flink.version>
+ <hudi.flink.module>hudi-flink1.15.x</hudi.flink.module>
+ <flink.bundle.version>1.15</flink.bundle.version>
<flink.format.parquet.version>1.12.2</flink.format.parquet.version>
<flink.runtime.artifactId>flink-runtime</flink.runtime.artifactId>
-
<flink.table.runtime.artifactId>flink-table-runtime_${scala.binary.version}</flink.table.runtime.artifactId>
-
<flink.table.planner.artifactId>flink-table-planner_${scala.binary.version}</flink.table.planner.artifactId>
+
<flink.table.runtime.artifactId>flink-table-runtime</flink.table.runtime.artifactId>
+
<flink.table.planner.artifactId>flink-table-planner_2.12</flink.table.planner.artifactId>
<flink.parquet.artifactId>flink-parquet</flink.parquet.artifactId>
<flink.statebackend.rocksdb.artifactId>flink-statebackend-rocksdb</flink.statebackend.rocksdb.artifactId>
<flink.test.utils.artifactId>flink-test-utils</flink.test.utils.artifactId>
<flink.streaming.java.artifactId>flink-streaming-java</flink.streaming.java.artifactId>
<flink.clients.artifactId>flink-clients</flink.clients.artifactId>
<flink.connector.kafka.artifactId>flink-connector-kafka</flink.connector.kafka.artifactId>
-
<flink.hadoop.compatibility.artifactId>flink-hadoop-compatibility_${scala.binary.version}</flink.hadoop.compatibility.artifactId>
+
<flink.hadoop.compatibility.artifactId>flink-hadoop-compatibility_2.12</flink.hadoop.compatibility.artifactId>
<spark31.version>3.1.3</spark31.version>
<spark32.version>3.2.1</spark32.version>
<spark33.version>3.3.0</spark33.version>
@@ -1828,20 +1828,6 @@
<profile>
<id>flink1.15</id>
- <properties>
- <flink.version>${flink1.15.version}</flink.version>
-
<flink.table.runtime.artifactId>flink-table-runtime</flink.table.runtime.artifactId>
- <flink.parquet.artifactId>flink-parquet</flink.parquet.artifactId>
-
<flink.statebackend.rocksdb.artifactId>flink-statebackend-rocksdb</flink.statebackend.rocksdb.artifactId>
-
<flink.test.utils.artifactId>flink-test-utils</flink.test.utils.artifactId>
-
<flink.streaming.java.artifactId>flink-streaming-java</flink.streaming.java.artifactId>
- <flink.clients.artifactId>flink-clients</flink.clients.artifactId>
-
<flink.connector.kafka.artifactId>flink-connector-kafka</flink.connector.kafka.artifactId>
- <!-- 1.15 only supports scala2.12 -->
-
<flink.hadoop.compatibility.artifactId>flink-hadoop-compatibility_2.12</flink.hadoop.compatibility.artifactId>
- <hudi.flink.module>hudi-flink1.15.x</hudi.flink.module>
- <flink.bundle.version>1.15</flink.bundle.version>
- </properties>
<activation>
<property>
<name>flink1.15</name>
@@ -1852,7 +1838,10 @@
<id>flink1.14</id>
<properties>
<flink.version>${flink1.14.version}</flink.version>
+ <hudi.flink.module>hudi-flink1.14.x</hudi.flink.module>
+ <flink.bundle.version>1.14</flink.bundle.version>
<flink.table.runtime.artifactId>flink-table-runtime_${scala.binary.version}</flink.table.runtime.artifactId>
+
<flink.table.planner.artifactId>flink-table-planner_${scala.binary.version}</flink.table.planner.artifactId>
<flink.parquet.artifactId>flink-parquet_${scala.binary.version}</flink.parquet.artifactId>
<flink.statebackend.rocksdb.artifactId>flink-statebackend-rocksdb_${scala.binary.version}</flink.statebackend.rocksdb.artifactId>
<flink.test.utils.artifactId>flink-test-utils_${scala.binary.version}</flink.test.utils.artifactId>
@@ -1860,23 +1849,19 @@
<flink.clients.artifactId>flink-clients_${scala.binary.version}</flink.clients.artifactId>
<flink.connector.kafka.artifactId>flink-connector-kafka_${scala.binary.version}</flink.connector.kafka.artifactId>
<flink.hadoop.compatibility.artifactId>flink-hadoop-compatibility_${scala.binary.version}</flink.hadoop.compatibility.artifactId>
- <hudi.flink.module>hudi-flink1.14.x</hudi.flink.module>
- <flink.bundle.version>1.14</flink.bundle.version>
</properties>
<activation>
- <activeByDefault>true</activeByDefault>
<property>
<name>flink1.14</name>
- <!-- add flink1.14 module to all profile -->
- <value>!disabled</value>
</property>
</activation>
</profile>
<profile>
<id>flink1.13</id>
<properties>
- <flink.scala.bnary.version>2.11</flink.scala.bnary.version>
<flink.version>${flink1.13.version}</flink.version>
+ <hudi.flink.module>hudi-flink1.13.x</hudi.flink.module>
+ <flink.bundle.version>1.13</flink.bundle.version>
<flink.runtime.artifactId>flink-runtime_${scala.binary.version}</flink.runtime.artifactId>
<flink.table.runtime.artifactId>flink-table-runtime-blink_${scala.binary.version}</flink.table.runtime.artifactId>
<flink.table.planner.artifactId>flink-table-planner-blink_${scala.binary.version}</flink.table.planner.artifactId>
@@ -1887,8 +1872,6 @@
<flink.clients.artifactId>flink-clients_${scala.binary.version}</flink.clients.artifactId>
<flink.connector.kafka.artifactId>flink-connector-kafka_${scala.binary.version}</flink.connector.kafka.artifactId>
<flink.hadoop.compatibility.artifactId>flink-hadoop-compatibility_${scala.binary.version}</flink.hadoop.compatibility.artifactId>
- <hudi.flink.module>hudi-flink1.13.x</hudi.flink.module>
- <flink.bundle.version>1.13</flink.bundle.version>
<skipITs>true</skipITs>
</properties>
<activation>