This is an automated email from the ASF dual-hosted git repository.
lresende pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/bahir-flink.git
The following commit(s) were added to refs/heads/master by this push:
new bdb39f3 [BAHIR-207] Add tests for scala 2.12 on travis (#59)
bdb39f3 is described below
commit bdb39f3ba8cc2ed3abbe3c6bbfedc7bda3f3c8af
Author: Joao Boto <[email protected]>
AuthorDate: Thu Jul 4 00:13:51 2019 +0200
[BAHIR-207] Add tests for scala 2.12 on travis (#59)
---
.travis.yml | 29 ++++++-------
flink-connector-activemq/pom.xml | 7 ---
.../streaming/connectors/flume/FlumeSink.java | 6 +--
.../connectors/kudu/KuduOuputFormatTest.java | 6 +--
.../streaming/connectors/kudu/KuduSinkTest.java | 6 +--
flink-connector-redis/pom.xml | 1 +
.../streaming/siddhi/utils/SiddhiTupleFactory.java | 50 +++++++++++-----------
.../flink/streaming/siddhi/SiddhiCEPITCase.java | 2 +-
.../streaming/siddhi/schema/StreamSchemaTest.java | 4 +-
.../flink/streaming/siddhi/source/Event.java | 2 +-
.../siddhi/utils/SiddhiTupleFactoryTest.java | 2 +-
pom.xml | 5 +--
12 files changed, 54 insertions(+), 66 deletions(-)
diff --git a/.travis.yml b/.travis.yml
index aaab50d..fd8c519 100644
--- a/.travis.yml
+++ b/.travis.yml
@@ -28,6 +28,8 @@ cache:
directories:
- $HOME/.m2
+install: true
+
services:
- docker
@@ -36,26 +38,19 @@ jdk:
- openjdk8
env:
- - |
- FLINK_VERSION="1.8.0" SCALA_VERSION="2.11" DOCKER="false"
-
PROJECTS="flink-connector-activemq,flink-connector-akka,flink-connector-influxdb,flink-connector-netty,flink-connector-redis,flink-library-siddhi,flink-connector-kudu"
- - |
- FLINK_VERSION="1.8.0" SCALA_VERSION="2.11" DOCKER="true"
- PROJECTS="flink-connector-flume"
+ - FLINK_VERSION="1.8.0" SCALA_VERSION="2.11"
+ - FLINK_VERSION="1.8.0" SCALA_VERSION="2.11" PROJECT="flink-connector-flume"
DOCKER="true"
+ - FLINK_VERSION="1.8.0" SCALA_VERSION="2.12"
+ - FLINK_VERSION="1.8.0" SCALA_VERSION="2.12" PROJECT="flink-connector-flume"
DOCKER="true"
before_install:
- ./dev/change-scala-version.sh $SCALA_VERSION
-install: true
-
-before_script:
- - if [[ $DOCKER == "true" ]]; then
- docker-compose -f "$PROJECTS/dockers/docker-compose.yml" up -d;
- fi
-
-script: mvn clean verify -pl $PROJECTS -Pscala-$SCALA_VERSION
-Dflink.version=$FLINK_VERSION
-
-after_script:
+script:
- if [[ $DOCKER == "true" ]]; then
- docker-compose -f "$PROJECTS/dockers/docker-compose.yml" down;
+ docker-compose -f "$PROJECT/dockers/docker-compose.yml" up -d;
+ mvn clean verify -pl $PROJECT -Dscala-$SCALA_VERSION
-Dflink.version=$FLINK_VERSION ;
+ docker-compose -f "$PROJECT/dockers/docker-compose.yml" down;
+ else
+ mvn clean verify -Dscala-$SCALA_VERSION -Dflink.version=$FLINK_VERSION ;
fi
diff --git a/flink-connector-activemq/pom.xml b/flink-connector-activemq/pom.xml
index 8f7cdaf..8efbedb 100644
--- a/flink-connector-activemq/pom.xml
+++ b/flink-connector-activemq/pom.xml
@@ -78,13 +78,6 @@ under the License.
<dependency>
<groupId>org.apache.flink</groupId>
- <artifactId>flink-test-utils_${scala.binary.version}</artifactId>
- <version>${flink.version}</version>
- <scope>test</scope>
- </dependency>
-
- <dependency>
- <groupId>org.apache.flink</groupId>
<artifactId>flink-runtime_${scala.binary.version}</artifactId>
<version>${flink.version}</version>
<type>test-jar</type>
diff --git
a/flink-connector-flume/src/main/java/org/apache/flink/streaming/connectors/flume/FlumeSink.java
b/flink-connector-flume/src/main/java/org/apache/flink/streaming/connectors/flume/FlumeSink.java
index 992d19f..13136ba 100644
---
a/flink-connector-flume/src/main/java/org/apache/flink/streaming/connectors/flume/FlumeSink.java
+++
b/flink-connector-flume/src/main/java/org/apache/flink/streaming/connectors/flume/FlumeSink.java
@@ -47,7 +47,7 @@ public class FlumeSink<IN> extends RichSinkFunction<IN> {
private int maxRetryAttempts;
private long waitTimeMs;
private List<IN> incomingList;
- private FlumeEventBuilder eventBuilder;
+ private FlumeEventBuilder<IN> eventBuilder;
private RpcClient client;
public FlumeSink(String clientType, String hostname, int port,
FlumeEventBuilder<IN> eventBuilder) {
@@ -71,7 +71,7 @@ public class FlumeSink<IN> extends RichSinkFunction<IN> {
@Override
public void open(Configuration parameters) throws Exception {
super.open(parameters);
- incomingList = new ArrayList();
+ incomingList = new ArrayList<>();
client = FlumeUtils.getRpcClient(clientType, hostname, port,
batchSize);
}
@@ -104,7 +104,7 @@ public class FlumeSink<IN> extends RichSinkFunction<IN> {
return;
}
toFlushList = incomingList;
- incomingList = new ArrayList();
+ incomingList = new ArrayList<>();
}
for (IN value: toFlushList) {
diff --git
a/flink-connector-kudu/src/test/java/org/apache/flink/streaming/connectors/kudu/KuduOuputFormatTest.java
b/flink-connector-kudu/src/test/java/org/apache/flink/streaming/connectors/kudu/KuduOuputFormatTest.java
index b9aaa40..4e91310 100644
---
a/flink-connector-kudu/src/test/java/org/apache/flink/streaming/connectors/kudu/KuduOuputFormatTest.java
+++
b/flink-connector-kudu/src/test/java/org/apache/flink/streaming/connectors/kudu/KuduOuputFormatTest.java
@@ -45,7 +45,7 @@ public class KuduOuputFormatTest extends KuduDatabase {
public void testNotTableExist() throws IOException {
String masterAddresses = harness.getMasterAddressesAsString();
KuduTableInfo tableInfo =
booksTableInfo(UUID.randomUUID().toString(),false);
- KuduOutputFormat outputFormat = new
KuduOutputFormat<>(masterAddresses, tableInfo, new DefaultSerDe());
+ KuduOutputFormat<KuduRow> outputFormat = new
KuduOutputFormat<>(masterAddresses, tableInfo, new DefaultSerDe());
Assertions.assertThrows(UnsupportedOperationException.class, () ->
outputFormat.open(0,1));
}
@@ -54,7 +54,7 @@ public class KuduOuputFormatTest extends KuduDatabase {
String masterAddresses = harness.getMasterAddressesAsString();
KuduTableInfo tableInfo =
booksTableInfo(UUID.randomUUID().toString(),true);
- KuduOutputFormat outputFormat = new
KuduOutputFormat<>(masterAddresses, tableInfo, new DefaultSerDe())
+ KuduOutputFormat<KuduRow> outputFormat = new
KuduOutputFormat<>(masterAddresses, tableInfo, new DefaultSerDe())
.withStrongConsistency();
outputFormat.open(0,1);
@@ -74,7 +74,7 @@ public class KuduOuputFormatTest extends KuduDatabase {
String masterAddresses = harness.getMasterAddressesAsString();
KuduTableInfo tableInfo =
booksTableInfo(UUID.randomUUID().toString(),true);
- KuduOutputFormat outputFormat = new
KuduOutputFormat<>(masterAddresses, tableInfo, new DefaultSerDe())
+ KuduOutputFormat<KuduRow> outputFormat = new
KuduOutputFormat<>(masterAddresses, tableInfo, new DefaultSerDe())
.withEventualConsistency();
outputFormat.open(0,1);
diff --git
a/flink-connector-kudu/src/test/java/org/apache/flink/streaming/connectors/kudu/KuduSinkTest.java
b/flink-connector-kudu/src/test/java/org/apache/flink/streaming/connectors/kudu/KuduSinkTest.java
index 83e060d..225bf7c 100644
---
a/flink-connector-kudu/src/test/java/org/apache/flink/streaming/connectors/kudu/KuduSinkTest.java
+++
b/flink-connector-kudu/src/test/java/org/apache/flink/streaming/connectors/kudu/KuduSinkTest.java
@@ -58,7 +58,7 @@ public class KuduSinkTest extends KuduDatabase {
public void testNotTableExist() throws IOException {
String masterAddresses = harness.getMasterAddressesAsString();
KuduTableInfo tableInfo =
booksTableInfo(UUID.randomUUID().toString(),false);
- KuduSink sink = new KuduSink<>(masterAddresses, tableInfo, new
DefaultSerDe());
+ KuduSink<KuduRow> sink = new KuduSink<>(masterAddresses, tableInfo,
new DefaultSerDe());
sink.setRuntimeContext(context);
Assertions.assertThrows(UnsupportedOperationException.class, () ->
sink.open(new Configuration()));
}
@@ -68,7 +68,7 @@ public class KuduSinkTest extends KuduDatabase {
String masterAddresses = harness.getMasterAddressesAsString();
KuduTableInfo tableInfo =
booksTableInfo(UUID.randomUUID().toString(),true);
- KuduSink sink = new KuduSink<>(masterAddresses, tableInfo, new
DefaultSerDe())
+ KuduSink<KuduRow> sink = new KuduSink<>(masterAddresses, tableInfo,
new DefaultSerDe())
.withStrongConsistency();
sink.setRuntimeContext(context);
sink.open(new Configuration());
@@ -88,7 +88,7 @@ public class KuduSinkTest extends KuduDatabase {
String masterAddresses = harness.getMasterAddressesAsString();
KuduTableInfo tableInfo =
booksTableInfo(UUID.randomUUID().toString(),true);
- KuduSink sink = new KuduSink<>(masterAddresses, tableInfo, new
DefaultSerDe())
+ KuduSink<KuduRow> sink = new KuduSink<>(masterAddresses, tableInfo,
new DefaultSerDe())
.withEventualConsistency();
sink.setRuntimeContext(context);
sink.open(new Configuration());
diff --git a/flink-connector-redis/pom.xml b/flink-connector-redis/pom.xml
index c9b7a63..46d546b 100644
--- a/flink-connector-redis/pom.xml
+++ b/flink-connector-redis/pom.xml
@@ -78,6 +78,7 @@ under the License.
<plugin>
<groupId>org.apache.felix</groupId>
<artifactId>maven-bundle-plugin</artifactId>
+ <version>4.2.0</version>
<inherited>true</inherited>
<extensions>true</extensions>
</plugin>
diff --git
a/flink-library-siddhi/src/main/java/org/apache/flink/streaming/siddhi/utils/SiddhiTupleFactory.java
b/flink-library-siddhi/src/main/java/org/apache/flink/streaming/siddhi/utils/SiddhiTupleFactory.java
index 88c15eb..6e96345 100644
---
a/flink-library-siddhi/src/main/java/org/apache/flink/streaming/siddhi/utils/SiddhiTupleFactory.java
+++
b/flink-library-siddhi/src/main/java/org/apache/flink/streaming/siddhi/utils/SiddhiTupleFactory.java
@@ -61,55 +61,55 @@ public class SiddhiTupleFactory {
case 0:
return setTupleValue(new Tuple0(), row);
case 1:
- return setTupleValue(new Tuple1(), row);
+ return setTupleValue(new Tuple1<>(), row);
case 2:
- return setTupleValue(new Tuple2(), row);
+ return setTupleValue(new Tuple2<>(), row);
case 3:
- return setTupleValue(new Tuple3(), row);
+ return setTupleValue(new Tuple3<>(), row);
case 4:
- return setTupleValue(new Tuple4(), row);
+ return setTupleValue(new Tuple4<>(), row);
case 5:
- return setTupleValue(new Tuple5(), row);
+ return setTupleValue(new Tuple5<>(), row);
case 6:
- return setTupleValue(new Tuple6(), row);
+ return setTupleValue(new Tuple6<>(), row);
case 7:
- return setTupleValue(new Tuple7(), row);
+ return setTupleValue(new Tuple7<>(), row);
case 8:
- return setTupleValue(new Tuple8(), row);
+ return setTupleValue(new Tuple8<>(), row);
case 9:
- return setTupleValue(new Tuple9(), row);
+ return setTupleValue(new Tuple9<>(), row);
case 10:
- return setTupleValue(new Tuple10(), row);
+ return setTupleValue(new Tuple10<>(), row);
case 11:
- return setTupleValue(new Tuple11(), row);
+ return setTupleValue(new Tuple11<>(), row);
case 12:
- return setTupleValue(new Tuple12(), row);
+ return setTupleValue(new Tuple12<>(), row);
case 13:
- return setTupleValue(new Tuple13(), row);
+ return setTupleValue(new Tuple13<>(), row);
case 14:
- return setTupleValue(new Tuple14(), row);
+ return setTupleValue(new Tuple14<>(), row);
case 15:
- return setTupleValue(new Tuple15(), row);
+ return setTupleValue(new Tuple15<>(), row);
case 16:
- return setTupleValue(new Tuple16(), row);
+ return setTupleValue(new Tuple16<>(), row);
case 17:
- return setTupleValue(new Tuple17(), row);
+ return setTupleValue(new Tuple17<>(), row);
case 18:
- return setTupleValue(new Tuple18(), row);
+ return setTupleValue(new Tuple18<>(), row);
case 19:
- return setTupleValue(new Tuple19(), row);
+ return setTupleValue(new Tuple19<>(), row);
case 20:
- return setTupleValue(new Tuple20(), row);
+ return setTupleValue(new Tuple20<>(), row);
case 21:
- return setTupleValue(new Tuple21(), row);
+ return setTupleValue(new Tuple21<>(), row);
case 22:
- return setTupleValue(new Tuple22(), row);
+ return setTupleValue(new Tuple22<>(), row);
case 23:
- return setTupleValue(new Tuple23(), row);
+ return setTupleValue(new Tuple23<>(), row);
case 24:
- return setTupleValue(new Tuple24(), row);
+ return setTupleValue(new Tuple24<>(), row);
case 25:
- return setTupleValue(new Tuple25(), row);
+ return setTupleValue(new Tuple25<>(), row);
default:
throw new IllegalArgumentException("Too long row: " +
row.length + ", unable to convert to Tuple");
}
diff --git
a/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/SiddhiCEPITCase.java
b/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/SiddhiCEPITCase.java
index 651288f..9b38539 100755
---
a/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/SiddhiCEPITCase.java
+++
b/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/SiddhiCEPITCase.java
@@ -253,7 +253,7 @@ public class SiddhiCEPITCase extends AbstractTestBase
implements Serializable {
DataStream<Event> input1 = env.addSource(new RandomEventSource(5),
"input1");
DataStream<Event> input2 = env.addSource(new RandomEventSource(5),
"input2");
- DataStream<? extends Map> output = SiddhiCEP
+ DataStream<? extends Map<?,?>> output = SiddhiCEP
.define("inputStream1", input1.keyBy("id"), "id", "name", "price",
"timestamp")
.union("inputStream2", input2.keyBy("id"), "id", "name", "price",
"timestamp")
.cql(
diff --git
a/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/schema/StreamSchemaTest.java
b/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/schema/StreamSchemaTest.java
index b9dcac7..73888fe 100644
---
a/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/schema/StreamSchemaTest.java
+++
b/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/schema/StreamSchemaTest.java
@@ -41,8 +41,8 @@ public class StreamSchemaTest {
@Test
public void testStreamSchemaWithTuple() {
- TypeInformation<Tuple4> typeInfo = Types.TUPLE(Types.INT, Types.LONG,
Types.STRING, Types.DOUBLE);
- StreamSchema<Tuple4> schema = new StreamSchema<>(typeInfo, "id",
"timestamp", "name", "price");
+ TypeInformation<Tuple4<Integer,Long,String,Double>> typeInfo =
Types.TUPLE(Types.INT, Types.LONG, Types.STRING, Types.DOUBLE);
+ StreamSchema<Tuple4<Integer,Long,String,Double>> schema = new
StreamSchema<>(typeInfo, "id", "timestamp", "name", "price");
assertEquals(Tuple4.class, schema.getTypeInfo().getTypeClass());
assertEquals(4, schema.getFieldIndexes().length);
assertEquals(Tuple4.class, schema.getTypeInfo().getTypeClass());
diff --git
a/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/source/Event.java
b/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/source/Event.java
index 357e1d2..125b771 100644
---
a/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/source/Event.java
+++
b/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/source/Event.java
@@ -83,7 +83,7 @@ public class Event {
}
public static TypeSerializer<Event> createTypeSerializer() {
- TypeInformation<Event> typeInformation = (TypeInformation<Event>)
TypeExtractor.createTypeInfo(Event.class);
+ TypeInformation<Event> typeInformation =
TypeExtractor.createTypeInfo(Event.class);
return typeInformation.createSerializer(new ExecutionConfig());
}
diff --git
a/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/utils/SiddhiTupleFactoryTest.java
b/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/utils/SiddhiTupleFactoryTest.java
index 4753a3f..2d8f04b 100644
---
a/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/utils/SiddhiTupleFactoryTest.java
+++
b/flink-library-siddhi/src/test/java/org/apache/flink/streaming/siddhi/utils/SiddhiTupleFactoryTest.java
@@ -27,7 +27,7 @@ public class SiddhiTupleFactoryTest {
@Test
public void testConvertObjectArrayToTuple() {
Object[] row = new Object[]{1, "message", 1234567L, true, new
Object()};
- Tuple5 tuple5 = SiddhiTupleFactory.newTuple(row);
+ Tuple5<Integer,String,Long,Boolean,Object> tuple5 =
SiddhiTupleFactory.newTuple(row);
assertEquals(5, tuple5.getArity());
assertArrayEquals(row, new Object[]{
tuple5.f0,
diff --git a/pom.xml b/pom.xml
index 9f735d2..111e8a9 100644
--- a/pom.xml
+++ b/pom.xml
@@ -88,8 +88,7 @@
<!-- General project dependencies version -->
<java.version>1.8</java.version>
- <scala.version>2.11.8</scala.version>
-
+ <scala.version>2.11.12</scala.version>
<scala.binary.version>2.11</scala.binary.version>
<slf4j.version>1.7.16</slf4j.version>
@@ -703,7 +702,7 @@
<id>scala-2.11</id>
<activation>
<property>
- <name>!scala-2.12</name>
+ <name>scala-2.11</name>
</property>
</activation>
<properties>