This is an automated email from the ASF dual-hosted git repository.

dianfu pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/flink-playgrounds.git


The following commit(s) were added to refs/heads/master by this push:
     new 6e440d7  [FLINK-23100] Update pyflink-walkthrough playground for Flink 
1.13 (#24)
6e440d7 is described below

commit 6e440d7c508f57482d4a514456bf10418f39b6ac
Author: David Anderson <[email protected]>
AuthorDate: Wed Jun 23 10:38:46 2021 +0200

    [FLINK-23100] Update pyflink-walkthrough playground for Flink 1.13 (#24)
---
 pyflink-walkthrough/Dockerfile         | 10 +++++-----
 pyflink-walkthrough/README.md          |  2 +-
 pyflink-walkthrough/docker-compose.yml |  6 +++---
 3 files changed, 9 insertions(+), 9 deletions(-)

diff --git a/pyflink-walkthrough/Dockerfile b/pyflink-walkthrough/Dockerfile
index 3612785..ae9ef03 100644
--- a/pyflink-walkthrough/Dockerfile
+++ b/pyflink-walkthrough/Dockerfile
@@ -20,8 +20,8 @@
 # Build PyFlink Playground Image
 ###############################################################################
 
-FROM apache/flink:1.12.1-scala_2.11-java8
-ARG FLINK_VERSION=1.12.1
+FROM apache/flink:1.13.1-scala_2.12-java8
+ARG FLINK_VERSION=1.13.1
 
 # Install pyflink
 RUN set -ex; \
@@ -33,14 +33,14 @@ RUN set -ex; \
   ln -s /usr/bin/pip3 /usr/bin/pip; \
   apt-get update; \
   python -m pip install --upgrade pip; \
-  pip install apache-flink==1.12.1; \
+  pip install apache-flink==1.13.1; \
   pip install kafka-python;
 
 
 # Download connector libraries
 RUN wget -P /opt/flink/lib/ 
https://repo.maven.apache.org/maven2/org/apache/flink/flink-json/${FLINK_VERSION}/flink-json-${FLINK_VERSION}.jar;
 \
-    wget -P /opt/flink/lib/ 
https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-kafka_2.11/${FLINK_VERSION}/flink-sql-connector-kafka_2.11-${FLINK_VERSION}.jar;
 \
-    wget -P /opt/flink/lib/ 
https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7_2.11/${FLINK_VERSION}/flink-sql-connector-elasticsearch7_2.11-${FLINK_VERSION}.jar;
+    wget -P /opt/flink/lib/ 
https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-kafka_2.12/${FLINK_VERSION}/flink-sql-connector-kafka_2.12-${FLINK_VERSION}.jar;
 \
+    wget -P /opt/flink/lib/ 
https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-elasticsearch7_2.12/${FLINK_VERSION}/flink-sql-connector-elasticsearch7_2.12-${FLINK_VERSION}.jar;
 
 
 RUN echo "taskmanager.memory.jvm-metaspace.size: 512m" >> 
/opt/flink/conf/flink-conf.yaml;
diff --git a/pyflink-walkthrough/README.md b/pyflink-walkthrough/README.md
index fa2f3e5..c3c447f 100644
--- a/pyflink-walkthrough/README.md
+++ b/pyflink-walkthrough/README.md
@@ -5,7 +5,7 @@
 In this playground, you will learn how to build and run an end-to-end PyFlink 
pipeline for data analytics, covering the following steps:
 
 * Reading data from a Kafka source;
-* Creating data using a 
[UDF](https://ci.apache.org/projects/flink/flink-docs-release-1.12/dev/python/table-api-users-guide/udfs/python_udfs.html);
+* Creating data using a 
[UDF](https://ci.apache.org/projects/flink/flink-docs-release-1.13/dev/python/table-api-users-guide/udfs/python_udfs.html);
 * Performing a simple aggregation over the source data;
 * Writing the results to Elasticsearch and visualizing them in Kibana.
 
diff --git a/pyflink-walkthrough/docker-compose.yml 
b/pyflink-walkthrough/docker-compose.yml
index ef07632..a8369fd 100644
--- a/pyflink-walkthrough/docker-compose.yml
+++ b/pyflink-walkthrough/docker-compose.yml
@@ -20,7 +20,7 @@ version: '2.1'
 services:
   jobmanager:
     build: .
-    image: pyflink/pyflink:1.12.1-scala_2.11
+    image: pyflink/pyflink:1.13.1-scala_2.12
     volumes:
       - .:/opt/pyflink-walkthrough
     hostname: "jobmanager"
@@ -32,7 +32,7 @@ services:
     environment:
       - JOB_MANAGER_RPC_ADDRESS=jobmanager
   taskmanager:
-    image: pyflink/pyflink:1.12.1-scala_2.11
+    image: pyflink/pyflink:1.13.1-scala_2.12
     volumes:
     - .:/opt/pyflink-walkthrough
     expose:
@@ -52,7 +52,7 @@ services:
   kafka:
     image: wurstmeister/kafka:2.12-2.2.1
     ports:
-      - "9092"
+      - "9092:9092"
     depends_on:
       - zookeeper
     environment:

Reply via email to