[jira] [Commented] (FLINK-19004) Fail to call Hive percentile function together with distinct aggregate call
[ https://issues.apache.org/jira/browse/FLINK-19004?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181194#comment-17181194 ] Rui Li commented on FLINK-19004: It seems the {{MIN}} is introduced during {{FlinkAggregateExpandDistinctAggregatesRule}} rewrites the original aggregate with grouping sets. It generates a {{MIN}} for the non-distinct aggregate calls. And since the return type of {{percentile}} is an array, thus the error. > Fail to call Hive percentile function together with distinct aggregate call > --- > > Key: FLINK-19004 > URL: https://issues.apache.org/jira/browse/FLINK-19004 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive, Table SQL / Planner >Reporter: Rui Li >Priority: Major > > The following test case would fail: > {code} > @Test > public void test() throws Exception { > TableEnvironment tableEnv = getTableEnvWithHiveCatalog(); > tableEnv.unloadModule("core"); > tableEnv.loadModule("hive", new HiveModule()); > tableEnv.loadModule("core", CoreModule.INSTANCE); > tableEnv.executeSql("create table src(x int,y int)"); > tableEnv.executeSql("select count(distinct > y),`percentile`(y,`array`(0.5,0.99)) from src group by x").collect(); > } > {code} > The error is: > {noformat} > org.apache.flink.table.api.TableException: Cannot generate a valid execution > plan for the given query: > FlinkLogicalLegacySink(name=[collect], fields=[EXPR$0, EXPR$1]) > +- FlinkLogicalCalc(select=[EXPR$0, EXPR$1]) >+- FlinkLogicalAggregate(group=[{0}], EXPR$0=[COUNT($1) FILTER $3], > EXPR$1=[MIN($2) FILTER $4]) > +- FlinkLogicalCalc(select=[x, y, EXPR$1, =(CASE(=($e, 0:BIGINT), > 0:BIGINT, 1:BIGINT), 0) AS $g_0, =(CASE(=($e, 0:BIGINT), 0:BIGINT, 1:BIGINT), > 1) AS $g_1]) > +- FlinkLogicalAggregate(group=[{0, 1, 3}], EXPR$1=[percentile($4, > $2)]) > +- FlinkLogicalExpand(projects=[x, y, $f2, $e, y_0]) >+- FlinkLogicalCalc(select=[x, y, array(0.5:DECIMAL(2, 1), > 0.99:DECIMAL(3, 2)) AS $f2]) > +- FlinkLogicalLegacyTableSourceScan(table=[[test-catalog, > default, src, source: [HiveTableSource(x, y) TablePath: default.src, > PartitionPruned: false, PartitionNums: null]]], fields=[x, y]) > Min aggregate function does not support type: ''ARRAY''. > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-18850) Add late records dropped metric for row time over windows
[ https://issues.apache.org/jira/browse/FLINK-18850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181190#comment-17181190 ] ZhuShang commented on FLINK-18850: -- sorry,i get it. > Add late records dropped metric for row time over windows > - > > Key: FLINK-18850 > URL: https://issues.apache.org/jira/browse/FLINK-18850 > Project: Flink > Issue Type: Improvement > Components: Table SQL / Runtime >Reporter: Benchao Li >Priority: Major > Labels: Starter > > Currently all the row time over windows in blink planner runtime discards > late records silently, it would be good to have a metric about the late > records dropping. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] danny0405 commented on a change in pull request #13050: [FLINK-18750][table] SqlValidatorException thrown when select from a …
danny0405 commented on a change in pull request #13050: URL: https://github.com/apache/flink/pull/13050#discussion_r473975682 ## File path: flink-table/flink-table-planner-blink/src/main/java/org/apache/flink/table/planner/utils/Expander.java ## @@ -0,0 +1,126 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.table.planner.utils; + +import org.apache.flink.table.planner.calcite.FlinkPlannerImpl; + +import org.apache.flink.shaded.guava18.com.google.common.collect.ImmutableMap; + +import org.apache.calcite.sql.SqlIdentifier; +import org.apache.calcite.sql.SqlNode; +import org.apache.calcite.sql.parser.SqlParser; +import org.apache.calcite.sql.parser.SqlParserPos; +import org.apache.calcite.sql.util.SqlBasicVisitor; +import org.apache.calcite.sql.util.SqlShuttle; + +import java.util.HashMap; +import java.util.Map; +import java.util.Objects; +import java.util.function.Function; + +/** + * Utility that expand SQL identifiers from a SQL query. + * + * Simple use: + * + * + * final String sql = + * "select ename from emp where deptno 10"; + * final Expander.Expanded expanded = + * Expander.create(planner).expanded(sql); + * print(expanded); // "select `emp`.`ename` from `catalog`.`db`.`emp` where `emp`.`deptno` 10" + * + * + * Calling {@link Expanded#toString()} generates a string that is similar to + * SQL where a user has manually converted all identifiers as expanded, and + * which could then be persisted as expanded query of a Catalog view. + * + * For more advanced formatting, use {@link Expanded#substitute(Function)}. + * + * Adjust {@link SqlParser.Config} to use a different parser or parsing options. + */ +public class Expander { Review comment: Yes, the `BridgingSqlFunction` did the function id expanding out of the scope of the `SqlValidator`, so in the code base of this patch, it does not work because `BridgingSqlFunction` id position was always `SqlParserPos.ZERO` so it can not match the original identifier. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] xccui commented on a change in pull request #13089: [FLINK-18813][docs-zh] Translate the 'Local Installation' page of 'Try Flink' into Chinese
xccui commented on a change in pull request #13089: URL: https://github.com/apache/flink/pull/13089#discussion_r473944845 ## File path: docs/try-flink/local_installation.zh.md ## @@ -26,36 +26,35 @@ under the License. {% if site.version contains "SNAPSHOT" %} - NOTE: The Apache Flink community only publishes official builds for - released versions of Apache Flink. + 注意:Apache Flink 社区只发布 Apache Flink 的 release 版本。 - Since you are currently looking at the latest SNAPSHOT - version of the documentation, all version references below will not work. - Please switch the documentation to the latest released version via the release picker which you - find on the left side below the menu. + 由于你当前正在查看的是文档最新的 SNAPSHOT 版本,因此相关内容会被隐藏。请通过左侧菜单底部的版本选择将文档切换到最新的 release 版本。 {% else %} -Follow these few steps to download the latest stable versions and get started. +请按照以下几个步骤下载最新的稳定版本并开始使用。 -## Step 1: Download + -To be able to run Flink, the only requirement is to have a working __Java 8 or 11__ installation. -You can check the correct installation of Java by issuing the following command: +## 步骤 1:下载 + +为了能够运行 Flink,唯一的要求就是安装有效的 __Java 8 或者 Java 11__。你可以通过运行以下命令来检查 Java 的正确安装。 Review comment: 为了运行Flink,只需提前安装好 __Java 8 或者 Java 11__ 。你可以通过以下命令来检查 Java 是否已经安装正确。 ## File path: docs/try-flink/local_installation.zh.md ## @@ -26,36 +26,35 @@ under the License. {% if site.version contains "SNAPSHOT" %} - NOTE: The Apache Flink community only publishes official builds for - released versions of Apache Flink. + 注意:Apache Flink 社区只发布 Apache Flink 的 release 版本。 - Since you are currently looking at the latest SNAPSHOT - version of the documentation, all version references below will not work. - Please switch the documentation to the latest released version via the release picker which you - find on the left side below the menu. + 由于你当前正在查看的是文档最新的 SNAPSHOT 版本,因此相关内容会被隐藏。请通过左侧菜单底部的版本选择将文档切换到最新的 release 版本。 {% else %} -Follow these few steps to download the latest stable versions and get started. +请按照以下几个步骤下载最新的稳定版本并开始使用。 Review comment: Remove “并” ## File path: docs/try-flink/local_installation.zh.md ## @@ -64,10 +63,11 @@ Starting standalonesession daemon on host. Starting taskexecutor daemon on host. {% endhighlight %} -## Step 3: Submit a Job + -Releases of Flink come with a number of example Jobs. -You can quickly deploy one of these applications to the running cluster. +## 步骤 3:提交作业(Job) + +Flink 的 Releases 附带了许多的示例作业。你可以将这些应用程序之一快速部署到正在运行的集群。 Review comment: Flink 的 Releases 附带了许多的示例作业。你可以任意选择一个,快速部署到已运行的集群上。 ## File path: docs/try-flink/local_installation.zh.md ## @@ -80,11 +80,13 @@ $ tail log/flink-*-taskexecutor-*.out (be,2) {% endhighlight %} -Additionally, you can check Flink's [Web UI](http://localhost:8080) to monitor the status of the Cluster and running Job. +另外,你可以查看 Flink 的 [Web UI](http://localhost:8080) 来监视集群的状态和正在运行的作业。 Review comment: 查看 -> 通过 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13208: [FLINK-18676] [FileSystem] Bump s3 aws version
flinkbot edited a comment on pull request #13208: URL: https://github.com/apache/flink/pull/13208#issuecomment-677651717 ## CI report: * 3c1d30e68c960e1009b93d51582053ccf67ceb37 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5746) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #12880: [FLINK-18555][table sql/api] Make TableConfig options can be configur…
flinkbot edited a comment on pull request #12880: URL: https://github.com/apache/flink/pull/12880#issuecomment-657437173 ## CI report: * 72e7dcb66f3a2c6eda4899d51ddd6420d2d5822b Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5744) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13207: [FLINK-18952][python][docs] Add 10 minutes to DataStream API documentation
flinkbot edited a comment on pull request #13207: URL: https://github.com/apache/flink/pull/13207#issuecomment-677643498 ## CI report: * 371eaa2083054d66ef256a9a0ef7a90e116fee0d Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5745) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #13208: [FLINK-18676] [FileSystem] Bump s3 aws version
flinkbot commented on pull request #13208: URL: https://github.com/apache/flink/pull/13208#issuecomment-677651717 ## CI report: * 3c1d30e68c960e1009b93d51582053ccf67ceb37 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-18850) Add late records dropped metric for row time over windows
[ https://issues.apache.org/jira/browse/FLINK-18850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181158#comment-17181158 ] Benchao Li commented on FLINK-18850: [~ZhuShang] Seems that you misunderstand this issue. What I want to do in this issue is to add a metric for over window, not the window operator. > Add late records dropped metric for row time over windows > - > > Key: FLINK-18850 > URL: https://issues.apache.org/jira/browse/FLINK-18850 > Project: Flink > Issue Type: Improvement > Components: Table SQL / Runtime >Reporter: Benchao Li >Priority: Major > Labels: Starter > > Currently all the row time over windows in blink planner runtime discards > late records silently, it would be good to have a metric about the late > records dropping. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19006) project transformation does not support conversion to Tuple25 type
ming li created FLINK-19006: --- Summary: project transformation does not support conversion to Tuple25 type Key: FLINK-19006 URL: https://issues.apache.org/jira/browse/FLINK-19006 Project: Flink Issue Type: Improvement Components: API / DataStream Affects Versions: 1.11.1 Reporter: ming li In the {{DataStream#project}} method, it will judge whether the length of {{fieldIndexes}} is between 1 and {{Tuple.MAX_ARITY-1}}, and then call {{projectTupleXX}} according to the length of {{fieldIndexes}}. This limits the maximum length of {{Tuple}} to 24. {code:java} protected StreamProjection(DataStream dataStream, int[] fieldIndexes) { if (!dataStream.getType().isTupleType()) { throw new RuntimeException("Only Tuple DataStreams can be projected"); } if (fieldIndexes.length == 0) { throw new IllegalArgumentException("project() needs to select at least one (1) field."); } else if (fieldIndexes.length > Tuple.MAX_ARITY - 1) { throw new IllegalArgumentException( "project() may select only up to (" + (Tuple.MAX_ARITY - 1) + ") fields."); } int maxFieldIndex = (dataStream.getType()).getArity(); for (int i = 0; i < fieldIndexes.length; i++) { Preconditions.checkElementIndex(fieldIndexes[i], maxFieldIndex); } this.dataStream = dataStream; this.fieldIndexes = fieldIndexes; }{code} This problem also appears in {{ProjectOperator}}. {code:java} public Projection(DataSet ds, int[] fieldIndexes) { if (!(ds.getType() instanceof TupleTypeInfo)) { throw new UnsupportedOperationException("project() can only be applied to DataSets of Tuples."); } if (fieldIndexes.length == 0) { throw new IllegalArgumentException("project() needs to select at least one (1) field."); } else if (fieldIndexes.length > Tuple.MAX_ARITY - 1) { throw new IllegalArgumentException( "project() may select only up to (" + (Tuple.MAX_ARITY - 1) + ") fields."); } int maxFieldIndex = ds.getType().getArity(); for (int fieldIndexe : fieldIndexes) { Preconditions.checkElementIndex(fieldIndexe, maxFieldIndex); } this.ds = ds; this.fieldIndexes = fieldIndexes; }{code} I think the length we limit should be 1 to {{Tuple.MAX_ARITY}} instead of 1 to {{Tuple.MAX_ARITY-1}}. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #13207: [FLINK-18952][python][docs] Add 10 minutes to DataStream API documentation
flinkbot commented on pull request #13207: URL: https://github.com/apache/flink/pull/13207#issuecomment-677643498 ## CI report: * 371eaa2083054d66ef256a9a0ef7a90e116fee0d UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13205: [FLINK-17330[runtime] Merge cyclic dependent pipelined regions into one region
flinkbot edited a comment on pull request #13205: URL: https://github.com/apache/flink/pull/13205#issuecomment-677463515 ## CI report: * e11ffec55b9151857069d64c806b47cc98d9679d Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5739) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #13208: [FLINK-18676] [FileSystem] Bump s3 aws version
flinkbot commented on pull request #13208: URL: https://github.com/apache/flink/pull/13208#issuecomment-677643552 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit 3c1d30e68c960e1009b93d51582053ccf67ceb37 (Thu Aug 20 12:47:30 UTC 2020) **Warnings:** * **1 pom.xml files were touched**: Check for build and licensing issues. * No documentation files were touched! Remember to keep the Flink docs up to date! * **This pull request references an unassigned [Jira ticket](https://issues.apache.org/jira/browse/FLINK-18676).** According to the [code contribution guide](https://flink.apache.org/contributing/contribute-code.html), tickets need to be assigned before starting with the implementation work. Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-18676) Update version of aws to support use of default constructor of "WebIdentityTokenCredentialsProvider"
[ https://issues.apache.org/jira/browse/FLINK-18676?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-18676: --- Labels: pull-request-available (was: ) > Update version of aws to support use of default constructor of > "WebIdentityTokenCredentialsProvider" > > > Key: FLINK-18676 > URL: https://issues.apache.org/jira/browse/FLINK-18676 > Project: Flink > Issue Type: Improvement > Components: Connectors / FileSystem >Affects Versions: 1.11.0 >Reporter: Ravi Bhushan Ratnakar >Priority: Minor > Labels: pull-request-available > > *Background:* > I am using Flink 1.11.0 on kubernetes platform. To give access of aws > services to taskmanager/jobmanager, we are using "IAM Roles for Service > Accounts" . I have configured below property in flink-conf.yaml to use > credential provider. > fs.s3a.aws.credentials.provider: > com.amazonaws.auth.WebIdentityTokenCredentialsProvider > > *Issue:* > When taskmanager/jobmanager is starting up, during this it complains that > "WebIdentityTokenCredentialsProvider" doesn't have "public constructor" and > container doesn't come up. > > *Solution:* > Currently the above credential's class is being used from > "*flink-s3-fs-hadoop"* which gets "aws-java-sdk-core" dependency from > "*flink-s3-fs-base*". In *"flink-s3-fs-base",* version of aws is 1.11.754 . > The support of default constructor for "WebIdentityTokenCredentialsProvider" > is provided from aws version 1.11.788 and onward. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-18938) Throw better exception message for quering sink-only connector
[ https://issues.apache.org/jira/browse/FLINK-18938?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181156#comment-17181156 ] liufangliang commented on FLINK-18938: -- [~jark] I understand what you mean. when we are quering a sink-only connector , if we don't have a dependency jar in the classpath ,return the following excception message: {code:java} Caused by: org.apache.flink.table.api.ValidationException: Could not find any factory for identifier 'elasticsearch-7' that implements 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. {code} if we have a dependency jar in the classpath ,return the following excception message: {code:java} Caused by: org.apache.flink.table.api.ValidationException: The connector named 'elasticsearch-7' is only supported as sink,can n't be used as a source. {code} Bisides,Tips(support as sink or source) can follow the identifier,For example: {code:java} Available factory identifiers are: datagen (source,sink) elasticsearch-7 (sink-only) test-connector (source-only) {code} What do you think of ? > Throw better exception message for quering sink-only connector > -- > > Key: FLINK-18938 > URL: https://issues.apache.org/jira/browse/FLINK-18938 > Project: Flink > Issue Type: Improvement >Reporter: Jark Wu >Priority: Major > > Currently, if we are quering a sink-only connector, for example: {{SELECT * > FROM elasticsearch_sink}}, a following exception will be thrown: > {code} > Caused by: org.apache.flink.table.api.ValidationException: Could not find any > factory for identifier 'elasticsearch-7' that implements > 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. > Available factory identifiers are: > datagen > {code} > The above exception is very misleading, it sounds like that the elasticsearch > jar is not loaded, however the elasticsearch jar is in the lib directory of > Flink cluster. > We can improve the exception that explicitly telling users the found > connector only support as sink, can't be used as a source. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] batas opened a new pull request #13208: [FLINK-18676] [FileSystem] Bump s3 aws version
batas opened a new pull request #13208: URL: https://github.com/apache/flink/pull/13208 ## What is the purpose of the change This pull request bumps AWS SDK version to minimal version supported WebIdentityTokenCredentialsProvider in s3-fs-package, which is described in an issue. ## Brief change log - flink-s3-fs-base - bumped aws dependency ## Verifying this change This change is a trivial rework / code cleanup without any test coverage. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (yes) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (don't know) - The runtime per-record code paths (performance sensitive): (don't know) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (don't know) - The S3 file system connector: (yes) ## Documentation - Does this pull request introduce a new feature? (no) - If yes, how is the feature documented? (not applicable) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink-web] Myasuka commented on a change in pull request #366: Add Apache Flink release 1.10.2
Myasuka commented on a change in pull request #366: URL: https://github.com/apache/flink-web/pull/366#discussion_r473940917 ## File path: _posts/2020-08-11-release-1.10.2.md ## @@ -0,0 +1,199 @@ +--- +layout: post +title: "Apache Flink 1.10.2 Released" +date: 2020-08-11 18:00:00 +categories: news +authors: +- Zhu Zhu: + name: "Zhu Zhu" + twitter: "zhuzhv" +--- + +The Apache Flink community released the second bugfix version of the Apache Flink 1.10 series. + +This release includes 73 fixes and minor improvements for Flink 1.10.1. The list below includes a detailed list of all fixes and improvements. + +We highly recommend all users to upgrade to Flink 1.10.2. + +Updated Maven dependencies: + +```xml + + org.apache.flink + flink-java + 1.10.2 + + + org.apache.flink + flink-streaming-java_2.11 + 1.10.2 + + + org.apache.flink + flink-clients_2.11 + 1.10.2 + +``` + +You can find the binaries on the updated [Downloads page]({{ site.baseurl }}/downloads.html). + +List of resolved issues: + +Sub-task + + +[FLINK-15836] - Throw fatal error in KubernetesResourceManager when the pods watcher is closed with exception + +[FLINK-16160] - Schema#proctime and Schema#rowtime dont work in TableEnvironment#connect code path + + + +Bug + + +[FLINK-13689] - Rest High Level Client for Elasticsearch6.x connector leaks threads if no connection could be established + +[FLINK-14369] - KafkaProducerAtLeastOnceITCaseKafkaProducerTestBase.testOneToOneAtLeastOnceCustomOperator fails on Travis + +[FLINK-14836] - Unable to set yarn container number for scala shell in yarn mode + +[FLINK-14894] - HybridOffHeapUnsafeMemorySegmentTest#testByteBufferWrap failed on Travis + +[FLINK-15758] - Investigate potential out-of-memory problems due to managed unsafe memory allocation + +[FLINK-15849] - Update SQL-CLIENT document from type to data-type + +[FLINK-16309] - ElasticSearch 7 connector is missing in SQL connector list + +[FLINK-16346] - BlobsCleanupITCase.testBlobServerCleanupCancelledJob fails on Travis + +[FLINK-16432] - Building Hive connector gives problems + +[FLINK-16451] - Fix IndexOutOfBoundsException for DISTINCT AGG with constants + +[FLINK-16510] - Task manager safeguard shutdown may not be reliable + +[FLINK-17092] - Pyflink test BlinkStreamDependencyTests is instable + +[FLINK-17322] - Enable latency tracker would corrupt the broadcast state + +[FLINK-17420] - Cannot alias Tuple and Row fields when converting DataStream to Table + +[FLINK-17466] - toRetractStream doesnt work correctly with Pojo conversion class + +[FLINK-17555] - docstring of pyflink.table.descriptors.FileSystem:1:duplicate object description of pyflink.table.descriptors.FileSystem + +[FLINK-17558] - Partitions are released in TaskExecutor Main Thread + +[FLINK-17562] - POST /jars/:jarid/plan is not working + +[FLINK-17578] - Union of 2 SideOutputs behaviour incorrect + +[FLINK-17639] - Document which FileSystems are supported by the StreamingFileSink + +[FLINK-17643] - LaunchCoordinatorTest fails + +[FLINK-17700] - The callback client of JavaGatewayServer should run in a daemon thread + +[FLINK-17744] - StreamContextEnvironment#execute cannot be call JobListener#onJobExecuted + +[FLINK-17763] - No log files when starting scala-shell + +[FLINK-17788] - scala shell in yarn mode is broken + +[FLINK-17800] - RocksDB optimizeForPointLookup results in missing time windows + +[FLINK-17801] - TaskExecutorTest.testHeartbeatTimeoutWithResourceManager timeout + +[FLINK-17809] - BashJavaUtil script logic does not work for paths with spaces + +[FLINK-17822] - Nightly Flink CLI end-to-end test failed with JavaGcCleanerWrapper$PendingCleanersRunner cannot access class jdk.internal.misc.SharedSecrets in Java 11 + +[FLINK-17870] - dependent jars are missing to be shipped to cluster in scala shell + +[FLINK-17891] - FlinkYarnSessionCli sets wrong execution.target type + +[FLINK-17959] - Exception: CANCELLED: call already cancelled is thrown when run python udf + +[FLINK-18008] - HistoryServer does not log environment information on startup + +[FLINK-18012] - Deactivate slot timeout if TaskSlotTable.tryMarkSlotActive is called + +[FLINK-18035] - Executors#newCachedThreadPool could not work as expected + +[FLINK-18045] - Fix Kerberos credentials checking to unblock Flink on secured MapR + +[FLINK-18048] - --host option could not take effect for standalone application cluster + +[FLINK-18097] - History server doesnt clean all job json files + +[FLINK-18168] - Error results when use UDAF with Object Array return type + +[FLINK-18223] -
[jira] [Updated] (FLINK-18800) Avro serialization schema doesn't support Kafka key/value serialization
[ https://issues.apache.org/jira/browse/FLINK-18800?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mohammad Hossein Gerami updated FLINK-18800: Priority: Critical (was: Major) > Avro serialization schema doesn't support Kafka key/value serialization > > > Key: FLINK-18800 > URL: https://issues.apache.org/jira/browse/FLINK-18800 > Project: Flink > Issue Type: Improvement > Components: Connectors / Kafka, Formats (JSON, Avro, Parquet, ORC, > SequenceFile) >Affects Versions: 1.11.0, 1.11.1 >Reporter: Mohammad Hossein Gerami >Priority: Critical > > {color:#ff8b00}AvroSerializationSchema{color} and > {color:#ff8b00}ConfluentRegistryAvroSerializationSchema{color} doesn't > support Kafka key/value serialization. I implemented a custom Avro > serialization schema for solving this problem. > please consensus to implement new class to support kafka key/value > serialization. > for example in the Flink must implement a class like this: > {code:java} > public class KafkaAvroRegistrySchemaSerializationSchema extends > RegistryAvroSerializationSchema implements > KafkaSerializationSchema{code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-18850) Add late records dropped metric for row time over windows
[ https://issues.apache.org/jira/browse/FLINK-18850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181153#comment-17181153 ] ZhuShang commented on FLINK-18850: -- hi [~libenchao], can i take this,if no one else working on this issus. my solution is : when the element is late to be droped,inc the numLateRecordsDropped accumulator. like this {code:java} org.apache.flink.table.runtime.operators.window.WindowOperator if (isElementDropped) { numLateRecordsDropped.inc();// ← add here // markEvent will increase numLateRecordsDropped lateRecordsDroppedRate.markEvent(); } {code} what do you think? > Add late records dropped metric for row time over windows > - > > Key: FLINK-18850 > URL: https://issues.apache.org/jira/browse/FLINK-18850 > Project: Flink > Issue Type: Improvement > Components: Table SQL / Runtime >Reporter: Benchao Li >Priority: Major > Labels: Starter > > Currently all the row time over windows in blink planner runtime discards > late records silently, it would be good to have a metric about the late > records dropping. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-18800) Avro serialization schema doesn't support Kafka key/value serialization
[ https://issues.apache.org/jira/browse/FLINK-18800?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mohammad Hossein Gerami updated FLINK-18800: Issue Type: Task (was: Improvement) > Avro serialization schema doesn't support Kafka key/value serialization > > > Key: FLINK-18800 > URL: https://issues.apache.org/jira/browse/FLINK-18800 > Project: Flink > Issue Type: Task > Components: Connectors / Kafka, Formats (JSON, Avro, Parquet, ORC, > SequenceFile) >Affects Versions: 1.11.0, 1.11.1 >Reporter: Mohammad Hossein Gerami >Priority: Critical > > {color:#ff8b00}AvroSerializationSchema{color} and > {color:#ff8b00}ConfluentRegistryAvroSerializationSchema{color} doesn't > support Kafka key/value serialization. I implemented a custom Avro > serialization schema for solving this problem. > please consensus to implement new class to support kafka key/value > serialization. > for example in the Flink must implement a class like this: > {code:java} > public class KafkaAvroRegistrySchemaSerializationSchema extends > RegistryAvroSerializationSchema implements > KafkaSerializationSchema{code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-18800) Avro serialization schema doesn't support Kafka key/value serialization
[ https://issues.apache.org/jira/browse/FLINK-18800?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mohammad Hossein Gerami updated FLINK-18800: Description: {color:#ff8b00}AvroSerializationSchema{color} and {color:#ff8b00}ConfluentRegistryAvroSerializationSchema{color} doesn't support Kafka key/value serialization. I implemented a custom Avro serialization schema for solving this problem. please consensus to implement new class to support kafka key/value serialization. for example in the Flink must implement a class like this: {code:java} public class KafkaAvroRegistrySchemaSerializationSchema extends RegistryAvroSerializationSchema implements KafkaSerializationSchema{code} was: {color:#ff8b00}AvroSerializationSchema{color} and {color:#ff8b00}ConfluentRegistryAvroSerializationSchema{color} doesn't support Kafka key/value serialization. I implemented a custom Avro serialization schema for solving this problem. for example in the Flink must implement a class like this. {code:java} public class KafkaAvroRegistrySchemaSerializationSchema extends RegistryAvroSerializationSchema implements KafkaSerializationSchema{code} > Avro serialization schema doesn't support Kafka key/value serialization > > > Key: FLINK-18800 > URL: https://issues.apache.org/jira/browse/FLINK-18800 > Project: Flink > Issue Type: Improvement > Components: Connectors / Kafka, Formats (JSON, Avro, Parquet, ORC, > SequenceFile) >Affects Versions: 1.11.0, 1.11.1 >Reporter: Mohammad Hossein Gerami >Priority: Major > > {color:#ff8b00}AvroSerializationSchema{color} and > {color:#ff8b00}ConfluentRegistryAvroSerializationSchema{color} doesn't > support Kafka key/value serialization. I implemented a custom Avro > serialization schema for solving this problem. > please consensus to implement new class to support kafka key/value > serialization. > for example in the Flink must implement a class like this: > {code:java} > public class KafkaAvroRegistrySchemaSerializationSchema extends > RegistryAvroSerializationSchema implements > KafkaSerializationSchema{code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-18960) flink sideoutput union
[ https://issues.apache.org/jira/browse/FLINK-18960?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zhu Zhu updated FLINK-18960: Fix Version/s: (was: 1.10.2) > flink sideoutput union > -- > > Key: FLINK-18960 > URL: https://issues.apache.org/jira/browse/FLINK-18960 > Project: Flink > Issue Type: Bug > Components: API / DataStream >Affects Versions: 1.10.1 >Reporter: xiaohang.li >Priority: Minor > > Flink sideoutput union seems not works right. If we union the sideoutput from > the same operator, the output is the result of last side output times by the > number of unions, which is not expected. For example, > {code:java} > val side = new OutputTag[String]("side") > val side2 = new OutputTag[String]("side2") > val side3 = new OutputTag[String]("side3") > val ds = env.socketTextStream("master",9001) > val res = ds.process(new ProcessFunction[String,String] { > override def processElement(value: String, ctx: ProcessFunction[String, > String]#Context, out: Collector[String]): Unit = { > if(value.contains("hello")) > { ctx.output(side,value) } > else if(value.contains("world")) > { ctx.output(side2,value) } > else if(value.contains("flink")) > { ctx.output(side3,value) } > out.collect(value) > } > }) > val res1 = res.getSideOutput(side) > val res2 = res.getSideOutput(side2) > val res3 = res.getSideOutput(side3) > println( ">"+res1.getClass) > println( ">"+res2.getClass) > res1.print("res1") > res2.print("res2") > res3.print("res3") > res2.union(res1).union(res3).print("all") > {code} > > If we input > {code:java} > hello > world > flink > {code} > The output will be > > {code:java} > res1> hello > res2> world > res3> flink > all> flink > all> flink > all> flink > {code} > > But the expected output would be > {code:java} > res1> hello > res2> world > res3> flink > all> hello > all> world > all> flink > {code} > > > if we add a _map_ after the sideoutput and then union them, the output would > be right, but adding map should be not needed. > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] curcur commented on a change in pull request #13175: [FLINK-18955][Checkpointing] Add checkpoint path to job startup/restore message
curcur commented on a change in pull request #13175: URL: https://github.com/apache/flink/pull/13175#discussion_r473935762 ## File path: flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CompletedCheckpoint.java ## @@ -320,6 +320,12 @@ void setDiscardCallback(@Nullable CompletedCheckpointStats.DiscardCallback disca @Override public String toString() { - return String.format("Checkpoint %d @ %d for %s", checkpointID, timestamp, job); + return String.format( + "%s %d @ %d for %s located at %s", + props.getCheckpointType(), Review comment: Please let me know what do you think :-). I would be happy to change it. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] curcur commented on a change in pull request #13175: [FLINK-18955][Checkpointing] Add checkpoint path to job startup/restore message
curcur commented on a change in pull request #13175: URL: https://github.com/apache/flink/pull/13175#discussion_r473929844 ## File path: flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CompletedCheckpoint.java ## @@ -320,6 +320,12 @@ void setDiscardCallback(@Nullable CompletedCheckpointStats.DiscardCallback disca @Override public String toString() { - return String.format("Checkpoint %d @ %d for %s", checkpointID, timestamp, job); + return String.format( + "%s %d @ %d for %s located at %s", + props.getCheckpointType(), Review comment: I do not have a strong opinion on this as well. But since you ask, let me have a try LOL :-) The only concern is that we "might" have more types in the future 1. checkpoint or savepoint; 2. Synchronous or not; 3. Global or Individual; 4. unaligned or not (I mean it could be); - It is a bit difficult to categorize the enum of CheckpointType to just "Checkpoint" and "Savepoint" - It is also difficult to tell what level of technicality a user needs. Or in other words, if we think SYNC_SAVEPOINT is not that understandable, we can improve its expressiveness. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-18850) Add late records dropped metric for row time over windows
[ https://issues.apache.org/jira/browse/FLINK-18850?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ZhuShang updated FLINK-18850: - Attachment: (was: image-2020-08-20-20-25-03-981.png) > Add late records dropped metric for row time over windows > - > > Key: FLINK-18850 > URL: https://issues.apache.org/jira/browse/FLINK-18850 > Project: Flink > Issue Type: Improvement > Components: Table SQL / Runtime >Reporter: Benchao Li >Priority: Major > Labels: Starter > > Currently all the row time over windows in blink planner runtime discards > late records silently, it would be good to have a metric about the late > records dropping. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] curcur commented on a change in pull request #13175: [FLINK-18955][Checkpointing] Add checkpoint path to job startup/restore message
curcur commented on a change in pull request #13175: URL: https://github.com/apache/flink/pull/13175#discussion_r473929844 ## File path: flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CompletedCheckpoint.java ## @@ -320,6 +320,12 @@ void setDiscardCallback(@Nullable CompletedCheckpointStats.DiscardCallback disca @Override public String toString() { - return String.format("Checkpoint %d @ %d for %s", checkpointID, timestamp, job); + return String.format( + "%s %d @ %d for %s located at %s", + props.getCheckpointType(), Review comment: I do not have a strong opinion on this as well. But since you ask, let me have a try LOL :-) The only concern is that we "might" have more types in the future 1. checkpoint or savepoint; 2. Synchronous or not; 3. Global or Individual; 4. unaligned or not (I mean it could be); - It is a bit difficult to categorize the enum of CheckpointType to just "Checkpoint" and "Savepoint" - It is also difficult to tell what level of technicality a user needs. For a user that makes an SYNC_SAVEPOINT, he/she probably has enough knowledge to understand what an SYNC_SAVEPOINT is. Or in other words, if we think SYNC_SAVEPOINT is not that understandable, we can improve its expressiveness. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #13207: [FLINK-18952][python][docs] Add 10 minutes to DataStream API documentation
flinkbot commented on pull request #13207: URL: https://github.com/apache/flink/pull/13207#issuecomment-677633881 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit 371eaa2083054d66ef256a9a0ef7a90e116fee0d (Thu Aug 20 12:26:13 UTC 2020) ✅no warnings Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Issue Comment Deleted] (FLINK-18850) Add late records dropped metric for row time over windows
[ https://issues.apache.org/jira/browse/FLINK-18850?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ZhuShang updated FLINK-18850: - Comment: was deleted (was: hi,[~libenchao] i found the accumulator in WindowOperator {code:java} private static final String LATE_ELEMENTS_DROPPED_METRIC_NAME = "numLateRecordsDropped"; protected transient Counter numLateRecordsDropped; {code} and also ,in flink ui has the metrics as below !image-2020-08-20-20-25-03-981.png! ) > Add late records dropped metric for row time over windows > - > > Key: FLINK-18850 > URL: https://issues.apache.org/jira/browse/FLINK-18850 > Project: Flink > Issue Type: Improvement > Components: Table SQL / Runtime >Reporter: Benchao Li >Priority: Major > Labels: Starter > > Currently all the row time over windows in blink planner runtime discards > late records silently, it would be good to have a metric about the late > records dropping. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-18850) Add late records dropped metric for row time over windows
[ https://issues.apache.org/jira/browse/FLINK-18850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181149#comment-17181149 ] ZhuShang commented on FLINK-18850: -- hi,[~libenchao] i found the accumulator in WindowOperator {code:java} private static final String LATE_ELEMENTS_DROPPED_METRIC_NAME = "numLateRecordsDropped"; protected transient Counter numLateRecordsDropped; {code} and also ,in flink ui has the metrics as below !image-2020-08-20-20-25-03-981.png! > Add late records dropped metric for row time over windows > - > > Key: FLINK-18850 > URL: https://issues.apache.org/jira/browse/FLINK-18850 > Project: Flink > Issue Type: Improvement > Components: Table SQL / Runtime >Reporter: Benchao Li >Priority: Major > Labels: Starter > Attachments: image-2020-08-20-20-25-03-981.png > > > Currently all the row time over windows in blink planner runtime discards > late records silently, it would be good to have a metric about the late > records dropping. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-18850) Add late records dropped metric for row time over windows
[ https://issues.apache.org/jira/browse/FLINK-18850?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ZhuShang updated FLINK-18850: - Attachment: image-2020-08-20-20-25-03-981.png > Add late records dropped metric for row time over windows > - > > Key: FLINK-18850 > URL: https://issues.apache.org/jira/browse/FLINK-18850 > Project: Flink > Issue Type: Improvement > Components: Table SQL / Runtime >Reporter: Benchao Li >Priority: Major > Labels: Starter > Attachments: image-2020-08-20-20-25-03-981.png > > > Currently all the row time over windows in blink planner runtime discards > late records silently, it would be good to have a metric about the late > records dropping. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19005) used metaspace grow on every execution
Guillermo Sánchez created FLINK-19005: - Summary: used metaspace grow on every execution Key: FLINK-19005 URL: https://issues.apache.org/jira/browse/FLINK-19005 Project: Flink Issue Type: Bug Components: API / DataSet, Client / Job Submission Affects Versions: 1.11.1 Reporter: Guillermo Sánchez Hi ! Im running a 1.11.1 flink cluster, where I execute batch jobs made with DataSet API. I submit these jobs every day to calculate daily data. In every execution, cluster's used metaspace increase by 7MB and its never released. This ends up with an OutOfMemoryError caused by Metaspace every 15 days and i need to restart the cluster to clean the metaspace taskmanager.memory.jvm-metaspace.size is set to 512mb Any idea of what could be causing this metaspace grow and why is it not released ? -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] curcur commented on a change in pull request #13175: [FLINK-18955][Checkpointing] Add checkpoint path to job startup/restore message
curcur commented on a change in pull request #13175: URL: https://github.com/apache/flink/pull/13175#discussion_r473929844 ## File path: flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CompletedCheckpoint.java ## @@ -320,6 +320,12 @@ void setDiscardCallback(@Nullable CompletedCheckpointStats.DiscardCallback disca @Override public String toString() { - return String.format("Checkpoint %d @ %d for %s", checkpointID, timestamp, job); + return String.format( + "%s %d @ %d for %s located at %s", + props.getCheckpointType(), Review comment: I do not have a strong opinion on this as well. But since you ask, let me have a try LOL :-) The only concern is that we "might" have more types in the future 1. checkpoint or savepoint; 2. Synchronous or not; 3. Global or Individual; 4. unaligned or not (I mean it could be); - It is a bit difficult to categorize the enum of CheckpointType to just "Checkpoint" and "Savepoint" - It is also difficult to tell what level of technicality a user needs. For a user that makes an SYNC_SAVEPOINT, he/she probably has enough knowledge to understand what an SYNC_SAVEPOINT is. Or in other words, if we think SYNC_SAVEPOINT is not that understandable, we probably need to improve its expressiveness. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-18952) Add 10 minutes to DataStream API documentation
[ https://issues.apache.org/jira/browse/FLINK-18952?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-18952: --- Labels: pull-request-available (was: ) > Add 10 minutes to DataStream API documentation > -- > > Key: FLINK-18952 > URL: https://issues.apache.org/jira/browse/FLINK-18952 > Project: Flink > Issue Type: Sub-task > Components: API / Python >Reporter: Hequn Cheng >Assignee: Hequn Cheng >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0 > > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] hequn8128 opened a new pull request #13207: [FLINK-18952][python][docs] Add 10 minutes to DataStream API documentation
hequn8128 opened a new pull request #13207: URL: https://github.com/apache/flink/pull/13207 ## What is the purpose of the change This pull request adds "10 minutes to DataStream API" documentation. This page a short introduction to Python DataStream, geared mainly for new users. ## Brief change log - Adds "10 minutes to DataStream API" documentation ## Verifying this change This change is a trivial rework / code cleanup without any test coverage. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (no) - The S3 file system connector: (no) ## Documentation - Does this pull request introduce a new feature? (no) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Comment Edited] (FLINK-18938) Throw better exception message for quering sink-only connector
[ https://issues.apache.org/jira/browse/FLINK-18938?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181135#comment-17181135 ] liufangliang edited comment on FLINK-18938 at 8/20/20, 12:22 PM: - Hi [~jark], Can you assign it to me ? The following tips are very complete and precise, I think there is no need to optimize. {code:java} Caused by: org.apache.flink.table.api.ValidationException: Could not find any factory for identifier 'elasticsearch-7' that implements 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. {code} Tips(support as sink or source) can follow the identifier. For example: {code:java} Available factory identifiers are: datagen (source,sink) elasticsearch-7 (sink-only) test-connector (source-only) {code} [~jark] [~lzljs3620320] What do you think of my solution? was (Author: liufangliang): Hi [~jark], Can you assign it to me ? The following tips are very complete and precise, I think there is no need to optimize. {code:java} Caused by: org.apache.flink.table.api.ValidationException: Could not find any factory for identifier 'elasticsearch-7' that implements 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. {code} Tips(support as sink or source) can follow the identifier. For example: {code:java} Available factory identifiers are: datagen (source,sink) elasticsearch-7 (sink-only) test-connector (source-only) {code} [~jark] [~lzljs3620320] What do you think of my solution? > Throw better exception message for quering sink-only connector > -- > > Key: FLINK-18938 > URL: https://issues.apache.org/jira/browse/FLINK-18938 > Project: Flink > Issue Type: Improvement >Reporter: Jark Wu >Priority: Major > > Currently, if we are quering a sink-only connector, for example: {{SELECT * > FROM elasticsearch_sink}}, a following exception will be thrown: > {code} > Caused by: org.apache.flink.table.api.ValidationException: Could not find any > factory for identifier 'elasticsearch-7' that implements > 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. > Available factory identifiers are: > datagen > {code} > The above exception is very misleading, it sounds like that the elasticsearch > jar is not loaded, however the elasticsearch jar is in the lib directory of > Flink cluster. > We can improve the exception that explicitly telling users the found > connector only support as sink, can't be used as a source. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-18938) Throw better exception message for quering sink-only connector
[ https://issues.apache.org/jira/browse/FLINK-18938?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181145#comment-17181145 ] Jark Wu commented on FLINK-18938: - The exception is exactly the same if we don't have a dependency jar in the classpath. So many people are misled to think it is because of missing dependency and not notice the "DynamicTableSourceFactory". What I want to improve is to make the exception message explicit for this case. > Throw better exception message for quering sink-only connector > -- > > Key: FLINK-18938 > URL: https://issues.apache.org/jira/browse/FLINK-18938 > Project: Flink > Issue Type: Improvement >Reporter: Jark Wu >Priority: Major > > Currently, if we are quering a sink-only connector, for example: {{SELECT * > FROM elasticsearch_sink}}, a following exception will be thrown: > {code} > Caused by: org.apache.flink.table.api.ValidationException: Could not find any > factory for identifier 'elasticsearch-7' that implements > 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. > Available factory identifiers are: > datagen > {code} > The above exception is very misleading, it sounds like that the elasticsearch > jar is not loaded, however the elasticsearch jar is in the lib directory of > Flink cluster. > We can improve the exception that explicitly telling users the found > connector only support as sink, can't be used as a source. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot edited a comment on pull request #12880: [FLINK-18555][table sql/api] Make TableConfig options can be configur…
flinkbot edited a comment on pull request #12880: URL: https://github.com/apache/flink/pull/12880#issuecomment-657437173 ## CI report: * 94e8eb39032665f77e897cb572cd4320b66635f9 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5703) * 72e7dcb66f3a2c6eda4899d51ddd6420d2d5822b Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5744) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #12880: [FLINK-18555][table sql/api] Make TableConfig options can be configur…
flinkbot edited a comment on pull request #12880: URL: https://github.com/apache/flink/pull/12880#issuecomment-657437173 ## CI report: * 94e8eb39032665f77e897cb572cd4320b66635f9 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5703) * 72e7dcb66f3a2c6eda4899d51ddd6420d2d5822b UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-18938) Throw better exception message for quering sink-only connector
[ https://issues.apache.org/jira/browse/FLINK-18938?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181135#comment-17181135 ] liufangliang commented on FLINK-18938: -- Hi [~jark], Can you assign it to me ? The following tips are very complete and precise, I think there is no need to optimize. {code:java} Caused by: org.apache.flink.table.api.ValidationException: Could not find any factory for identifier 'elasticsearch-7' that implements 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. {code} Tips(support as sink or source) can follow the identifier. For example: {code:java} Available factory identifiers are: datagen (source,sink) elasticsearch-7 (sink-only) test-connector (source-only) {code} [~jark] [~lzljs3620320] What do you think of my solution? > Throw better exception message for quering sink-only connector > -- > > Key: FLINK-18938 > URL: https://issues.apache.org/jira/browse/FLINK-18938 > Project: Flink > Issue Type: Improvement >Reporter: Jark Wu >Priority: Major > > Currently, if we are quering a sink-only connector, for example: {{SELECT * > FROM elasticsearch_sink}}, a following exception will be thrown: > {code} > Caused by: org.apache.flink.table.api.ValidationException: Could not find any > factory for identifier 'elasticsearch-7' that implements > 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. > Available factory identifiers are: > datagen > {code} > The above exception is very misleading, it sounds like that the elasticsearch > jar is not loaded, however the elasticsearch jar is in the lib directory of > Flink cluster. > We can improve the exception that explicitly telling users the found > connector only support as sink, can't be used as a source. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Comment Edited] (FLINK-18938) Throw better exception message for quering sink-only connector
[ https://issues.apache.org/jira/browse/FLINK-18938?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181135#comment-17181135 ] liufangliang edited comment on FLINK-18938 at 8/20/20, 11:49 AM: - Hi [~jark], Can you assign it to me ? The following tips are very complete and precise, I think there is no need to optimize. {code:java} Caused by: org.apache.flink.table.api.ValidationException: Could not find any factory for identifier 'elasticsearch-7' that implements 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. {code} Tips(support as sink or source) can follow the identifier. For example: {code:java} Available factory identifiers are: datagen (source,sink) elasticsearch-7 (sink-only) test-connector (source-only) {code} [~jark] [~lzljs3620320] What do you think of my solution? was (Author: liufangliang): Hi [~jark], Can you assign it to me ? The following tips are very complete and precise, I think there is no need to optimize. {code:java} Caused by: org.apache.flink.table.api.ValidationException: Could not find any factory for identifier 'elasticsearch-7' that implements 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. {code} Tips(support as sink or source) can follow the identifier. For example: {code:java} Available factory identifiers are: datagen (source,sink) elasticsearch-7 (sink-only) test-connector (source-only) {code} [~jark] [~lzljs3620320] What do you think of my solution? > Throw better exception message for quering sink-only connector > -- > > Key: FLINK-18938 > URL: https://issues.apache.org/jira/browse/FLINK-18938 > Project: Flink > Issue Type: Improvement >Reporter: Jark Wu >Priority: Major > > Currently, if we are quering a sink-only connector, for example: {{SELECT * > FROM elasticsearch_sink}}, a following exception will be thrown: > {code} > Caused by: org.apache.flink.table.api.ValidationException: Could not find any > factory for identifier 'elasticsearch-7' that implements > 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. > Available factory identifiers are: > datagen > {code} > The above exception is very misleading, it sounds like that the elasticsearch > jar is not loaded, however the elasticsearch jar is in the lib directory of > Flink cluster. > We can improve the exception that explicitly telling users the found > connector only support as sink, can't be used as a source. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Comment Edited] (FLINK-18938) Throw better exception message for quering sink-only connector
[ https://issues.apache.org/jira/browse/FLINK-18938?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181135#comment-17181135 ] liufangliang edited comment on FLINK-18938 at 8/20/20, 11:49 AM: - Hi [~jark], Can you assign it to me ? The following tips are very complete and precise, I think there is no need to optimize. {code:java} Caused by: org.apache.flink.table.api.ValidationException: Could not find any factory for identifier 'elasticsearch-7' that implements 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. {code} Tips(support as sink or source) can follow the identifier. For example: {code:java} Available factory identifiers are: datagen (source,sink) elasticsearch-7 (sink-only) test-connector (source-only) {code} [~jark] [~lzljs3620320] What do you think of my solution? was (Author: liufangliang): Hi [~jark], Can you assign it to me ? The following tips are very complete and precise, I think there is no need to optimize. {code:java} Caused by: org.apache.flink.table.api.ValidationException: Could not find any factory for identifier 'elasticsearch-7' that implements 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. {code} Tips(support as sink or source) can follow the identifier. For example: {code:java} Available factory identifiers are: datagen (source,sink) elasticsearch-7 (sink-only) test-connector (source-only) {code} [~jark] [~lzljs3620320] What do you think of my solution? > Throw better exception message for quering sink-only connector > -- > > Key: FLINK-18938 > URL: https://issues.apache.org/jira/browse/FLINK-18938 > Project: Flink > Issue Type: Improvement >Reporter: Jark Wu >Priority: Major > > Currently, if we are quering a sink-only connector, for example: {{SELECT * > FROM elasticsearch_sink}}, a following exception will be thrown: > {code} > Caused by: org.apache.flink.table.api.ValidationException: Could not find any > factory for identifier 'elasticsearch-7' that implements > 'org.apache.flink.table.factories.DynamicTableSourceFactory' in the classpath. > Available factory identifiers are: > datagen > {code} > The above exception is very misleading, it sounds like that the elasticsearch > jar is not loaded, however the elasticsearch jar is in the lib directory of > Flink cluster. > We can improve the exception that explicitly telling users the found > connector only support as sink, can't be used as a source. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] wangzzu removed a comment on pull request #13024: [FLINK-18742][flink-clients] Some configuration args do not take effect at client
wangzzu removed a comment on pull request #13024: URL: https://github.com/apache/flink/pull/13024#issuecomment-677524943 @flinkbot run azure This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink-web] zhuzhurk commented on a change in pull request #366: Add Apache Flink release 1.10.2
zhuzhurk commented on a change in pull request #366: URL: https://github.com/apache/flink-web/pull/366#discussion_r473908601 ## File path: _posts/2020-08-11-release-1.10.2.md ## @@ -0,0 +1,199 @@ +--- +layout: post +title: "Apache Flink 1.10.2 Released" +date: 2020-08-11 18:00:00 +categories: news +authors: +- Zhu Zhu: + name: "Zhu Zhu" + twitter: "zhuzhv" +--- + +The Apache Flink community released the second bugfix version of the Apache Flink 1.10 series. + +This release includes 73 fixes and minor improvements for Flink 1.10.1. The list below includes a detailed list of all fixes and improvements. + +We highly recommend all users to upgrade to Flink 1.10.2. + +Updated Maven dependencies: + +```xml + + org.apache.flink + flink-java + 1.10.2 + + + org.apache.flink + flink-streaming-java_2.11 + 1.10.2 + + + org.apache.flink + flink-clients_2.11 + 1.10.2 + +``` + +You can find the binaries on the updated [Downloads page]({{ site.baseurl }}/downloads.html). + +List of resolved issues: + +Sub-task + + +[FLINK-15836] - Throw fatal error in KubernetesResourceManager when the pods watcher is closed with exception + +[FLINK-16160] - Schema#proctime and Schema#rowtime dont work in TableEnvironment#connect code path + + + +Bug + + +[FLINK-13689] - Rest High Level Client for Elasticsearch6.x connector leaks threads if no connection could be established + +[FLINK-14369] - KafkaProducerAtLeastOnceITCaseKafkaProducerTestBase.testOneToOneAtLeastOnceCustomOperator fails on Travis + +[FLINK-14836] - Unable to set yarn container number for scala shell in yarn mode + +[FLINK-14894] - HybridOffHeapUnsafeMemorySegmentTest#testByteBufferWrap failed on Travis + +[FLINK-15758] - Investigate potential out-of-memory problems due to managed unsafe memory allocation + +[FLINK-15849] - Update SQL-CLIENT document from type to data-type + +[FLINK-16309] - ElasticSearch 7 connector is missing in SQL connector list + +[FLINK-16346] - BlobsCleanupITCase.testBlobServerCleanupCancelledJob fails on Travis + +[FLINK-16432] - Building Hive connector gives problems + +[FLINK-16451] - Fix IndexOutOfBoundsException for DISTINCT AGG with constants + +[FLINK-16510] - Task manager safeguard shutdown may not be reliable + +[FLINK-17092] - Pyflink test BlinkStreamDependencyTests is instable + +[FLINK-17322] - Enable latency tracker would corrupt the broadcast state + +[FLINK-17420] - Cannot alias Tuple and Row fields when converting DataStream to Table + +[FLINK-17466] - toRetractStream doesnt work correctly with Pojo conversion class + +[FLINK-17555] - docstring of pyflink.table.descriptors.FileSystem:1:duplicate object description of pyflink.table.descriptors.FileSystem + +[FLINK-17558] - Partitions are released in TaskExecutor Main Thread + +[FLINK-17562] - POST /jars/:jarid/plan is not working + +[FLINK-17578] - Union of 2 SideOutputs behaviour incorrect + +[FLINK-17639] - Document which FileSystems are supported by the StreamingFileSink + +[FLINK-17643] - LaunchCoordinatorTest fails + +[FLINK-17700] - The callback client of JavaGatewayServer should run in a daemon thread + +[FLINK-17744] - StreamContextEnvironment#execute cannot be call JobListener#onJobExecuted + +[FLINK-17763] - No log files when starting scala-shell + +[FLINK-17788] - scala shell in yarn mode is broken + +[FLINK-17800] - RocksDB optimizeForPointLookup results in missing time windows + +[FLINK-17801] - TaskExecutorTest.testHeartbeatTimeoutWithResourceManager timeout + +[FLINK-17809] - BashJavaUtil script logic does not work for paths with spaces + +[FLINK-17822] - Nightly Flink CLI end-to-end test failed with JavaGcCleanerWrapper$PendingCleanersRunner cannot access class jdk.internal.misc.SharedSecrets in Java 11 + +[FLINK-17870] - dependent jars are missing to be shipped to cluster in scala shell + +[FLINK-17891] - FlinkYarnSessionCli sets wrong execution.target type + +[FLINK-17959] - Exception: CANCELLED: call already cancelled is thrown when run python udf + +[FLINK-18008] - HistoryServer does not log environment information on startup + +[FLINK-18012] - Deactivate slot timeout if TaskSlotTable.tryMarkSlotActive is called + +[FLINK-18035] - Executors#newCachedThreadPool could not work as expected + +[FLINK-18045] - Fix Kerberos credentials checking to unblock Flink on secured MapR + +[FLINK-18048] - --host option could not take effect for standalone application cluster + +[FLINK-18097] - History server doesnt clean all job json files + +[FLINK-18168] - Error results when use UDAF with Object Array return type + +[FLINK-18223] -
[GitHub] [flink-web] asfgit closed pull request #370: Add a blog post about the current state of Flink on Docker
asfgit closed pull request #370: URL: https://github.com/apache/flink-web/pull/370 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] lirui-apache commented on a change in pull request #13157: [FLINK-18900][hive] HiveCatalog should error out when listing partitions with an invalid spec
lirui-apache commented on a change in pull request #13157: URL: https://github.com/apache/flink/pull/13157#discussion_r473905054 ## File path: flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/catalog/hive/HiveCatalog.java ## @@ -777,8 +777,11 @@ public void dropPartition(ObjectPath tablePath, CatalogPartitionSpec partitionSp // partition spec can be partial List partialVals = HiveReflectionUtils.getPvals(hiveShim, hiveTable.getPartitionKeys(), partitionSpec.getPartitionSpec()); + checkValidPartitionSpec(partitionSpec, getFieldNames(hiveTable.getPartitionKeys()), tablePath); return client.listPartitionNames(tablePath.getDatabaseName(), tablePath.getObjectName(), partialVals, (short) -1).stream().map(HiveCatalog::createPartitionSpec).collect(Collectors.toList()); + } catch (PartitionSpecInvalidException e) { Review comment: Why do we need to catch the exception and throw another one? ## File path: flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/catalog/hive/HiveCatalog.java ## @@ -777,8 +777,11 @@ public void dropPartition(ObjectPath tablePath, CatalogPartitionSpec partitionSp // partition spec can be partial List partialVals = HiveReflectionUtils.getPvals(hiveShim, hiveTable.getPartitionKeys(), partitionSpec.getPartitionSpec()); + checkValidPartitionSpec(partitionSpec, getFieldNames(hiveTable.getPartitionKeys()), tablePath); Review comment: This should be done before we call `HiveReflectionUtils.getPvals` This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13024: [FLINK-18742][flink-clients] Some configuration args do not take effect at client
flinkbot edited a comment on pull request #13024: URL: https://github.com/apache/flink/pull/13024#issuecomment-666148245 ## CI report: * eb611d5b39f997fa3e986b5d163cb65a44b4b0ba UNKNOWN * a5db01bd0f63289039dfcdae41eaf099ed28a812 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5742) Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5730) * 047cc5ec9e12641f2787ac6fa4a3525fd6fc7d70 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5743) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-19004) Fail to call Hive percentile function together with distinct aggregate call
[ https://issues.apache.org/jira/browse/FLINK-19004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rui Li updated FLINK-19004: --- Description: The following test case would fail: {code} @Test public void test() throws Exception { TableEnvironment tableEnv = getTableEnvWithHiveCatalog(); tableEnv.unloadModule("core"); tableEnv.loadModule("hive", new HiveModule()); tableEnv.loadModule("core", CoreModule.INSTANCE); tableEnv.executeSql("create table src(x int,y int)"); tableEnv.executeSql("select count(distinct y),`percentile`(y,`array`(0.5,0.99)) from src group by x").collect(); } {code} The error is: {noformat} org.apache.flink.table.api.TableException: Cannot generate a valid execution plan for the given query: FlinkLogicalLegacySink(name=[collect], fields=[EXPR$0, EXPR$1]) +- FlinkLogicalCalc(select=[EXPR$0, EXPR$1]) +- FlinkLogicalAggregate(group=[{0}], EXPR$0=[COUNT($1) FILTER $3], EXPR$1=[MIN($2) FILTER $4]) +- FlinkLogicalCalc(select=[x, y, EXPR$1, =(CASE(=($e, 0:BIGINT), 0:BIGINT, 1:BIGINT), 0) AS $g_0, =(CASE(=($e, 0:BIGINT), 0:BIGINT, 1:BIGINT), 1) AS $g_1]) +- FlinkLogicalAggregate(group=[{0, 1, 3}], EXPR$1=[percentile($4, $2)]) +- FlinkLogicalExpand(projects=[x, y, $f2, $e, y_0]) +- FlinkLogicalCalc(select=[x, y, array(0.5:DECIMAL(2, 1), 0.99:DECIMAL(3, 2)) AS $f2]) +- FlinkLogicalLegacyTableSourceScan(table=[[test-catalog, default, src, source: [HiveTableSource(x, y) TablePath: default.src, PartitionPruned: false, PartitionNums: null]]], fields=[x, y]) Min aggregate function does not support type: ''ARRAY''. {noformat} > Fail to call Hive percentile function together with distinct aggregate call > --- > > Key: FLINK-19004 > URL: https://issues.apache.org/jira/browse/FLINK-19004 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive, Table SQL / Planner >Reporter: Rui Li >Priority: Major > > The following test case would fail: > {code} > @Test > public void test() throws Exception { > TableEnvironment tableEnv = getTableEnvWithHiveCatalog(); > tableEnv.unloadModule("core"); > tableEnv.loadModule("hive", new HiveModule()); > tableEnv.loadModule("core", CoreModule.INSTANCE); > tableEnv.executeSql("create table src(x int,y int)"); > tableEnv.executeSql("select count(distinct > y),`percentile`(y,`array`(0.5,0.99)) from src group by x").collect(); > } > {code} > The error is: > {noformat} > org.apache.flink.table.api.TableException: Cannot generate a valid execution > plan for the given query: > FlinkLogicalLegacySink(name=[collect], fields=[EXPR$0, EXPR$1]) > +- FlinkLogicalCalc(select=[EXPR$0, EXPR$1]) >+- FlinkLogicalAggregate(group=[{0}], EXPR$0=[COUNT($1) FILTER $3], > EXPR$1=[MIN($2) FILTER $4]) > +- FlinkLogicalCalc(select=[x, y, EXPR$1, =(CASE(=($e, 0:BIGINT), > 0:BIGINT, 1:BIGINT), 0) AS $g_0, =(CASE(=($e, 0:BIGINT), 0:BIGINT, 1:BIGINT), > 1) AS $g_1]) > +- FlinkLogicalAggregate(group=[{0, 1, 3}], EXPR$1=[percentile($4, > $2)]) > +- FlinkLogicalExpand(projects=[x, y, $f2, $e, y_0]) >+- FlinkLogicalCalc(select=[x, y, array(0.5:DECIMAL(2, 1), > 0.99:DECIMAL(3, 2)) AS $f2]) > +- FlinkLogicalLegacyTableSourceScan(table=[[test-catalog, > default, src, source: [HiveTableSource(x, y) TablePath: default.src, > PartitionPruned: false, PartitionNums: null]]], fields=[x, y]) > Min aggregate function does not support type: ''ARRAY''. > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19004) Fail to call Hive percentile function together with distinct aggregate call
Rui Li created FLINK-19004: -- Summary: Fail to call Hive percentile function together with distinct aggregate call Key: FLINK-19004 URL: https://issues.apache.org/jira/browse/FLINK-19004 Project: Flink Issue Type: Bug Components: Connectors / Hive, Table SQL / Planner Reporter: Rui Li -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink-web] Myasuka commented on a change in pull request #366: Add Apache Flink release 1.10.2
Myasuka commented on a change in pull request #366: URL: https://github.com/apache/flink-web/pull/366#discussion_r473897004 ## File path: _posts/2020-08-11-release-1.10.2.md ## @@ -0,0 +1,199 @@ +--- +layout: post +title: "Apache Flink 1.10.2 Released" +date: 2020-08-11 18:00:00 +categories: news +authors: +- Zhu Zhu: + name: "Zhu Zhu" + twitter: "zhuzhv" +--- + +The Apache Flink community released the second bugfix version of the Apache Flink 1.10 series. + +This release includes 73 fixes and minor improvements for Flink 1.10.1. The list below includes a detailed list of all fixes and improvements. + +We highly recommend all users to upgrade to Flink 1.10.2. + +Updated Maven dependencies: + +```xml + + org.apache.flink + flink-java + 1.10.2 + + + org.apache.flink + flink-streaming-java_2.11 + 1.10.2 + + + org.apache.flink + flink-clients_2.11 + 1.10.2 + +``` + +You can find the binaries on the updated [Downloads page]({{ site.baseurl }}/downloads.html). + +List of resolved issues: + +Sub-task + + +[FLINK-15836] - Throw fatal error in KubernetesResourceManager when the pods watcher is closed with exception + +[FLINK-16160] - Schema#proctime and Schema#rowtime dont work in TableEnvironment#connect code path + + + +Bug + + +[FLINK-13689] - Rest High Level Client for Elasticsearch6.x connector leaks threads if no connection could be established + +[FLINK-14369] - KafkaProducerAtLeastOnceITCaseKafkaProducerTestBase.testOneToOneAtLeastOnceCustomOperator fails on Travis + +[FLINK-14836] - Unable to set yarn container number for scala shell in yarn mode + +[FLINK-14894] - HybridOffHeapUnsafeMemorySegmentTest#testByteBufferWrap failed on Travis + +[FLINK-15758] - Investigate potential out-of-memory problems due to managed unsafe memory allocation + +[FLINK-15849] - Update SQL-CLIENT document from type to data-type + +[FLINK-16309] - ElasticSearch 7 connector is missing in SQL connector list + +[FLINK-16346] - BlobsCleanupITCase.testBlobServerCleanupCancelledJob fails on Travis + +[FLINK-16432] - Building Hive connector gives problems + +[FLINK-16451] - Fix IndexOutOfBoundsException for DISTINCT AGG with constants + +[FLINK-16510] - Task manager safeguard shutdown may not be reliable + +[FLINK-17092] - Pyflink test BlinkStreamDependencyTests is instable + +[FLINK-17322] - Enable latency tracker would corrupt the broadcast state + +[FLINK-17420] - Cannot alias Tuple and Row fields when converting DataStream to Table + +[FLINK-17466] - toRetractStream doesnt work correctly with Pojo conversion class + +[FLINK-17555] - docstring of pyflink.table.descriptors.FileSystem:1:duplicate object description of pyflink.table.descriptors.FileSystem + +[FLINK-17558] - Partitions are released in TaskExecutor Main Thread + +[FLINK-17562] - POST /jars/:jarid/plan is not working + +[FLINK-17578] - Union of 2 SideOutputs behaviour incorrect + +[FLINK-17639] - Document which FileSystems are supported by the StreamingFileSink + +[FLINK-17643] - LaunchCoordinatorTest fails + +[FLINK-17700] - The callback client of JavaGatewayServer should run in a daemon thread + +[FLINK-17744] - StreamContextEnvironment#execute cannot be call JobListener#onJobExecuted + +[FLINK-17763] - No log files when starting scala-shell + +[FLINK-17788] - scala shell in yarn mode is broken + +[FLINK-17800] - RocksDB optimizeForPointLookup results in missing time windows + +[FLINK-17801] - TaskExecutorTest.testHeartbeatTimeoutWithResourceManager timeout + +[FLINK-17809] - BashJavaUtil script logic does not work for paths with spaces + +[FLINK-17822] - Nightly Flink CLI end-to-end test failed with JavaGcCleanerWrapper$PendingCleanersRunner cannot access class jdk.internal.misc.SharedSecrets in Java 11 + +[FLINK-17870] - dependent jars are missing to be shipped to cluster in scala shell + +[FLINK-17891] - FlinkYarnSessionCli sets wrong execution.target type + +[FLINK-17959] - Exception: CANCELLED: call already cancelled is thrown when run python udf + +[FLINK-18008] - HistoryServer does not log environment information on startup + +[FLINK-18012] - Deactivate slot timeout if TaskSlotTable.tryMarkSlotActive is called + +[FLINK-18035] - Executors#newCachedThreadPool could not work as expected + +[FLINK-18045] - Fix Kerberos credentials checking to unblock Flink on secured MapR + +[FLINK-18048] - --host option could not take effect for standalone application cluster + +[FLINK-18097] - History server doesnt clean all job json files + +[FLINK-18168] - Error results when use UDAF with Object Array return type + +[FLINK-18223] -
[GitHub] [flink] flinkbot edited a comment on pull request #13024: [FLINK-18742][flink-clients] Some configuration args do not take effect at client
flinkbot edited a comment on pull request #13024: URL: https://github.com/apache/flink/pull/13024#issuecomment-666148245 ## CI report: * eb611d5b39f997fa3e986b5d163cb65a44b4b0ba UNKNOWN * a5db01bd0f63289039dfcdae41eaf099ed28a812 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5742) Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5730) * 047cc5ec9e12641f2787ac6fa4a3525fd6fc7d70 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-18959) Fail to archiveExecutionGraph because job is not finished when dispatcher close
[ https://issues.apache.org/jira/browse/FLINK-18959?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181104#comment-17181104 ] ZhuShang commented on FLINK-18959: -- i also found that archiveExecutionGraph can not be reached when the job is running > Fail to archiveExecutionGraph because job is not finished when dispatcher > close > --- > > Key: FLINK-18959 > URL: https://issues.apache.org/jira/browse/FLINK-18959 > Project: Flink > Issue Type: Bug > Components: Runtime / Coordination >Affects Versions: 1.10.0, 1.12.0, 1.11.1 >Reporter: Liu >Priority: Critical > Fix For: 1.12.0, 1.11.2, 1.10.3 > > Attachments: flink-debug-log > > > When job is cancelled, we expect to see it in flink's history server. But I > can not see my job after it is cancelled. > After digging into the problem, I find that the function > archiveExecutionGraph is not executed. Below is the brief log: > {panel:title=log} > 2020-08-14 15:10:06,406 INFO > org.apache.flink.runtime.executiongraph.ExecutionGraph > [flink-akka.actor.default-dispatcher- 15] - Job EtlAndWindow > (6f784d4cc5bae88a332d254b21660372) switched from state RUNNING to CANCELLING. > 2020-08-14 15:10:06,415 DEBUG > org.apache.flink.runtime.dispatcher.MiniDispatcher > [flink-akka.actor.default-dispatcher-3] - Shutting down per-job cluster > because the job was canceled. > 2020-08-14 15:10:06,629 INFO > org.apache.flink.runtime.dispatcher.MiniDispatcher > [flink-akka.actor.default-dispatcher-3] - Stopping dispatcher > akka.tcp://flink@bjfk-c9865.yz02:38663/user/dispatcher. > 2020-08-14 15:10:06,629 INFO > org.apache.flink.runtime.dispatcher.MiniDispatcher > [flink-akka.actor.default-dispatcher-3] - Stopping all currently running jobs > of dispatcher akka.tcp://flink@bjfk-c9865.yz02:38663/user/dispatcher. > 2020-08-14 15:10:06,631 INFO org.apache.flink.runtime.jobmaster.JobMaster > [flink-akka.actor.default-dispatcher-29] - Stopping the JobMaster for job > EtlAndWindow(6f784d4cc5bae88a332d254b21660372). > 2020-08-14 15:10:06,632 DEBUG org.apache.flink.runtime.jobmaster.JobMaster > [flink-akka.actor.default-dispatcher-29] - Disconnect TaskExecutor > container_e144_1590060720089_2161_01_06 because: Stopping JobMaster for > job EtlAndWindow(6f784d4cc5bae88a332d254b21660372). > 2020-08-14 15:10:06,646 INFO > org.apache.flink.runtime.executiongraph.ExecutionGraph > [flink-akka.actor.default-dispatcher-29] - Job EtlAndWindow > (6f784d4cc5bae88a332d254b21660372) switched from state CANCELLING to CANCELED. > 2020-08-14 15:10:06,664 DEBUG > org.apache.flink.runtime.dispatcher.MiniDispatcher > [flink-akka.actor.default-dispatcher-4] - There is a newer JobManagerRunner > for the job 6f784d4cc5bae88a332d254b21660372. > {panel} > From the log, we can see that job is not finished when dispatcher closes. The > process is as following: > * Receive cancel command and send it to all tasks async. > * In MiniDispatcher, begin to shutting down per-job cluster. > * Stopping dispatcher and remove job. > * Job is cancelled and callback is executed in method startJobManagerRunner. > * Because job is removed before, so currentJobManagerRunner is null which > not equals to the original jobManagerRunner. In this case, > archivedExecutionGraph will not be uploaded. > In normal cases, I find that job is cancelled first and then dispatcher is > stopped so that archivedExecutionGraph will succeed. But I think that the > order is not constrained and it is hard to know which comes first. > Above is what I suspected. If so, then we should fix it. > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] RocMarshal commented on pull request #13172: [FLINK-18854][docs-zh] Translate the 'API Migration Guides' page of 'Application Development' into Chinese
RocMarshal commented on pull request #13172: URL: https://github.com/apache/flink/pull/13172#issuecomment-677536186 Hi, @XBaith @klion26 . Could you help me to review this PR if you have free time? Thank you. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] RocMarshal removed a comment on pull request #13172: [FLINK-18854][docs-zh] Translate the 'API Migration Guides' page of 'Application Development' into Chinese
RocMarshal removed a comment on pull request #13172: URL: https://github.com/apache/flink/pull/13172#issuecomment-675233767 Hi, @XBaith @klion26 . Could you help me to review this PR if you have free time? Thank you. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13024: [FLINK-18742][flink-clients] Some configuration args do not take effect at client
flinkbot edited a comment on pull request #13024: URL: https://github.com/apache/flink/pull/13024#issuecomment-666148245 ## CI report: * eb611d5b39f997fa3e986b5d163cb65a44b4b0ba UNKNOWN * a5db01bd0f63289039dfcdae41eaf099ed28a812 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5730) Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5742) * 047cc5ec9e12641f2787ac6fa4a3525fd6fc7d70 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] wangzzu commented on pull request #13024: [FLINK-18742][flink-clients] Some configuration args do not take effect at client
wangzzu commented on pull request #13024: URL: https://github.com/apache/flink/pull/13024#issuecomment-677524943 @flinkbot run azure This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink-web] rmetzger commented on a change in pull request #370: Add a blog post about the current state of Flink on Docker
rmetzger commented on a change in pull request #370: URL: https://github.com/apache/flink-web/pull/370#discussion_r473867137 ## File path: _posts/2020-08-20-flink-docker.md ## @@ -0,0 +1,90 @@ +--- +layout: post +title: "The State of Flink on Docker" +date: 2020-08-08T00:00:00.000Z +authors: +- rmetzger: + name: "Robert Metzger" + twitter: rmetzger_ +categories: news + +excerpt: This blog post gives an update on the recent developments of Flink's support for Docker. +--- + +The Flink community recently put some effort into upgrading the Docker experience for our users. The goal was to reduce confusion and improve usability. With over 50 million downloads from Docker Hub, the Flink docker images are a very popular deployment option. Review comment: okay, I'll use "improve" This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-12351) AsyncWaitOperator should deep copy StreamElement when object reuse is enabled
[ https://issues.apache.org/jira/browse/FLINK-12351?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181094#comment-17181094 ] Piotr Nowojski commented on FLINK-12351: Yes that's true. I was more worried about relaying on an assumption that network stack is not re using the records in any way. But maybe this is not big of an issue and could be guarded by some unit test for `AsyncWaitOperator`. > AsyncWaitOperator should deep copy StreamElement when object reuse is enabled > - > > Key: FLINK-12351 > URL: https://issues.apache.org/jira/browse/FLINK-12351 > Project: Flink > Issue Type: Bug > Components: API / DataStream >Reporter: Jark Wu >Assignee: Jark Wu >Priority: Major > Labels: pull-request-available > Time Spent: 10m > Remaining Estimate: 0h > > Currently, AsyncWaitOperator directly put the input StreamElement into > {{StreamElementQueue}}. But when object reuse is enabled, the StreamElement > is reused, which means the element in {{StreamElementQueue}} will be > modified. As a result, the output of AsyncWaitOperator might be wrong. > An easy way to fix this might be deep copy the input StreamElement when > object reuse is enabled, like this: > https://github.com/apache/flink/blob/blink/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/operators/async/AsyncWaitOperator.java#L209 -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink-web] rmetzger commented on pull request #370: Add a blog post about the current state of Flink on Docker
rmetzger commented on pull request #370: URL: https://github.com/apache/flink-web/pull/370#issuecomment-677515170 Thanks a lot for your reviews! I pushed a new version. Unless there's any more feedback, I'll merge it later today. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink-benchmarks] pnowojski commented on a change in pull request #2: [FLINK-19003][checkpointing] Add micro-benchmark for unaligned checkpoints
pnowojski commented on a change in pull request #2: URL: https://github.com/apache/flink-benchmarks/pull/2#discussion_r473849350 ## File path: src/main/java/org/apache/flink/benchmark/UnalignedCheckpointBenchmark.java ## @@ -0,0 +1,164 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.benchmark; + +import org.apache.flink.api.common.functions.MapFunction; +import org.apache.flink.configuration.ConfigConstants; +import org.apache.flink.configuration.Configuration; +import org.apache.flink.configuration.TaskManagerOptions; +import org.apache.flink.runtime.state.CheckpointListener; +import org.apache.flink.streaming.api.datastream.DataStreamSource; +import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment; +import org.apache.flink.streaming.api.functions.sink.SinkFunction; +import org.apache.flink.streaming.api.functions.source.RichParallelSourceFunction; + +import org.openjdk.jmh.annotations.Benchmark; +import org.openjdk.jmh.annotations.OperationsPerInvocation; +import org.openjdk.jmh.annotations.Setup; +import org.openjdk.jmh.runner.Runner; +import org.openjdk.jmh.runner.RunnerException; +import org.openjdk.jmh.runner.options.Options; +import org.openjdk.jmh.runner.options.OptionsBuilder; +import org.openjdk.jmh.runner.options.VerboseMode; + +import java.io.IOException; + +@OperationsPerInvocation(value = UnalignedCheckpointBenchmark.RECORDS_PER_INVOCATION) Review comment: This is not correct, you are not using `RECORDS_PER_INVOCATION` to control the number of records per invocation. ## File path: src/main/java/org/apache/flink/benchmark/UnalignedCheckpointBenchmark.java ## @@ -0,0 +1,164 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.benchmark; + +import org.apache.flink.api.common.functions.MapFunction; +import org.apache.flink.configuration.ConfigConstants; +import org.apache.flink.configuration.Configuration; +import org.apache.flink.configuration.TaskManagerOptions; +import org.apache.flink.runtime.state.CheckpointListener; +import org.apache.flink.streaming.api.datastream.DataStreamSource; +import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment; +import org.apache.flink.streaming.api.functions.sink.SinkFunction; +import org.apache.flink.streaming.api.functions.source.RichParallelSourceFunction; + +import org.openjdk.jmh.annotations.Benchmark; +import org.openjdk.jmh.annotations.OperationsPerInvocation; +import org.openjdk.jmh.annotations.Setup; +import org.openjdk.jmh.runner.Runner; +import org.openjdk.jmh.runner.RunnerException; +import org.openjdk.jmh.runner.options.Options; +import org.openjdk.jmh.runner.options.OptionsBuilder; +import org.openjdk.jmh.runner.options.VerboseMode; + +import java.io.IOException; + +@OperationsPerInvocation(value = UnalignedCheckpointBenchmark.RECORDS_PER_INVOCATION) +public class UnalignedCheckpointBenchmark extends BenchmarkBase { +public static final int RECORDS_PER_INVOCATION = 10_000_000; +private static final int NUM_VERTICES = 3; +private static final int PARALLELISM = 4; +private static final long CHECKPOINT_INTERVAL_MS = 100; + +public static void main(String[] args) throws RunnerException { +Options options = new OptionsBuilder() +.verbosity(VerboseMode.NORMAL) +.include(UnalignedCheckpointBenchmark.class.getCanonicalName()) +
[GitHub] [flink] flinkbot edited a comment on pull request #13024: [FLINK-18742][flink-clients] Some configuration args do not take effect at client
flinkbot edited a comment on pull request #13024: URL: https://github.com/apache/flink/pull/13024#issuecomment-666148245 ## CI report: * eb611d5b39f997fa3e986b5d163cb65a44b4b0ba UNKNOWN * a5db01bd0f63289039dfcdae41eaf099ed28a812 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5730) * 047cc5ec9e12641f2787ac6fa4a3525fd6fc7d70 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink-web] rmetzger commented on a change in pull request #370: Add a blog post about the current state of Flink on Docker
rmetzger commented on a change in pull request #370: URL: https://github.com/apache/flink-web/pull/370#discussion_r473850331 ## File path: _posts/2020-08-20-flink-docker.md ## @@ -0,0 +1,90 @@ +--- +layout: post +title: "The State of Flink on Docker" +date: 2020-08-08T00:00:00.000Z +authors: +- rmetzger: + name: "Robert Metzger" + twitter: rmetzger_ +categories: news + +excerpt: This blog post gives an update on the recent developments of Flink's support for Docker. +--- + +The Flink community recently put some effort into upgrading the Docker experience for our users. The goal was to reduce confusion and improve usability. With over 50 million downloads from Docker Hub, the Flink docker images are a very popular deployment option. Review comment: ... elevating the Docker experience to a new level ... ? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink-web] rmetzger commented on a change in pull request #370: Add a blog post about the current state of Flink on Docker
rmetzger commented on a change in pull request #370: URL: https://github.com/apache/flink-web/pull/370#discussion_r473849162 ## File path: _posts/2020-08-20-flink-docker.md ## @@ -0,0 +1,90 @@ +--- +layout: post +title: "The State of Flink on Docker" +date: 2020-08-08T00:00:00.000Z +authors: +- rmetzger: + name: "Robert Metzger" + twitter: rmetzger_ +categories: news + +excerpt: This blog post gives an update on the recent developments of Flink's support for Docker. +--- + +The Flink community recently put some effort into upgrading the Docker experience for our users. The goal was to reduce confusion and improve usability. With over 50 million downloads from Docker Hub, the Flink docker images are a very popular deployment option. Review comment: Seth didn't correct the word in his review This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13206: [FLINK-18948][python] Add end to end test for Python DataStream API.
flinkbot edited a comment on pull request #13206: URL: https://github.com/apache/flink/pull/13206#issuecomment-677500353 ## CI report: * bea57ee386d4ffdd264e3cc0116a653454103c42 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5741) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13204: [FLINK-16080][docs-zh]Translate \docs\dev\table\hive\hive_streaming.zh.md into Chinese
flinkbot edited a comment on pull request #13204: URL: https://github.com/apache/flink/pull/13204#issuecomment-677254693 ## CI report: * f3ff67413489210527790afa12a9f50b63ced883 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5737) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink-web] rmetzger commented on a change in pull request #370: Add a blog post about the current state of Flink on Docker
rmetzger commented on a change in pull request #370: URL: https://github.com/apache/flink-web/pull/370#discussion_r473846656 ## File path: _posts/2020-08-20-flink-docker.md ## @@ -0,0 +1,90 @@ +--- +layout: post +title: "The State of Flink on Docker" +date: 2020-08-08T00:00:00.000Z +authors: +- rmetzger: + name: "Robert Metzger" + twitter: rmetzger_ +categories: news + +excerpt: This blog post gives an update on the recent developments of Flink's support for Docker. +--- + +The Flink community recently put some effort into upgrading the Docker experience for our users. The goal was to reduce confusion and improve usability. With over 50 million downloads from Docker Hub, the Flink docker images are a very popular deployment option. Review comment: Damn, me trying to sound eloquent often fires back :) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-12351) AsyncWaitOperator should deep copy StreamElement when object reuse is enabled
[ https://issues.apache.org/jira/browse/FLINK-12351?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181081#comment-17181081 ] Wenlong Lyu commented on FLINK-12351: - [~pnowojski] As I known, it is easy to get whether the op is the head of chain by: StreamConfig#isChainStart > AsyncWaitOperator should deep copy StreamElement when object reuse is enabled > - > > Key: FLINK-12351 > URL: https://issues.apache.org/jira/browse/FLINK-12351 > Project: Flink > Issue Type: Bug > Components: API / DataStream >Reporter: Jark Wu >Assignee: Jark Wu >Priority: Major > Labels: pull-request-available > Time Spent: 10m > Remaining Estimate: 0h > > Currently, AsyncWaitOperator directly put the input StreamElement into > {{StreamElementQueue}}. But when object reuse is enabled, the StreamElement > is reused, which means the element in {{StreamElementQueue}} will be > modified. As a result, the output of AsyncWaitOperator might be wrong. > An easy way to fix this might be deep copy the input StreamElement when > object reuse is enabled, like this: > https://github.com/apache/flink/blob/blink/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/operators/async/AsyncWaitOperator.java#L209 -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #13206: [FLINK-18948][python] Add end to end test for Python DataStream API.
flinkbot commented on pull request #13206: URL: https://github.com/apache/flink/pull/13206#issuecomment-677500353 ## CI report: * bea57ee386d4ffdd264e3cc0116a653454103c42 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] RocMarshal commented on pull request #13089: [FLINK-18813][docs-zh] Translate the 'Local Installation' page of 'Try Flink' into Chinese
RocMarshal commented on pull request #13089: URL: https://github.com/apache/flink/pull/13089#issuecomment-677495556 ping @xccui This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] kl0u commented on pull request #6613: [FLINK-9940] [API/DataStream][File source] out-of-order files were missed in continuous monitoring
kl0u commented on pull request #6613: URL: https://github.com/apache/flink/pull/6613#issuecomment-677494490 Sorry @lvhuyen for closing this, I am reopening it because I just noticed there is discussion on the related JIRA. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] lvhuyen opened a new pull request #6613: [FLINK-9940] [API/DataStream][File source] out-of-order files were missed in continuous monitoring
lvhuyen opened a new pull request #6613: URL: https://github.com/apache/flink/pull/6613 [FLINK-9940] Fix - File-source continuous monitoring mode - out-of-order files were missed ## Fix the issue with ContinuousFileMonitoringFunction - out-of-order files were missed in continuous directory scanning mode. - _Cause_: In the existing directory monitoring mechanism, Flink was maintaining the maximum last-modified-timestamp of all identified files (_globalModificationTime_) so that in the next scan, all files with last-modified-timestamp equal or earlier than that _globalModificationTime_ will be ignored. - _Fix_: This fix provides an additional param when creating a ContinuousFileMonitoringFunction: readConsistencyOffset. Every scan now starts from that max last-modified-timestamp minus this offset. A new list of processedFiles is also maintained, which consists of all known files having modTimestamp in that offset period. - For testing this fix, a change to flink-fs-tests has also been made: The collection of seenFiles is changed from a TreeSet to a SortedList. This change is to verify the ExactOnce of file scanning, instead of AtLeastOnce. ## Verifying this change This change is already covered by existing tests with slight update. - ContinuousFileProcessingMigrationTest.testMonitoringSourceRestore. - ContinuousFileProcessingTest.{testFunctionRestore, testProcessContinuously} This change also added test: - ContinuousFileProcessingTest.testProcessContinuouslyWithNoteTooLateFile ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): no - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: yes - The serializers: no - The runtime per-record code paths (performance sensitive): yes (per-file). This is expected to have minimal impact. - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: no - The S3 file system connector: no ## Documentation - Does this pull request introduce a new feature? yes - If yes, how is the feature documented? JavaDocs This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] zhuxiaoshang commented on pull request #13162: [FLINK-18685][API / DataStream]JobClient.getAccumulators() blocks until streaming job has finished in local environment
zhuxiaoshang commented on pull request #13162: URL: https://github.com/apache/flink/pull/13162#issuecomment-677492835 @rmetzger Sorry for the failing,what i think is make the getAccumulators() can be async invoked.But the actual behavior is not my expect. When i debug the code,the returned CompletableFuture of 'miniCluster.getExecutionGraph(jobID)' is 'not completed',so the accumulators is empty. Could you give some suggestions to me?Thanks a lot. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #13206: [FLINK-18948][python] Add end to end test for Python DataStream API.
flinkbot commented on pull request #13206: URL: https://github.com/apache/flink/pull/13206#issuecomment-677491849 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit bea57ee386d4ffdd264e3cc0116a653454103c42 (Thu Aug 20 09:37:21 UTC 2020) **Warnings:** * No documentation files were touched! Remember to keep the Flink docs up to date! Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] kl0u closed pull request #6613: [FLINK-9940] [API/DataStream][File source] out-of-order files were missed in continuous monitoring
kl0u closed pull request #6613: URL: https://github.com/apache/flink/pull/6613 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] kl0u commented on pull request #6613: [FLINK-9940] [API/DataStream][File source] out-of-order files were missed in continuous monitoring
kl0u commented on pull request #6613: URL: https://github.com/apache/flink/pull/6613#issuecomment-677491493 I'm closing this as "Abandoned", since there is no more activity. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-18948) Add end to end test for Python DataStream API
[ https://issues.apache.org/jira/browse/FLINK-18948?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-18948: --- Labels: pull-request-available (was: ) > Add end to end test for Python DataStream API > - > > Key: FLINK-18948 > URL: https://issues.apache.org/jira/browse/FLINK-18948 > Project: Flink > Issue Type: Sub-task > Components: API / Python >Reporter: Hequn Cheng >Assignee: Shuiqiang Chen >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0 > > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] shuiqiangchen opened a new pull request #13206: [FLINK-18948][python] Add end to end test for Python DataStream API.
shuiqiangchen opened a new pull request #13206: URL: https://github.com/apache/flink/pull/13206 ## What is the purpose of the change Add end to end test for Python DataStream API. ## Brief change log Added a simple Python DataStream Job, which will consume datas from specific Kafka topic, after a series of transformations, then, write the result data back to Kafka. ## Verifying this change This pull request is an end to end test, it will be started by running test_python_datastream.sh script. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): ( no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: ( no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (no) - The S3 file system connector: (no) ## Documentation - Does this pull request introduce a new feature? (no) - If yes, how is the feature documented? ( not documented) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Closed] (FLINK-18962) Improve error message if checkpoint directory is not writable
[ https://issues.apache.org/jira/browse/FLINK-18962?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Piotr Nowojski closed FLINK-18962. -- Fix Version/s: 1.12.0 Resolution: Fixed Merged to master as f8ce30a50b^^..f8ce30a50b. Thanks for submitting the idea [~NicoK] and addressing the issue [~roman_khachatryan] :) > Improve error message if checkpoint directory is not writable > - > > Key: FLINK-18962 > URL: https://issues.apache.org/jira/browse/FLINK-18962 > Project: Flink > Issue Type: Improvement > Components: Runtime / Checkpointing >Affects Versions: 1.11.1 >Reporter: Nico Kruber >Assignee: Roman Khachatryan >Priority: Major > Labels: pull-request-available, usability > Fix For: 1.12.0 > > > If the checkpoint directory from {{state.checkpoints.dir}} is not writable by > the user that Flink is running with, checkpoints will be declined, but the > real cause is not mentioned anywhere: > * the Web UI says: "Cause: The job has failed" (the Flink job is running > though) > * the JM log says: > {code} > 2020-08-14 12:13:18,820 INFO > org.apache.flink.runtime.checkpoint.CheckpointCoordinator[] - Triggering > checkpoint 2 (type=CHECKPOINT) @ 159738819 for job > 2c567b14e8d0833404931ef47dfec266. > 2020-08-14 12:13:18,921 INFO > org.apache.flink.runtime.checkpoint.CheckpointCoordinator[] - Decline > checkpoint 2 by task 0d4fd75374ad16c8d963679e3c2171ec of job > 2c567b14e8d0833404931ef47dfec266 at a184deea621e3923fbfcb1d899348448 @ > Nico-PC.lan (dataPort=35531). > {code} > * the TM log says: > {code} > 2020-08-14 12:13:14,102 INFO > org.apache.flink.streaming.runtime.tasks.SubtaskCheckpointCoordinatorImpl [] > - Checkpoint 1 has been notified as aborted, would not trigger any checkpoint. > {code} > And that's it. It should have a real error message indicating that the > checkpoint (sub)-directory could not be created. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] pnowojski merged pull request #13180: [FLINK-18962][checkpointing] Improve logging when checkpoint declined
pnowojski merged pull request #13180: URL: https://github.com/apache/flink/pull/13180 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] pnowojski commented on a change in pull request #13180: [FLINK-18962][checkpointing] Improve logging when checkpoint declined
pnowojski commented on a change in pull request #13180: URL: https://github.com/apache/flink/pull/13180#discussion_r473803693 ## File path: flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/tasks/AsyncCheckpointRunnable.java ## @@ -129,12 +129,10 @@ public void run() { checkpointMetaData.getCheckpointId()); } } catch (Exception e) { - if (LOG.isDebugEnabled()) { - LOG.debug("{} - asynchronous part of checkpoint {} could not be completed.", - taskName, - checkpointMetaData.getCheckpointId(), - e); - } + LOG.info("{} - asynchronous part of checkpoint {} could not be completed.", Review comment: > Honestly, I have not had this to me, I’m not against this change here. Ok, in that case let's not overthink it and let's try this out :) Thanks for your inputs @NicoK and @klion26 . This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13175: [FLINK-18955][Checkpointing] Add checkpoint path to job startup/restore message
flinkbot edited a comment on pull request #13175: URL: https://github.com/apache/flink/pull/13175#issuecomment-674839926 ## CI report: * 635f839466122e36674a38a9845c2d6b5eb5c244 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5735) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13188: [hotfix][docs] Translate building.md to building.zh.md
flinkbot edited a comment on pull request #13188: URL: https://github.com/apache/flink/pull/13188#issuecomment-675323573 ## CI report: * 8bbc06f7a8931c9e02130f668805a82ccc00a9e8 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5738) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink-web] azagrebin commented on a change in pull request #370: Add a blog post about the current state of Flink on Docker
azagrebin commented on a change in pull request #370: URL: https://github.com/apache/flink-web/pull/370#discussion_r473789842 ## File path: _posts/2020-08-20-flink-docker.md ## @@ -0,0 +1,90 @@ +--- +layout: post +title: "The State of Flink on Docker" +date: 2020-08-08T00:00:00.000Z +authors: +- rmetzger: + name: "Robert Metzger" + twitter: rmetzger_ +categories: news + +excerpt: This blog post gives an update on the recent developments of Flink's support for Docker. +--- + +The Flink community recently put some effort into upgrading the Docker experience for our users. The goal was to reduce confusion and improve usability. With over 50 million downloads from Docker Hub, the Flink docker images are a very popular deployment option. + +Let's quickly break down the recent improvements: + +- Reduce confusion: Flink used to have 2 Dockerfiles and a 3rd file maintained outside of the official repository — all with different features and varying stability. Now, we have one central place for all images: [apache/flink-docker](https://github.com/apache/flink-docker). + + Here, we keep all the Dockerfiles for the different releases. Check out the [detailed readme](https://github.com/apache/flink-docker/blob/master/README.md) of that repository for further explanation on the different branches, as well as the [Flink Improvement Proposal (FLIP-111)](https://cwiki.apache.org/confluence/display/FLINK/FLIP-111%3A+Docker+image+unification) that contains the detailed planning. + + The apache/flink-docker repository also seeds the o[fficial Flink image on Docker Hub](https://hub.docker.com/_/flink). + +- Improve Usability: The Dockerfiles are used for various purposes: [Native docker deployments](https://ci.apache.org/projects/flink/flink-docs-master/ops/deployment/docker.html), [Flink on Kubernetes](https://ci.apache.org/projects/flink/flink-docs-master/ops/deployment/native_kubernetes.html), the (unofficial) [Flink helm example](https://github.com/docker-flink/examples) and the project's [internal end to end tests](https://github.com/apache/flink/tree/master/flink-end-to-end-tests). With one unified image, all these consumers of the images benefit from the same set of features, documentation and testing. + + The new images support passing configuration variables via a `FLINK_PROPERTIES` environment variable. Users can enable default plugins with the `ENABLE_BUILT_IN_PLUGINS`environment variable. The images also allow loading custom jar paths and configuration files. + +Looking into the future, there are already some interesting potential improvements lined up: + +- [Java 11 Docker images](https://issues.apache.org/jira/browse/FLINK-16260) (already completed) +- [Use vanilla docker-entrypoint with flink-kubernetes](https://issues.apache.org/jira/browse/FLINK-15793) (in progress) +- [History server support](https://issues.apache.org/jira/browse/FLINK-17167) +- [Support for OpenShift](https://issues.apache.org/jira/browse/FLINK-15587) + +## How do I get started? + +This is a short tutorial on how to start a Flink Session Cluster with docker. Review comment: again I would either make `a Flink Session Cluster with docker` clickable into the docs or add `see also details in docs link`. ## File path: _posts/2020-08-20-flink-docker.md ## @@ -0,0 +1,90 @@ +--- +layout: post +title: "The State of Flink on Docker" +date: 2020-08-08T00:00:00.000Z +authors: +- rmetzger: + name: "Robert Metzger" + twitter: rmetzger_ +categories: news + +excerpt: This blog post gives an update on the recent developments of Flink's support for Docker. +--- + +The Flink community recently put some effort into upgrading the Docker experience for our users. The goal was to reduce confusion and improve usability. With over 50 million downloads from Docker Hub, the Flink docker images are a very popular deployment option. + +Let's quickly break down the recent improvements: + +- Reduce confusion: Flink used to have 2 Dockerfiles and a 3rd file maintained outside of the official repository — all with different features and varying stability. Now, we have one central place for all images: [apache/flink-docker](https://github.com/apache/flink-docker). + + Here, we keep all the Dockerfiles for the different releases. Check out the [detailed readme](https://github.com/apache/flink-docker/blob/master/README.md) of that repository for further explanation on the different branches, as well as the [Flink Improvement Proposal (FLIP-111)](https://cwiki.apache.org/confluence/display/FLINK/FLIP-111%3A+Docker+image+unification) that contains the detailed planning. + + The apache/flink-docker repository also seeds the o[fficial Flink image on Docker Hub](https://hub.docker.com/_/flink). Review comment: ```suggestion The `apache/flink-docker` repository also seeds the [official Flink image on Docker Hub](https://hub.docker.com/_/flink). ``` ## File
[jira] [Commented] (FLINK-18643) Migrate Jenkins jobs to ci-builds.apache.org
[ https://issues.apache.org/jira/browse/FLINK-18643?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181055#comment-17181055 ] Robert Metzger commented on FLINK-18643: I've disabled the Jenkins profiles for Flink (except the statefun snapshot deployment ones) > Migrate Jenkins jobs to ci-builds.apache.org > > > Key: FLINK-18643 > URL: https://issues.apache.org/jira/browse/FLINK-18643 > Project: Flink > Issue Type: Improvement > Components: Release System >Reporter: Chesnay Schepler >Assignee: Robert Metzger >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0, 1.11.2 > > > Infra is [reworking the Jenkins > setup|https://lists.apache.org/thread.html/re974eed417a1bc294694701d5c91b4bf92689fcf32a4c91f169be87d%40%3Cbuilds.apache.org%3E], > so we have to migrate our jobs that do the snapshot deployments. > Alternatively, find other ways to do this (Azure?) to reduce number of used > infrastructure services. > /cc [~rmetzger] -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot edited a comment on pull request #13205: [FLINK-17330[runtime] Merge cyclic dependent pipelined regions into one region
flinkbot edited a comment on pull request #13205: URL: https://github.com/apache/flink/pull/13205#issuecomment-677463515 ## CI report: * e11ffec55b9151857069d64c806b47cc98d9679d Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5739) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #13205: [FLINK-17330[runtime] Merge cyclic dependent pipelined regions into one region
flinkbot commented on pull request #13205: URL: https://github.com/apache/flink/pull/13205#issuecomment-677463515 ## CI report: * e11ffec55b9151857069d64c806b47cc98d9679d UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] AHeise commented on a change in pull request #13175: [FLINK-18955][Checkpointing] Add checkpoint path to job startup/restore message
AHeise commented on a change in pull request #13175: URL: https://github.com/apache/flink/pull/13175#discussion_r473766209 ## File path: flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CompletedCheckpoint.java ## @@ -320,6 +320,12 @@ void setDiscardCallback(@Nullable CompletedCheckpointStats.DiscardCallback disca @Override public String toString() { - return String.format("Checkpoint %d @ %d for %s", checkpointID, timestamp, job); + return String.format( + "%s %d @ %d for %s located at %s", + props.getCheckpointType(), Review comment: Currently that would display: "CHECKPOINT" "SAVEPOINT" "SYNC_SAVEPOINT" I'm proposing to translate them into just "Checkpoint" and "Savepoint" to reduce the technicality. However, no strong feelings, so you can also convince me that it's better to leave as is. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #13205: [FLINK-17330[runtime] Merge cyclic dependent pipelined regions into one region
flinkbot commented on pull request #13205: URL: https://github.com/apache/flink/pull/13205#issuecomment-677459348 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit e11ffec55b9151857069d64c806b47cc98d9679d (Thu Aug 20 08:33:48 UTC 2020) **Warnings:** * No documentation files were touched! Remember to keep the Flink docs up to date! Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] zhuzhurk opened a new pull request #13205: [FLINK-17330[runtime] Merge cyclic dependent pipelined regions into one region
zhuzhurk opened a new pull request #13205: URL: https://github.com/apache/flink/pull/13205 ## What is the purpose of the change This PR changes regions building to merge cyclic dependent pipelined regions into one region. This is to avoid scheduling deadlocks due to cyclic dependencies. More details see FLINK-17330. ## Brief change log - *Added StronglyConnectedComponentsComputeUtils to find out cyclic dependent regions* - *Changed PipelinedRegionComputeUtil to merge cyclic dependent pipelined regions into one region* ## Verifying this change - *Added UT StronglyConnectedComponentsComputeUtilsTest* - *Added test case in PipelinedRegionComputeUtilTest to verify regions on the same cycles are merged* ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (yes / **no**) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (yes / **no**) - The serializers: (yes / **no** / don't know) - The runtime per-record code paths (performance sensitive): (yes / **no** / don't know) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (**yes** / no / don't know) - The S3 file system connector: (yes / **no** / don't know) ## Documentation - Does this pull request introduce a new feature? (yes / **no**) - If yes, how is the feature documented? (**not applicable** / docs / JavaDocs / not documented) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-17330) Avoid scheduling deadlocks caused by cyclic input dependencies between regions
[ https://issues.apache.org/jira/browse/FLINK-17330?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-17330: --- Labels: pull-request-available (was: ) > Avoid scheduling deadlocks caused by cyclic input dependencies between regions > -- > > Key: FLINK-17330 > URL: https://issues.apache.org/jira/browse/FLINK-17330 > Project: Flink > Issue Type: Sub-task > Components: Runtime / Coordination >Reporter: Zhu Zhu >Assignee: Zhu Zhu >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0 > > > Imagine a job like this: > A -- (pipelined FORWARD) --> B -- (blocking ALL-to-ALL) --> D > A -- (pipelined FORWARD) --> C -- (pipelined FORWARD) --> D > parallelism=2 for all vertices. > We will have 2 execution pipelined regions: > R1 = {A1, B1, C1, D1} > R2 = {A2, B2, C2, D2} > R1 has a cross-region input edge (B2->D1). > R2 has a cross-region input edge (B1->D2). > Scheduling deadlock will happen since we schedule a region only when all its > inputs are consumable (i.e. blocking partitions to be finished). This is > because R1 can be scheduled only if R2 finishes, while R2 can be scheduled > only if R1 finishes. > To avoid this, one solution is to force a logical pipelined region with > intra-region ALL-to-ALL blocking edges to form one only execution pipelined > region, so that there would not be cyclic input dependency between regions. > Besides that, we should also pay attention to avoid cyclic cross-region > POINTWISE blocking edges. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Closed] (FLINK-18960) flink sideoutput union
[ https://issues.apache.org/jira/browse/FLINK-18960?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dawid Wysakowicz closed FLINK-18960. Fix Version/s: 1.10.2 Resolution: Duplicate > flink sideoutput union > -- > > Key: FLINK-18960 > URL: https://issues.apache.org/jira/browse/FLINK-18960 > Project: Flink > Issue Type: Bug > Components: API / DataStream >Affects Versions: 1.10.1 >Reporter: xiaohang.li >Priority: Minor > Fix For: 1.10.2 > > > Flink sideoutput union seems not works right. If we union the sideoutput from > the same operator, the output is the result of last side output times by the > number of unions, which is not expected. For example, > {code:java} > val side = new OutputTag[String]("side") > val side2 = new OutputTag[String]("side2") > val side3 = new OutputTag[String]("side3") > val ds = env.socketTextStream("master",9001) > val res = ds.process(new ProcessFunction[String,String] { > override def processElement(value: String, ctx: ProcessFunction[String, > String]#Context, out: Collector[String]): Unit = { > if(value.contains("hello")) > { ctx.output(side,value) } > else if(value.contains("world")) > { ctx.output(side2,value) } > else if(value.contains("flink")) > { ctx.output(side3,value) } > out.collect(value) > } > }) > val res1 = res.getSideOutput(side) > val res2 = res.getSideOutput(side2) > val res3 = res.getSideOutput(side3) > println( ">"+res1.getClass) > println( ">"+res2.getClass) > res1.print("res1") > res2.print("res2") > res3.print("res3") > res2.union(res1).union(res3).print("all") > {code} > > If we input > {code:java} > hello > world > flink > {code} > The output will be > > {code:java} > res1> hello > res2> world > res3> flink > all> flink > all> flink > all> flink > {code} > > But the expected output would be > {code:java} > res1> hello > res2> world > res3> flink > all> hello > all> world > all> flink > {code} > > > if we add a _map_ after the sideoutput and then union them, the output would > be right, but adding map should be not needed. > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] kl0u commented on pull request #13024: [FLINK-18742][flink-clients] Some configuration args do not take effect at client
kl0u commented on pull request #13024: URL: https://github.com/apache/flink/pull/13024#issuecomment-677440255 @wangzzu There is a test failure related to your changes. Please fix it and then I think I can merge :) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13188: [hotfix][docs] Translate building.md to building.zh.md
flinkbot edited a comment on pull request #13188: URL: https://github.com/apache/flink/pull/13188#issuecomment-675323573 ## CI report: * 8301e1e53770c95e2f903526ac93e80e039cf4f2 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5733) * 8bbc06f7a8931c9e02130f668805a82ccc00a9e8 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5738) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-18960) flink sideoutput union
[ https://issues.apache.org/jira/browse/FLINK-18960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181004#comment-17181004 ] Yun Gao commented on FLINK-18960: - hi [~xiaohang.li], I checked the issue, it does existed in old versions, but it should be fixed in https://issues.apache.org/jira/browse/FLINK-17578, you may try to upgrade to 1.10.2 or 1.11. > flink sideoutput union > -- > > Key: FLINK-18960 > URL: https://issues.apache.org/jira/browse/FLINK-18960 > Project: Flink > Issue Type: Bug > Components: API / DataStream >Affects Versions: 1.10.1 >Reporter: xiaohang.li >Priority: Minor > > Flink sideoutput union seems not works right. If we union the sideoutput from > the same operator, the output is the result of last side output times by the > number of unions, which is not expected. For example, > {code:java} > val side = new OutputTag[String]("side") > val side2 = new OutputTag[String]("side2") > val side3 = new OutputTag[String]("side3") > val ds = env.socketTextStream("master",9001) > val res = ds.process(new ProcessFunction[String,String] { > override def processElement(value: String, ctx: ProcessFunction[String, > String]#Context, out: Collector[String]): Unit = { > if(value.contains("hello")) > { ctx.output(side,value) } > else if(value.contains("world")) > { ctx.output(side2,value) } > else if(value.contains("flink")) > { ctx.output(side3,value) } > out.collect(value) > } > }) > val res1 = res.getSideOutput(side) > val res2 = res.getSideOutput(side2) > val res3 = res.getSideOutput(side3) > println( ">"+res1.getClass) > println( ">"+res2.getClass) > res1.print("res1") > res2.print("res2") > res3.print("res3") > res2.union(res1).union(res3).print("all") > {code} > > If we input > {code:java} > hello > world > flink > {code} > The output will be > > {code:java} > res1> hello > res2> world > res3> flink > all> flink > all> flink > all> flink > {code} > > But the expected output would be > {code:java} > res1> hello > res2> world > res3> flink > all> hello > all> world > all> flink > {code} > > > if we add a _map_ after the sideoutput and then union them, the output would > be right, but adding map should be not needed. > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-18800) Avro serialization schema doesn't support Kafka key/value serialization
[ https://issues.apache.org/jira/browse/FLINK-18800?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17181003#comment-17181003 ] Mohammad Hossein Gerami commented on FLINK-18800: - anybody doesn't have any idea? > Avro serialization schema doesn't support Kafka key/value serialization > > > Key: FLINK-18800 > URL: https://issues.apache.org/jira/browse/FLINK-18800 > Project: Flink > Issue Type: Improvement > Components: Connectors / Kafka, Formats (JSON, Avro, Parquet, ORC, > SequenceFile) >Affects Versions: 1.11.0, 1.11.1 >Reporter: Mohammad Hossein Gerami >Priority: Major > > {color:#ff8b00}AvroSerializationSchema{color} and > {color:#ff8b00}ConfluentRegistryAvroSerializationSchema{color} doesn't > support Kafka key/value serialization. I implemented a custom Avro > serialization schema for solving this problem. > for example in the Flink must implement a class like this. > {code:java} > public class KafkaAvroRegistrySchemaSerializationSchema extends > RegistryAvroSerializationSchema implements > KafkaSerializationSchema{code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-18800) Avro serialization schema doesn't support Kafka key/value serialization
[ https://issues.apache.org/jira/browse/FLINK-18800?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mohammad Hossein Gerami updated FLINK-18800: Description: {color:#ff8b00}AvroSerializationSchema{color} and {color:#ff8b00}ConfluentRegistryAvroSerializationSchema{color} doesn't support Kafka key/value serialization. I implemented a custom Avro serialization schema for solving this problem. for example in the Flink must implement a class like this. {code:java} public class KafkaAvroRegistrySchemaSerializationSchema extends RegistryAvroSerializationSchema implements KafkaSerializationSchema{code} was:{color:#ff8b00}[AvroSerializationSchema|[https://ci.apache.org/projects/flink/flink-docs-stable/api/java/org/apache/flink/formats/avro/AvroDeserializationSchema.html]]{color} and {color:#ff8b00}[ConfluentRegistryAvroSerializationSchema|[https://ci.apache.org/projects/flink/flink-docs-stable/api/java/org/apache/flink/formats/avro/registry/confluent/ConfluentRegistryAvroSerializationSchema.html]]{color} doesn't support Kafka key/value serialization. I implemented a custom Avro serialization schema for solving this problem. > Avro serialization schema doesn't support Kafka key/value serialization > > > Key: FLINK-18800 > URL: https://issues.apache.org/jira/browse/FLINK-18800 > Project: Flink > Issue Type: Improvement > Components: Connectors / Kafka, Formats (JSON, Avro, Parquet, ORC, > SequenceFile) >Affects Versions: 1.11.0, 1.11.1 >Reporter: Mohammad Hossein Gerami >Priority: Major > > {color:#ff8b00}AvroSerializationSchema{color} and > {color:#ff8b00}ConfluentRegistryAvroSerializationSchema{color} doesn't > support Kafka key/value serialization. I implemented a custom Avro > serialization schema for solving this problem. > for example in the Flink must implement a class like this. > {code:java} > public class KafkaAvroRegistrySchemaSerializationSchema extends > RegistryAvroSerializationSchema implements > KafkaSerializationSchema{code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] tillrohrmann commented on a change in pull request #13190: [FLINK-18986][kubernetes] Create ClusterClient only for attached deployments
tillrohrmann commented on a change in pull request #13190: URL: https://github.com/apache/flink/pull/13190#discussion_r473711834 ## File path: flink-kubernetes/src/main/java/org/apache/flink/kubernetes/cli/KubernetesSessionCli.java ## @@ -88,6 +88,7 @@ public Configuration getEffectiveConfiguration(String[] args) throws CliArgsExce private int run(String[] args) throws FlinkException, CliArgsException { final Configuration configuration = getEffectiveConfiguration(args); + KubernetesClusterClientFactory.ensureClusterIdIsSet(configuration); Review comment: Hmm, yeah one should either get the `clusterId` from the user or after a cluster is deployed via the `ClusterClient` or `ClusterClientProvider`. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] tillrohrmann commented on a change in pull request #13190: [FLINK-18986][kubernetes] Create ClusterClient only for attached deployments
tillrohrmann commented on a change in pull request #13190: URL: https://github.com/apache/flink/pull/13190#discussion_r473709875 ## File path: flink-kubernetes/src/main/java/org/apache/flink/kubernetes/KubernetesClusterClientFactory.java ## @@ -51,10 +51,7 @@ public boolean isCompatibleWith(Configuration configuration) { @Override public KubernetesClusterDescriptor createClusterDescriptor(Configuration configuration) { checkNotNull(configuration); - if (!configuration.contains(KubernetesConfigOptions.CLUSTER_ID)) { - final String clusterId = generateClusterId(); - configuration.setString(KubernetesConfigOptions.CLUSTER_ID, clusterId); - } + ensureClusterIdIsSet(configuration); return new KubernetesClusterDescriptor(configuration, KubeClientFactory.fromConfiguration(configuration)); Review comment: I think the idea was that the `ClusterDescriptor` will generate a cluster id when you call any of the `deploy*` methods. I believe what is set in the constructor of the `YarnClusterDescriptor` is fine and does not contradict what I was saying. If one really needs to configure the cluster id manually, then one could extend the `ClusterDescriptor` interface accordingly. I think that we are adding technical debt by working around broken contracts which might be fine now but we should definitely pull this straight eventually. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13188: [hotfix][docs] Translate building.md to building.zh.md
flinkbot edited a comment on pull request #13188: URL: https://github.com/apache/flink/pull/13188#issuecomment-675323573 ## CI report: * 8301e1e53770c95e2f903526ac93e80e039cf4f2 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5733) * 8bbc06f7a8931c9e02130f668805a82ccc00a9e8 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-18643) Migrate Jenkins jobs to ci-builds.apache.org
[ https://issues.apache.org/jira/browse/FLINK-18643?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17180998#comment-17180998 ] Chesnay Schepler commented on FLINK-18643: -- hmm...I guess it's okay to only cover 1.11+ with this. Go ahead and remove the jenkins setup. > Migrate Jenkins jobs to ci-builds.apache.org > > > Key: FLINK-18643 > URL: https://issues.apache.org/jira/browse/FLINK-18643 > Project: Flink > Issue Type: Improvement > Components: Release System >Reporter: Chesnay Schepler >Assignee: Robert Metzger >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0, 1.11.2 > > > Infra is [reworking the Jenkins > setup|https://lists.apache.org/thread.html/re974eed417a1bc294694701d5c91b4bf92689fcf32a4c91f169be87d%40%3Cbuilds.apache.org%3E], > so we have to migrate our jobs that do the snapshot deployments. > Alternatively, find other ways to do this (Azure?) to reduce number of used > infrastructure services. > /cc [~rmetzger] -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] rkhachatryan commented on a change in pull request #13180: [FLINK-18962][checkpointing] Improve logging when checkpoint declined
rkhachatryan commented on a change in pull request #13180: URL: https://github.com/apache/flink/pull/13180#discussion_r473689643 ## File path: flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/tasks/AsyncCheckpointRunnable.java ## @@ -129,12 +129,10 @@ public void run() { checkpointMetaData.getCheckpointId()); } } catch (Exception e) { - if (LOG.isDebugEnabled()) { - LOG.debug("{} - asynchronous part of checkpoint {} could not be completed.", - taskName, - checkpointMetaData.getCheckpointId(), - e); - } + LOG.info("{} - asynchronous part of checkpoint {} could not be completed.", Review comment: I think everything said above about failure frequency is also true for expirations (in fact, there is only one counter for all types of failures). This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org