[jira] [Assigned] (FLINK-34102) Invalid configuration when using 'env.log.max' on yarn application mode

2024-01-15 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34102?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan reassigned FLINK-34102:
---

Assignee: RocMarshal

> Invalid configuration when using 'env.log.max' on yarn application mode
> ---
>
> Key: FLINK-34102
> URL: https://issues.apache.org/jira/browse/FLINK-34102
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Configuration
>Affects Versions: 1.19.0
>Reporter: RocMarshal
>Assignee: RocMarshal
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33988][configuration] Fix the invalid configuration when using initialized root logger level on yarn deployment mode [flink]

2024-01-15 Thread via GitHub


1996fanrui commented on code in PR #24042:
URL: https://github.com/apache/flink/pull/24042#discussion_r1453038598


##
flink-yarn/src/main/java/org/apache/flink/yarn/YarnClusterDescriptor.java:
##
@@ -216,6 +229,24 @@ public YarnClusterDescriptor(
 this.nodeLabel = 
flinkConfiguration.getString(YarnConfigOptions.NODE_LABEL);
 }
 
+/** Adapt flink env setting. */
+private static  void adaptEnvSetting(
+Configuration config,
+ConfigOption configOption,
+String envKey,
+Function validator) {

Review Comment:
   As I understand, this `validator` wanna check whether the value of 
`configOption` in config is legal. 
   
   - If it's legal, we can add the env
   - If it's illegal, we will ignore.
   
   Is my understanding right?
   
   If right, I don't think we should do this check here. I mean we should cover 
Flink on Kubernetes if we wanna check whether it's legal. So check them in the 
yarn side doesn't make sense to me.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [revert][FLINK-33907][ci] Makes copying test jars being done in the package phase [flink]

2024-01-15 Thread via GitHub


XComp commented on PR #24095:
URL: https://github.com/apache/flink/pull/24095#issuecomment-1893225487

   I responded in the 
[FLINK-33907](https://issues.apache.org/jira/browse/FLINK-33907?focusedCommentId=17807087=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-17807087)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-34106) user defined source parallelism in ddl can't work if connector ScanRuntimeProvider is SourceFunctionProvider

2024-01-15 Thread Zhanghao Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34106?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807088#comment-17807088
 ] 

Zhanghao Chen commented on FLINK-34106:
---

[~Akihito Liang] Are you refering to the new API introduced in 
[FLIP-367|[https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=263429150]?]
 If so, it is still undergoing, but we'll make it in Flink 1.19. Please 
subscribe  FLINK-33261 for the latest progress

> user defined source parallelism in ddl can't work if connector 
> ScanRuntimeProvider is SourceFunctionProvider
> 
>
> Key: FLINK-34106
> URL: https://issues.apache.org/jira/browse/FLINK-34106
> Project: Flink
>  Issue Type: Bug
>Reporter: Junning Liang
>Priority: Critical
>
> I implemented a connector using the SourceFunctionProvider myself and found 
> that setting scan parallelism is not effective.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-34106) user defined source parallelism in ddl can't work if connector ScanRuntimeProvider is SourceFunctionProvider

2024-01-15 Thread Zhanghao Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34106?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807088#comment-17807088
 ] 

Zhanghao Chen edited comment on FLINK-34106 at 1/16/24 7:54 AM:


[~Akihito Liang] Are you refering to the new API introduced in 
[FLIP-367|[https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=263429150]]?
 If so, it is still undergoing, but we'll make it in Flink 1.19. Please 
subscribe  FLINK-33261 for the latest progress


was (Author: zhanghao chen):
[~Akihito Liang] Are you refering to the new API introduced in 
[FLIP-367|[https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=263429150]?]
 If so, it is still undergoing, but we'll make it in Flink 1.19. Please 
subscribe  FLINK-33261 for the latest progress

> user defined source parallelism in ddl can't work if connector 
> ScanRuntimeProvider is SourceFunctionProvider
> 
>
> Key: FLINK-34106
> URL: https://issues.apache.org/jira/browse/FLINK-34106
> Project: Flink
>  Issue Type: Bug
>Reporter: Junning Liang
>Priority: Critical
>
> I implemented a connector using the SourceFunctionProvider myself and found 
> that setting scan parallelism is not effective.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-34106) user defined source parallelism in ddl can't work if connector ScanRuntimeProvider is SourceFunctionProvider

2024-01-15 Thread Zhanghao Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34106?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807088#comment-17807088
 ] 

Zhanghao Chen edited comment on FLINK-34106 at 1/16/24 7:54 AM:


[~Akihito Liang] Are you refering to the new API introduced in 
[FLIP-367|https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=263429150]?
 If so, it is still undergoing, but we'll make it in Flink 1.19. Please 
subscribe  FLINK-33261 for the latest progress


was (Author: zhanghao chen):
[~Akihito Liang] Are you refering to the new API introduced in 
[FLIP-367|[https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=263429150]]?
 If so, it is still undergoing, but we'll make it in Flink 1.19. Please 
subscribe  FLINK-33261 for the latest progress

> user defined source parallelism in ddl can't work if connector 
> ScanRuntimeProvider is SourceFunctionProvider
> 
>
> Key: FLINK-34106
> URL: https://issues.apache.org/jira/browse/FLINK-34106
> Project: Flink
>  Issue Type: Bug
>Reporter: Junning Liang
>Priority: Critical
>
> I implemented a connector using the SourceFunctionProvider myself and found 
> that setting scan parallelism is not effective.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33907) Makes copying test jars being done later

2024-01-15 Thread Matthias Pohl (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807087#comment-17807087
 ] 

Matthias Pohl commented on FLINK-33907:
---

I had a discussion with [~chesnay] on that issue. His argument from the comment 
above is reasonable. We would need to improve the Maven setup to make the 
individual test execution work in this specific case (either through the 
utilization of the failsafe plugin or by introducing another profile). This is 
covered by FLINK-34107. There is a (admittedly not so convenient) workaround 
documented in FLINK-34107. Alternatively, you can run {{mvn -pl flink-client 
verify}}.

The actual goal of FLINK-33907 is to allow for an improved performance of the 
build pipeline. That outweighs the need to run two tests individually on the 
command line in my opinion. [~gyfora] WDYT?

> Makes copying test jars being done later
> 
>
> Key: FLINK-33907
> URL: https://issues.apache.org/jira/browse/FLINK-33907
> Project: Flink
>  Issue Type: Sub-task
>  Components: Build System / CI
>Affects Versions: 1.18.0, 1.19.0
>Reporter: Matthias Pohl
>Assignee: Matthias Pohl
>Priority: Major
>  Labels: github-actions, pull-request-available
> Fix For: 1.19.0, 1.18.1
>
>
> We experienced an issue in GHA which is due to the fact how test resources 
> are pre-computed in GHA:
> {code:java}
> This fixes the following error when compiling flink-clients:
> Error: 2.054 [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-dependency-plugin:3.2.0:copy-dependencies 
> (copy-dependencies) on project flink-clients: Artifact has not been packaged 
> yet. When used on reactor artifact, copy should be executed after packaging: 
> see MDEP-187. -> [Help 1] {code}
> We need to move this goal to a later phase.
> The reason why this popped up is (as far as I remember) that we do only do 
> test-compile in GitHub Actions.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33988][configuration] Fix the invalid configuration when using initialized root logger level on yarn deployment mode [flink]

2024-01-15 Thread via GitHub


1996fanrui commented on code in PR #24042:
URL: https://github.com/apache/flink/pull/24042#discussion_r1453038598


##
flink-yarn/src/main/java/org/apache/flink/yarn/YarnClusterDescriptor.java:
##
@@ -216,6 +229,24 @@ public YarnClusterDescriptor(
 this.nodeLabel = 
flinkConfiguration.getString(YarnConfigOptions.NODE_LABEL);
 }
 
+/** Adapt flink env setting. */
+private static  void adaptEnvSetting(
+Configuration config,
+ConfigOption configOption,
+String envKey,
+Function validator) {

Review Comment:
   As I understand, this `validator` wanna check whether the value of 
`configOption` in config is legal. 
   
   - If it's legal, we can add the env
   - If it's illegal, we will ignore.
   
   Is my understanding right?
   
   If right, I don't think we should do this check here. I means we should 
cover Flink on Kubernetes if we wanna check whether it's legal. So check them 
in the yarn side doesn't make sense to me.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-34107) Clearer separation of unit test and integration tests

2024-01-15 Thread Matthias Pohl (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807081#comment-17807081
 ] 

Matthias Pohl commented on FLINK-34107:
---

One approach here could be to use the [failsafe 
plugin|https://maven.apache.org/surefire/maven-failsafe-plugin/] for the 
integration test execution. This wasn't done so far to keep the number of 
plugins in the Flink build low (it was conveniently enough to have the test 
executions being covered by the surefire plugin).

> Clearer separation of unit test and integration tests
> -
>
> Key: FLINK-34107
> URL: https://issues.apache.org/jira/browse/FLINK-34107
> Project: Flink
>  Issue Type: Bug
>  Components: Build System, Build System / CI
>Affects Versions: 1.18.0, 1.17.2, 1.19.0
>Reporter: Matthias Pohl
>Priority: Major
>
> FLINK-33907 revealed an issue in the current Maven setup where the 
> surefire-plugin is in charge of executing both, the unit and the integration 
> tests in two separate executions: The {{-Dtest}} system parameter is 
> overwriting the {{}} and {{}} configuration of the two 
> executions. That results in a test being executed twice if specifically 
> selected for execution (i.e. {{mvn [...] verify -Dtest=???}}).
> That's especially problematic in cases where the ITCase relies on artifacts 
> that are build during the Maven run (see the discussion in FLINK-33907 where 
> {{DefaultPackagedProgramRetrieverITCase}} relies on 
> {{flink-client-test-utils}} to provide test jars). The artifacts should be 
> moved in the {{pre-integration-test}} phase from a conceptual point of view.
> A workaround for this issue is to directly call the execution 
> {{surefire:test@integration-tests}}:
> {code}
> mvn -pl flink-clients surefire:test@integration-tests 
> -Dtest=DefaultPackagedProgramRetrieverITCase -Dfast
> {code}
> This approach has the flow that the Maven lifecycle is ignored. You would 
> have to do a {{mvn -pl flink-clients clean verify}} run upfront to build all 
> the artifacts the test relies on. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34107) Clearer separation of unit test and integration tests

2024-01-15 Thread Matthias Pohl (Jira)
Matthias Pohl created FLINK-34107:
-

 Summary: Clearer separation of unit test and integration tests
 Key: FLINK-34107
 URL: https://issues.apache.org/jira/browse/FLINK-34107
 Project: Flink
  Issue Type: Bug
  Components: Build System, Build System / CI
Affects Versions: 1.17.2, 1.18.0, 1.19.0
Reporter: Matthias Pohl


FLINK-33907 revealed an issue in the current Maven setup where the 
surefire-plugin is in charge of executing both, the unit and the integration 
tests in two separate executions: The {{-Dtest}} system parameter is 
overwriting the {{}} and {{}} configuration of the two 
executions. That results in a test being executed twice if specifically 
selected for execution (i.e. {{mvn [...] verify -Dtest=???}}).

That's especially problematic in cases where the ITCase relies on artifacts 
that are build during the Maven run (see the discussion in FLINK-33907 where 
{{DefaultPackagedProgramRetrieverITCase}} relies on {{flink-client-test-utils}} 
to provide test jars). The artifacts should be moved in the 
{{pre-integration-test}} phase from a conceptual point of view.

A workaround for this issue is to directly call the execution 
{{surefire:test@integration-tests}}:
{code}
mvn -pl flink-clients surefire:test@integration-tests 
-Dtest=DefaultPackagedProgramRetrieverITCase -Dfast
{code}
This approach has the flow that the Maven lifecycle is ignored. You would have 
to do a {{mvn -pl flink-clients clean verify}} run upfront to build all the 
artifacts the test relies on. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-21949][table] Support ARRAY_AGG aggregate function [flink]

2024-01-15 Thread via GitHub


snuyanzin commented on code in PR #23411:
URL: https://github.com/apache/flink/pull/23411#discussion_r1452991614


##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/functions/ArrayAggFunctionITCase.java:
##
@@ -0,0 +1,95 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.functions;
+
+import org.apache.flink.table.functions.BuiltInFunctionDefinitions;
+import org.apache.flink.types.Row;
+
+import java.util.Arrays;
+import java.util.Collections;
+import java.util.stream.Stream;
+
+import static org.apache.flink.table.api.DataTypes.ARRAY;
+import static org.apache.flink.table.api.DataTypes.INT;
+import static org.apache.flink.table.api.DataTypes.ROW;
+import static org.apache.flink.table.api.DataTypes.STRING;
+import static org.apache.flink.table.api.Expressions.$;
+import static org.apache.flink.types.RowKind.DELETE;
+import static org.apache.flink.types.RowKind.INSERT;
+import static org.apache.flink.types.RowKind.UPDATE_AFTER;
+import static org.apache.flink.types.RowKind.UPDATE_BEFORE;
+
+/** Tests for built-in ARRAY_AGG aggregation functions. */
+class ArrayAggFunctionITCase extends BuiltInAggregateFunctionTestBase {
+
+@Override
+Stream getTestCaseSpecs() {
+return Stream.of(
+TestSpec.forFunction(BuiltInFunctionDefinitions.ARRAY_AGG)
+.withDescription("ARRAY changelog stream aggregation")
+.withSource(
+ROW(STRING(), INT()),
+Arrays.asList(
+Row.ofKind(INSERT, "A", 1),
+Row.ofKind(INSERT, "A", 2),
+Row.ofKind(INSERT, "B", 2),
+Row.ofKind(INSERT, "B", 2),
+Row.ofKind(INSERT, "B", 3),
+Row.ofKind(INSERT, "C", 3),
+Row.ofKind(INSERT, "C", null),
+Row.ofKind(INSERT, "D", null),
+Row.ofKind(INSERT, "E", 4),
+Row.ofKind(INSERT, "E", 5),
+Row.ofKind(DELETE, "E", 5),
+Row.ofKind(UPDATE_BEFORE, "E", 4),
+Row.ofKind(UPDATE_AFTER, "E", 6)))
+.testResult(
+source ->
+"SELECT f0, array_agg(f1) FROM " + 
source + " GROUP BY f0",
+TableApiAggSpec.groupBySelect(
+Collections.singletonList($("f0")),
+$("f0"),
+$("f1").arrayAgg()),
+ROW(STRING(), ARRAY(INT())),
+ROW(STRING(), ARRAY(INT())),
+Arrays.asList(
+Row.of("A", new Integer[] {1, 2}),
+Row.of("B", new Integer[] {2, 2, 3}),
+Row.of("C", new Integer[] {3}),

Review Comment:
   It's ok to support it only partially however the main concern here: 
   ~~I would expect more consistent result
   since there is also input `Row.ofKind(INSERT, "D", null),` and expected 
value `Row.of("D", null),`
   So I would expect for ths either containing nulls or both not containing 
null however not mixed~~
   
   UPD: to be more clear:
   after playing with Postgres and BigQuery I noticed that both respect nulls 
by default and for BigQuery to make it ignoring nulls it should be specified 
explicitely. Is there a reason why for Flink it is done differently? Probably 
it's better to have similar behaviour
   
   I don't tell that we need to support `RESPECT NULLS | IGNORE NULLS` syntax, 
however need to make `RESPECT NULLS` default behaviour to be on same page with 
vendors



-- 
This is an automated message from the Apache Git 

Re: [PR] [FLINK-21949][table] Support ARRAY_AGG aggregate function [flink]

2024-01-15 Thread via GitHub


snuyanzin commented on code in PR #23411:
URL: https://github.com/apache/flink/pull/23411#discussion_r1452991614


##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/functions/ArrayAggFunctionITCase.java:
##
@@ -0,0 +1,95 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.functions;
+
+import org.apache.flink.table.functions.BuiltInFunctionDefinitions;
+import org.apache.flink.types.Row;
+
+import java.util.Arrays;
+import java.util.Collections;
+import java.util.stream.Stream;
+
+import static org.apache.flink.table.api.DataTypes.ARRAY;
+import static org.apache.flink.table.api.DataTypes.INT;
+import static org.apache.flink.table.api.DataTypes.ROW;
+import static org.apache.flink.table.api.DataTypes.STRING;
+import static org.apache.flink.table.api.Expressions.$;
+import static org.apache.flink.types.RowKind.DELETE;
+import static org.apache.flink.types.RowKind.INSERT;
+import static org.apache.flink.types.RowKind.UPDATE_AFTER;
+import static org.apache.flink.types.RowKind.UPDATE_BEFORE;
+
+/** Tests for built-in ARRAY_AGG aggregation functions. */
+class ArrayAggFunctionITCase extends BuiltInAggregateFunctionTestBase {
+
+@Override
+Stream getTestCaseSpecs() {
+return Stream.of(
+TestSpec.forFunction(BuiltInFunctionDefinitions.ARRAY_AGG)
+.withDescription("ARRAY changelog stream aggregation")
+.withSource(
+ROW(STRING(), INT()),
+Arrays.asList(
+Row.ofKind(INSERT, "A", 1),
+Row.ofKind(INSERT, "A", 2),
+Row.ofKind(INSERT, "B", 2),
+Row.ofKind(INSERT, "B", 2),
+Row.ofKind(INSERT, "B", 3),
+Row.ofKind(INSERT, "C", 3),
+Row.ofKind(INSERT, "C", null),
+Row.ofKind(INSERT, "D", null),
+Row.ofKind(INSERT, "E", 4),
+Row.ofKind(INSERT, "E", 5),
+Row.ofKind(DELETE, "E", 5),
+Row.ofKind(UPDATE_BEFORE, "E", 4),
+Row.ofKind(UPDATE_AFTER, "E", 6)))
+.testResult(
+source ->
+"SELECT f0, array_agg(f1) FROM " + 
source + " GROUP BY f0",
+TableApiAggSpec.groupBySelect(
+Collections.singletonList($("f0")),
+$("f0"),
+$("f1").arrayAgg()),
+ROW(STRING(), ARRAY(INT())),
+ROW(STRING(), ARRAY(INT())),
+Arrays.asList(
+Row.of("A", new Integer[] {1, 2}),
+Row.of("B", new Integer[] {2, 2, 3}),
+Row.of("C", new Integer[] {3}),

Review Comment:
   It's ok to support it only partially however the main concern here: 
   ~~I would expect more consistent result
   since there is also input `Row.ofKind(INSERT, "D", null),` and expected 
value `Row.of("D", null),`
   So I would expect for ths either containing nulls or both not containing 
null however not mixed~~
   
   UPD: to be more clear:
   after playing with Postgres and BigQuery I noticed that both respect nulls 
by default and for BigQuery to make it ignoring nulls it should be specified 
explicitely. Is there a reason why for Flink it is done differently? Probably 
it's better to have similar behaviour



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: 

Re: [PR] [FLINK-21949][table] Support ARRAY_AGG aggregate function [flink]

2024-01-15 Thread via GitHub


snuyanzin commented on code in PR #23411:
URL: https://github.com/apache/flink/pull/23411#discussion_r1452991614


##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/functions/ArrayAggFunctionITCase.java:
##
@@ -0,0 +1,95 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.functions;
+
+import org.apache.flink.table.functions.BuiltInFunctionDefinitions;
+import org.apache.flink.types.Row;
+
+import java.util.Arrays;
+import java.util.Collections;
+import java.util.stream.Stream;
+
+import static org.apache.flink.table.api.DataTypes.ARRAY;
+import static org.apache.flink.table.api.DataTypes.INT;
+import static org.apache.flink.table.api.DataTypes.ROW;
+import static org.apache.flink.table.api.DataTypes.STRING;
+import static org.apache.flink.table.api.Expressions.$;
+import static org.apache.flink.types.RowKind.DELETE;
+import static org.apache.flink.types.RowKind.INSERT;
+import static org.apache.flink.types.RowKind.UPDATE_AFTER;
+import static org.apache.flink.types.RowKind.UPDATE_BEFORE;
+
+/** Tests for built-in ARRAY_AGG aggregation functions. */
+class ArrayAggFunctionITCase extends BuiltInAggregateFunctionTestBase {
+
+@Override
+Stream getTestCaseSpecs() {
+return Stream.of(
+TestSpec.forFunction(BuiltInFunctionDefinitions.ARRAY_AGG)
+.withDescription("ARRAY changelog stream aggregation")
+.withSource(
+ROW(STRING(), INT()),
+Arrays.asList(
+Row.ofKind(INSERT, "A", 1),
+Row.ofKind(INSERT, "A", 2),
+Row.ofKind(INSERT, "B", 2),
+Row.ofKind(INSERT, "B", 2),
+Row.ofKind(INSERT, "B", 3),
+Row.ofKind(INSERT, "C", 3),
+Row.ofKind(INSERT, "C", null),
+Row.ofKind(INSERT, "D", null),
+Row.ofKind(INSERT, "E", 4),
+Row.ofKind(INSERT, "E", 5),
+Row.ofKind(DELETE, "E", 5),
+Row.ofKind(UPDATE_BEFORE, "E", 4),
+Row.ofKind(UPDATE_AFTER, "E", 6)))
+.testResult(
+source ->
+"SELECT f0, array_agg(f1) FROM " + 
source + " GROUP BY f0",
+TableApiAggSpec.groupBySelect(
+Collections.singletonList($("f0")),
+$("f0"),
+$("f1").arrayAgg()),
+ROW(STRING(), ARRAY(INT())),
+ROW(STRING(), ARRAY(INT())),
+Arrays.asList(
+Row.of("A", new Integer[] {1, 2}),
+Row.of("B", new Integer[] {2, 2, 3}),
+Row.of("C", new Integer[] {3}),

Review Comment:
   It's ok to support it only partially however the main concern here: 
   ~~I would expect more consistent result
   since there is also input `Row.ofKind(INSERT, "D", null),` and expected 
value `Row.of("D", null),`
   So I would expect for ths either containing nulls or both not containing 
null however not mixed~~
   
   UPD: to be more clear:
   after playing with Postgres and BigQuery I noticed that both respect nulls 
by default and for BigQuery to make it ignoring nulls it should be specified 
explicitely. Is there a reason why for Flink it is done differently?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please 

Re: [PR] [FLINK-33450][autoscaler] Support the JDBCAutoScalerStateStore [flink-kubernetes-operator]

2024-01-15 Thread via GitHub


1996fanrui commented on code in PR #741:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/741#discussion_r1453006111


##
flink-autoscaler-plugin-jdbc/src/main/java/org/apache/flink/autoscaler/jdbc/state/JdbcAutoScalerStateStore.java:
##
@@ -0,0 +1,286 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.autoscaler.jdbc.state;
+
+import org.apache.flink.annotation.Experimental;
+import org.apache.flink.annotation.VisibleForTesting;
+import org.apache.flink.autoscaler.JobAutoScalerContext;
+import org.apache.flink.autoscaler.ScalingSummary;
+import org.apache.flink.autoscaler.ScalingTracking;
+import org.apache.flink.autoscaler.metrics.CollectedMetrics;
+import org.apache.flink.autoscaler.state.AutoScalerStateStore;
+import org.apache.flink.autoscaler.utils.AutoScalerSerDeModule;
+import org.apache.flink.configuration.ConfigurationUtils;
+import org.apache.flink.runtime.jobgraph.JobVertexID;
+
+import com.fasterxml.jackson.core.JacksonException;
+import com.fasterxml.jackson.core.type.TypeReference;
+import com.fasterxml.jackson.databind.ObjectMapper;
+import com.fasterxml.jackson.datatype.jsr310.JavaTimeModule;
+import org.apache.commons.io.IOUtils;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import javax.annotation.Nonnull;
+
+import java.io.ByteArrayInputStream;
+import java.io.ByteArrayOutputStream;
+import java.io.IOException;
+import java.nio.charset.StandardCharsets;
+import java.time.Instant;
+import java.util.Base64;
+import java.util.HashMap;
+import java.util.Map;
+import java.util.Optional;
+import java.util.SortedMap;
+import java.util.TreeMap;
+import java.util.zip.GZIPInputStream;
+import java.util.zip.GZIPOutputStream;
+
+import static 
org.apache.flink.autoscaler.jdbc.state.StateType.COLLECTED_METRICS;
+import static 
org.apache.flink.autoscaler.jdbc.state.StateType.PARALLELISM_OVERRIDES;
+import static org.apache.flink.autoscaler.jdbc.state.StateType.SCALING_HISTORY;
+import static 
org.apache.flink.autoscaler.jdbc.state.StateType.SCALING_TRACKING;
+
+/**
+ * The state store which persists its state in JDBC related database.
+ *
+ * @param  The job key.
+ * @param  The job autoscaler context.
+ */
+@Experimental
+public class JdbcAutoScalerStateStore>
+implements AutoScalerStateStore {
+
+private static final Logger LOG = 
LoggerFactory.getLogger(JdbcAutoScalerStateStore.class);
+
+private final JdbcStateStore jdbcStateStore;
+
+protected static final ObjectMapper YAML_MAPPER =
+new ObjectMapper()
+.registerModule(new JavaTimeModule())
+.registerModule(new AutoScalerSerDeModule());
+
+public JdbcAutoScalerStateStore(JdbcStateStore jdbcStateStore) {
+this.jdbcStateStore = jdbcStateStore;
+}
+
+@Override
+public void storeScalingHistory(
+Context jobContext, Map> scalingHistory)
+throws Exception {
+jdbcStateStore.putSerializedState(
+getSerializeKey(jobContext),
+SCALING_HISTORY,
+serializeScalingHistory(scalingHistory));
+}
+
+@Nonnull
+@Override
+public Map> 
getScalingHistory(
+Context jobContext) {
+Optional serializedScalingHistory =
+jdbcStateStore.getSerializedState(getSerializeKey(jobContext), 
SCALING_HISTORY);
+if (serializedScalingHistory.isEmpty()) {
+return new HashMap<>();
+}
+try {
+return deserializeScalingHistory(serializedScalingHistory.get());
+} catch (JacksonException e) {
+LOG.error(
+"Could not deserialize scaling history, possibly the 
format changed. Discarding...",
+e);
+jdbcStateStore.removeSerializedState(getSerializeKey(jobContext), 
SCALING_HISTORY);
+return new HashMap<>();
+}
+}
+
+@Override
+public void removeScalingHistory(Context jobContext) {
+jdbcStateStore.removeSerializedState(getSerializeKey(jobContext), 
SCALING_HISTORY);
+}
+
+@Override
+public void storeScalingTracking(Context jobContext, ScalingTracking 
scalingTrack)
+  

Re: [PR] [FLINK-33450][autoscaler] Support the JDBCAutoScalerStateStore [flink-kubernetes-operator]

2024-01-15 Thread via GitHub


1996fanrui commented on code in PR #741:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/741#discussion_r1453004963


##
docs/content/docs/custom-resource/autoscaler.md:
##
@@ -260,17 +260,46 @@ job.autoscaler.metrics.window : 3m
 > `ScalingReport` will show the recommended parallelism for each vertex.
 
 After the flink job starts, please start the StandaloneAutoscaler process by 
the
-following command.
+following command. Please download released autoscaler-standalone jar from 
+[here](https://repo.maven.apache.org/maven2/org/apache/flink/flink-autoscaler-standalone/)
 first.
 
 ```
 java -cp flink-autoscaler-standalone-{{< version >}}.jar \
 org.apache.flink.autoscaler.standalone.StandaloneAutoscalerEntrypoint \
---flinkClusterHost localhost \
---flinkClusterPort 8081
+--autoscaler.standalone.fetcher.flink-cluster.host localhost \
+--autoscaler.standalone.fetcher.flink-cluster.port 8081
 ```
 
-Updating the `flinkClusterHost` and `flinkClusterPort` based on your flink 
cluster.
-In general, the host and port are the same as Flink WebUI.
+Updating the `autoscaler.standalone.fetcher.flink-cluster.host` and 
`autoscaler.standalone.fetcher.flink-cluster.port` 
+based on your flink cluster. In general, the host and port are the same as 
Flink WebUI.
+
+### Using the JDBC Autoscaler State Store
+
+A driver dependency is required to connect to a specified database. Here are 
drivers currently supported, 
+please download JDBC driver and initialize database and table first.
+
+| Driver | Group Id   | Artifact Id| JAR   
  | Schema  
|
+|:---|:---|:---|:|-|
+| MySQL  | `mysql`| `mysql-connector-java` | 
[Download](https://repo.maven.apache.org/maven2/mysql/mysql-connector-java/)
| [Table 
DDL](https://github.com/apache/flink-kubernetes-operator/blob/main/flink-autoscaler-plugin-jdbc/src/main/resources/schema/mysql_schema.sql)
 |
+| PostgreSQL | `org.postgresql`   | `postgresql`   | 
[Download](https://jdbc.postgresql.org/download/)   
| [Table 
DDL](https://github.com/apache/flink-kubernetes-operator/blob/main/flink-autoscaler-plugin-jdbc/src/main/resources/schema/postgres_schema.sql)
  |
+| Derby  | `org.apache.derby` | `derby`| 
[Download](http://db.apache.org/derby/derby_downloads.html) 
| [Table 
DDL](https://github.com/apache/flink-kubernetes-operator/blob/main/flink-autoscaler-plugin-jdbc/src/main/resources/schema/derby_schema.sql)
 |
+
+```
+JDBC_DRIVER_JAR=./mysql-connector-java-8.0.30.jar
+# export the password of jdbc state store
+export STATE_STORE_JDBC_PWD=123456
+
+java -cp flink-autoscaler-standalone-{{< version >}}.jar:${JDBC_DRIVER_JAR} \
+org.apache.flink.autoscaler.standalone.StandaloneAutoscalerEntrypoint \
+--autoscaler.standalone.fetcher.flink-cluster.host localhost \
+--autoscaler.standalone.fetcher.flink-cluster.port 8081 \
+--autoscaler.standalone.state-store.type jdbc \
+--autoscaler.standalone.state-store.jdbc.url 
jdbc:mysql://localhost:3306/flink_autoscaler \
+--autoscaler.standalone.state-store.jdbc.username root

Review Comment:
   In general, the environment variable name doesn't need to be changed. Users 
need to export the password using this environment variable.
   
   So I didn't mention `password-env-variable` here. But in the beginning of 
this doc, it mentioned how to export password. WDYT?
   
   ```
   # export the password of jdbc state store
   export STATE_STORE_JDBC_PWD=123456
   ```



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Closed] (FLINK-33980) Reorganize job configuration

2024-01-15 Thread Zhu Zhu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33980?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhu Zhu closed FLINK-33980.
---
Fix Version/s: 1.19.0
   Resolution: Done

master/release-1.19:
06b37089f0c1cdf70ca21970a40d15c3eaba07ed
290b633c4265540d481ac563454c7f4a3e706b9d
65b8b3baad6a27e6253a93701508ba24dc8fcfe0
d02ef1cebf302c56a0d9d51664d2c7fb6f5be932
eb8af0c589ce46b091f403e848c7dc84b7e3ee8b

> Reorganize job configuration
> 
>
> Key: FLINK-33980
> URL: https://issues.apache.org/jira/browse/FLINK-33980
> Project: Flink
>  Issue Type: Technical Debt
>  Components: API / Core
>Reporter: Junrui Li
>Assignee: Junrui Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0
>
>
> Currently, job configuration in FLINK is spread out across different 
> components, including StreamExecutionEnvironment, CheckpointConfig, and 
> ExecutionConfig. This distribution leads to inconsistencies among the 
> configurations stored within these components. Furthermore, the methods used 
> to configure these components vary; some rely on complex Java objects, while 
> others use ConfigOption, which is a key-value configuration approach. This 
> variation complicates the effective management of job configurations. 
> Additionally, passing complex Java objects (e.g., StateBackend and 
> CheckpointStorage) between the environment, StreamGraph, and JobGraph adds 
> complexity to development.
> With the completion of FLIP-381, it is now time to standardize and unify job 
> configuration in FLINK. The goals of this JIRA are as follows:
>  # Migrate configuration from non-ConfigOption objects to use ConfigOption.
>  # Adopt a single Configuration object to house all configurations.
>  # Create complex Java objects, such as RestartBackoffTimeStrategyFactory, 
> CheckpointStorage, and StateBackend, directly from the configuration on the 
> JM side.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33980][core] Reorganize job configuration [flink]

2024-01-15 Thread via GitHub


zhuzhurk closed pull request #24025: [FLINK-33980][core] Reorganize job 
configuration
URL: https://github.com/apache/flink/pull/24025


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-21949][table] Support ARRAY_AGG aggregate function [flink]

2024-01-15 Thread via GitHub


snuyanzin commented on code in PR #23411:
URL: https://github.com/apache/flink/pull/23411#discussion_r1452991614


##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/functions/ArrayAggFunctionITCase.java:
##
@@ -0,0 +1,95 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.functions;
+
+import org.apache.flink.table.functions.BuiltInFunctionDefinitions;
+import org.apache.flink.types.Row;
+
+import java.util.Arrays;
+import java.util.Collections;
+import java.util.stream.Stream;
+
+import static org.apache.flink.table.api.DataTypes.ARRAY;
+import static org.apache.flink.table.api.DataTypes.INT;
+import static org.apache.flink.table.api.DataTypes.ROW;
+import static org.apache.flink.table.api.DataTypes.STRING;
+import static org.apache.flink.table.api.Expressions.$;
+import static org.apache.flink.types.RowKind.DELETE;
+import static org.apache.flink.types.RowKind.INSERT;
+import static org.apache.flink.types.RowKind.UPDATE_AFTER;
+import static org.apache.flink.types.RowKind.UPDATE_BEFORE;
+
+/** Tests for built-in ARRAY_AGG aggregation functions. */
+class ArrayAggFunctionITCase extends BuiltInAggregateFunctionTestBase {
+
+@Override
+Stream getTestCaseSpecs() {
+return Stream.of(
+TestSpec.forFunction(BuiltInFunctionDefinitions.ARRAY_AGG)
+.withDescription("ARRAY changelog stream aggregation")
+.withSource(
+ROW(STRING(), INT()),
+Arrays.asList(
+Row.ofKind(INSERT, "A", 1),
+Row.ofKind(INSERT, "A", 2),
+Row.ofKind(INSERT, "B", 2),
+Row.ofKind(INSERT, "B", 2),
+Row.ofKind(INSERT, "B", 3),
+Row.ofKind(INSERT, "C", 3),
+Row.ofKind(INSERT, "C", null),
+Row.ofKind(INSERT, "D", null),
+Row.ofKind(INSERT, "E", 4),
+Row.ofKind(INSERT, "E", 5),
+Row.ofKind(DELETE, "E", 5),
+Row.ofKind(UPDATE_BEFORE, "E", 4),
+Row.ofKind(UPDATE_AFTER, "E", 6)))
+.testResult(
+source ->
+"SELECT f0, array_agg(f1) FROM " + 
source + " GROUP BY f0",
+TableApiAggSpec.groupBySelect(
+Collections.singletonList($("f0")),
+$("f0"),
+$("f1").arrayAgg()),
+ROW(STRING(), ARRAY(INT())),
+ROW(STRING(), ARRAY(INT())),
+Arrays.asList(
+Row.of("A", new Integer[] {1, 2}),
+Row.of("B", new Integer[] {2, 2, 3}),
+Row.of("C", new Integer[] {3}),

Review Comment:
   It's ok to support it only partially however the main concern here: I would 
expect more consistent result
   since there is also input `Row.ofKind(INSERT, "D", null),` and expected 
value `Row.of("D", null),`
   So I would expect for ths either containing nulls or both not containing 
null however not mixed



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Resolved] (FLINK-33712) FLIP-391: Deprecate RuntimeContext#getExecutionConfig

2024-01-15 Thread Junrui Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33712?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junrui Li resolved FLINK-33712.
---
Fix Version/s: 1.19.0
 Release Note: RuntimeContext#getExecutionConfig is now be deprecated in 
FLINK-1.19. And this method is planned to be removed in Flink-2.0. More details 
can be found at 
https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=278465937.
   Resolution: Fixed

> FLIP-391: Deprecate RuntimeContext#getExecutionConfig
> -
>
> Key: FLINK-33712
> URL: https://issues.apache.org/jira/browse/FLINK-33712
> Project: Flink
>  Issue Type: Technical Debt
>  Components: API / Core
>Reporter: Junrui Li
>Assignee: Junrui Li
>Priority: Major
> Fix For: 1.19.0
>
>
> Deprecate the RuntimeContext#getExecutionConfig and  introduce alternative 
> getter methods that allow users to access specific information without 
> exposing unnecessary runtime details. More details see: 
> [FLIP-391|https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=278465937]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34106) user defined source parallelism in ddl can't work if connector ScanRuntimeProvider is SourceFunctionProvider

2024-01-15 Thread Junning Liang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34106?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807065#comment-17807065
 ] 

Junning Liang commented on FLINK-34106:
---

I have fixed the issue, please assign it to me~

> user defined source parallelism in ddl can't work if connector 
> ScanRuntimeProvider is SourceFunctionProvider
> 
>
> Key: FLINK-34106
> URL: https://issues.apache.org/jira/browse/FLINK-34106
> Project: Flink
>  Issue Type: Bug
>Reporter: Junning Liang
>Priority: Critical
>
> I implemented a connector using the SourceFunctionProvider myself and found 
> that setting scan parallelism is not effective.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34106) user defined source parallelism in ddl can't work if connector ScanRuntimeProvider is SourceFunctionProvider

2024-01-15 Thread Junning Liang (Jira)
Junning Liang created FLINK-34106:
-

 Summary: user defined source parallelism in ddl can't work if 
connector ScanRuntimeProvider is SourceFunctionProvider
 Key: FLINK-34106
 URL: https://issues.apache.org/jira/browse/FLINK-34106
 Project: Flink
  Issue Type: Bug
Reporter: Junning Liang


I implemented a connector using the SourceFunctionProvider myself and found 
that setting scan parallelism is not effective.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33803] Set observedGeneration at end of reconciliation [flink-kubernetes-operator]

2024-01-15 Thread via GitHub


gyfora commented on code in PR #755:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/755#discussion_r1452979030


##
flink-kubernetes-operator-api/src/main/java/org/apache/flink/kubernetes/operator/api/status/FlinkDeploymentStatus.java:
##
@@ -55,4 +55,7 @@ public class FlinkDeploymentStatus extends 
CommonStatus {
 
 /** Information about the TaskManagers for the scale subresource. */
 private TaskManagerInfo taskManager;
+
+/** Last observed generation of the FlinkDeployment. */
+private Long observedGeneration;

Review Comment:
   @justin-chen I would prefer to replace the implementation and call it out in 
the release notes instead. That way we end up with a cleaner codebase.
   
   Any concerns from your end @mxm ?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-21949][table] Support ARRAY_AGG aggregate function [flink]

2024-01-15 Thread via GitHub


Jiabao-Sun commented on PR #23411:
URL: https://github.com/apache/flink/pull/23411#issuecomment-1893141857

   Thanks @snuyanzin @dawidwys for the review.
   Could you help review it again?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-21949][table] Support ARRAY_AGG aggregate function [flink]

2024-01-15 Thread via GitHub


Jiabao-Sun commented on code in PR #23411:
URL: https://github.com/apache/flink/pull/23411#discussion_r1452975551


##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/functions/ArrayAggFunctionITCase.java:
##
@@ -0,0 +1,95 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.functions;
+
+import org.apache.flink.table.functions.BuiltInFunctionDefinitions;
+import org.apache.flink.types.Row;
+
+import java.util.Arrays;
+import java.util.Collections;
+import java.util.stream.Stream;
+
+import static org.apache.flink.table.api.DataTypes.ARRAY;
+import static org.apache.flink.table.api.DataTypes.INT;
+import static org.apache.flink.table.api.DataTypes.ROW;
+import static org.apache.flink.table.api.DataTypes.STRING;
+import static org.apache.flink.table.api.Expressions.$;
+import static org.apache.flink.types.RowKind.DELETE;
+import static org.apache.flink.types.RowKind.INSERT;
+import static org.apache.flink.types.RowKind.UPDATE_AFTER;
+import static org.apache.flink.types.RowKind.UPDATE_BEFORE;
+
+/** Tests for built-in ARRAY_AGG aggregation functions. */
+class ArrayAggFunctionITCase extends BuiltInAggregateFunctionTestBase {
+
+@Override
+Stream getTestCaseSpecs() {
+return Stream.of(
+TestSpec.forFunction(BuiltInFunctionDefinitions.ARRAY_AGG)
+.withDescription("ARRAY changelog stream aggregation")
+.withSource(
+ROW(STRING(), INT()),
+Arrays.asList(
+Row.ofKind(INSERT, "A", 1),
+Row.ofKind(INSERT, "A", 2),
+Row.ofKind(INSERT, "B", 2),
+Row.ofKind(INSERT, "B", 2),
+Row.ofKind(INSERT, "B", 3),
+Row.ofKind(INSERT, "C", 3),
+Row.ofKind(INSERT, "C", null),
+Row.ofKind(INSERT, "D", null),
+Row.ofKind(INSERT, "E", 4),
+Row.ofKind(INSERT, "E", 5),
+Row.ofKind(DELETE, "E", 5),
+Row.ofKind(UPDATE_BEFORE, "E", 4),
+Row.ofKind(UPDATE_AFTER, "E", 6)))
+.testResult(
+source ->
+"SELECT f0, array_agg(f1) FROM " + 
source + " GROUP BY f0",
+TableApiAggSpec.groupBySelect(
+Collections.singletonList($("f0")),
+$("f0"),
+$("f1").arrayAgg()),
+ROW(STRING(), ARRAY(INT())),
+ROW(STRING(), ARRAY(INT())),
+Arrays.asList(
+Row.of("A", new Integer[] {1, 2}),
+Row.of("B", new Integer[] {2, 2, 3}),
+Row.of("C", new Integer[] {3}),

Review Comment:
   ```sql
   -- calcite
   ARRAY_AGG([ ALL | DISTINCT ] value [ RESPECT NULLS | IGNORE NULLS ] [ ORDER 
BY orderItem [, orderItem ]* ] )
   -- flink
   ARRAY_AGG([ ALL | DISTINCT ] expression)
   ```
   
   This function simplifies some aspects compared to calcite. 
   1. Currently, the Flink parser does not support parsing RESPECT NULLS | 
IGNORE NULLS, but it is still possible to make changes to support it.
   2. AggregateUtil#extractDistinctInformation 848~858 ignores the ignoreNulls 
fields. 
https://github.com/apache/flink/blob/6bdb4f752adb2b43dbadd8ad4fffcb4c00568dd3/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/utils/AggregateUtil.scala#L848-L858
   3. ListView does not support null values. 

Re: [PR] [FLINK-33634] Add Conditions to Flink CRD's Status field [flink-kubernetes-operator]

2024-01-15 Thread via GitHub


lajith2006 commented on code in PR #749:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/749#discussion_r1452972476


##
flink-kubernetes-operator-api/src/main/java/org/apache/flink/kubernetes/operator/api/utils/ConditionUtils.java:
##
@@ -0,0 +1,52 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.kubernetes.operator.api.utils;
+
+import io.fabric8.kubernetes.api.model.Condition;
+import io.fabric8.kubernetes.api.model.ConditionBuilder;
+
+import java.text.SimpleDateFormat;
+import java.util.Date;
+
+/** Status of CR. */
+public class ConditionUtils {
+
+public static Condition ready(final String message) {
+return crCondition("Ready", "True", message, "Ready");
+}
+
+public static Condition notReady(final String message) {
+return crCondition("Ready", "False", message, "Progressing");
+}
+
+public static Condition error(final String message) {
+return crCondition("Error", "True", message, "The job terminally 
failed");
+}
+
+public static Condition crCondition(
+final String type, final String status, final String message, 
final String reason) {
+return new ConditionBuilder()
+.withType(type)
+.withStatus(status)
+.withMessage(message)
+.withReason(reason)

Review Comment:
   As per the doc , https://maelvls.dev/kubernetes-conditions/ ,  looks like we 
can keep reason as simple category of cause of the current status.
   
   `Reason is intended to be used in concise output, such as one-line kubectl 
get output, and in summarizing occurrences of causes, whereas Message is 
intended to be presented to users in detailed status explanations, such as 
kubectl describe output.`
   
   I was thinking probably  we can keep that for any not ready and error 
conditions, and not required for ready condition?. 
   
   



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-34088) fix the problems with special table name characters of postgres and oracle and sqlserver.

2024-01-15 Thread blackpighe (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34088?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807062#comment-17807062
 ] 

blackpighe commented on FLINK-34088:


https://github.com/apache/flink-connector-jdbc/pull/89

> fix the problems with special table name characters of postgres and oracle  
> and sqlserver.
> --
>
> Key: FLINK-34088
> URL: https://issues.apache.org/jira/browse/FLINK-34088
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC
>Affects Versions: jdbc-3.1.1
>Reporter: blackpighe
>Priority: Major
> Fix For: jdbc-3.1.1
>
> Attachments: image-2024-01-15-17-10-57-586.png, 
> image-2024-01-15-17-11-11-136.png, image-2024-01-15-17-13-16-448.png
>
>   Original Estimate: 48h
>  Remaining Estimate: 48h
>
>  
> !image-2024-01-15-17-11-11-136.png!  
> [https://github.com/apache/flink-connector-jdbc/]
> The new version of flink jdbc is stand-alone and does not have an open issue 
> entry.
>  
> The FLink-JDBC-connector does not handle special table names for oracle and 
> postgres and sqlserver, when the table name is a special character such as 
> default. jdbc will report an error. hopefully giving me a chance to fix the 
> problem, I can provide a commit
>  
>  
> !image-2024-01-15-17-13-16-448.png!
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [hotfix] fix the problems with special table name characters of postgres and oracle  and sqlserver. [flink-connector-jdbc]

2024-01-15 Thread via GitHub


boring-cyborg[bot] commented on PR #89:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/89#issuecomment-1893138077

   Thanks for opening this pull request! Please check out our contributing 
guidelines. (https://flink.apache.org/contributing/how-to-contribute.html)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[PR] [hotfix] fix the problems with special table name characters of postgres and oracle  and sqlserver. [flink-connector-jdbc]

2024-01-15 Thread via GitHub


BlackPigHe opened a new pull request, #89:
URL: https://github.com/apache/flink-connector-jdbc/pull/89

   [hotfix] fix the problems with special table name characters of postgres and 
oracle  and sqlserver.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33634] Add Conditions to Flink CRD's Status field [flink-kubernetes-operator]

2024-01-15 Thread via GitHub


lajith2006 commented on code in PR #749:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/749#discussion_r1452972476


##
flink-kubernetes-operator-api/src/main/java/org/apache/flink/kubernetes/operator/api/utils/ConditionUtils.java:
##
@@ -0,0 +1,52 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.kubernetes.operator.api.utils;
+
+import io.fabric8.kubernetes.api.model.Condition;
+import io.fabric8.kubernetes.api.model.ConditionBuilder;
+
+import java.text.SimpleDateFormat;
+import java.util.Date;
+
+/** Status of CR. */
+public class ConditionUtils {
+
+public static Condition ready(final String message) {
+return crCondition("Ready", "True", message, "Ready");
+}
+
+public static Condition notReady(final String message) {
+return crCondition("Ready", "False", message, "Progressing");
+}
+
+public static Condition error(final String message) {
+return crCondition("Error", "True", message, "The job terminally 
failed");
+}
+
+public static Condition crCondition(
+final String type, final String status, final String message, 
final String reason) {
+return new ConditionBuilder()
+.withType(type)
+.withStatus(status)
+.withMessage(message)
+.withReason(reason)

Review Comment:
   As per the doc , https://maelvls.dev/kubernetes-conditions/ ,  looks like we 
can keep reason as simple category of cause of the current status.
   
   `Reason is intended to be used in concise output, such as one-line kubectl 
get output, and in summarizing occurrences of causes, whereas Message is 
intended to be presented to users in detailed status explanations, such as 
kubectl describe output.`
   
   I was thinking we can keep that for any not ready and error conditions, and 
not required for ready conditions.
   
   



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33634] Add Conditions to Flink CRD's Status field [flink-kubernetes-operator]

2024-01-15 Thread via GitHub


gyfora commented on code in PR #749:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/749#discussion_r1452970570


##
flink-kubernetes-operator-api/src/main/java/org/apache/flink/kubernetes/operator/api/status/CommonStatus.java:
##
@@ -101,4 +108,67 @@ public ResourceLifecycleState getLifecycleState() {
  * loop immediately. For example autoscaler overrides have changed and we 
need to apply them.
  */
 @JsonIgnore @Internal private boolean immediateReconciliationNeeded = 
false;
+
+public List getConditions() {
+switch (getLifecycleState()) {
+case CREATED:
+updateConditionIfNotExist(
+conditions,
+ConditionUtils.notReady(
+"The resource was created in Kubernetes but 
not yet handled by the operator"));
+break;
+case SUSPENDED:
+updateConditionIfNotExist(
+conditions,
+ConditionUtils.notReady("The resource (job) has been 
suspended"));
+break;
+case UPGRADING:
+updateConditionIfNotExist(
+conditions, ConditionUtils.notReady("The resource is 
being upgraded"));
+break;
+case DEPLOYED:
+updateConditionIfNotExist(
+conditions,
+ConditionUtils.ready(
+"The resource is deployed, but it’s not yet 
considered to be stable and might be rolled back in the future"));
+break;
+case ROLLING_BACK:
+updateConditionIfNotExist(
+conditions,
+ConditionUtils.notReady(
+"The resource is being rolled back to the last 
stable spec"));
+break;
+case ROLLED_BACK:
+updateConditionIfNotExist(
+conditions,
+ConditionUtils.ready("The resource is deployed with 
the last stable spec"));
+break;
+case FAILED:
+updateConditionIfNotExist(conditions, 
ConditionUtils.error("failed"));
+break;
+case STABLE:
+updateConditionIfNotExist(
+conditions,
+ConditionUtils.ready(
+"The resource deployment is considered to be 
stable and won’t be rolled back"));
+break;
+}
+
+return conditions;
+}
+
+private void updateConditionIfNotExist(List conditions, 
Condition newCondition) {
+if (conditions.isEmpty()) {
+conditions.add(newCondition);
+}
+if (conditions.stream()
+.noneMatch(condition -> 
condition.getType().equals(newCondition.getType( {
+conditions.add(newCondition);
+} else if (conditions.removeIf(
+condition ->
+
!(condition.getReason().equals(newCondition.getReason())
+&& 
condition.getMessage().equals(newCondition.getMessage() {
+conditions.add(newCondition);
+}

Review Comment:
   As I wrote in the last comment, I think we need a FLIP for this instead of 
fixing up this PR



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-33973) Add new interfaces for SinkV2 to synchronize the API with the SourceV2 API

2024-01-15 Thread Gyula Fora (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33973?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17805957#comment-17805957
 ] 

Gyula Fora edited comment on FLINK-33973 at 1/16/24 6:18 AM:
-

Merged to master 40cbf231e3dc15129b69136941adc15beebe8ef6


was (Author: gyfora):
Merged to master 6afe98daf6190a77a67a1fa8ac8f12337a75f8e7

> Add new interfaces for SinkV2 to synchronize the API with the SourceV2 API
> --
>
> Key: FLINK-33973
> URL: https://issues.apache.org/jira/browse/FLINK-33973
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Peter Vary
>Assignee: Peter Vary
>Priority: Major
>  Labels: pull-request-available
>
> Create the new interfaces, set inheritance and deprecation to finalize the 
> interface.
> After this change the new interafaces will exits, but they will not be 
> functional.
> The existing interfaces, and test should be working without issue, to verify 
> that adding the API will be backward compatible.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33973] Add new interfaces for SinkV2 to synchronize the API with the SourceV2 API [flink]

2024-01-15 Thread via GitHub


gyfora merged PR #24080:
URL: https://github.com/apache/flink/pull/24080


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33634] Add Conditions to Flink CRD's Status field [flink-kubernetes-operator]

2024-01-15 Thread via GitHub


lajith2006 commented on code in PR #749:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/749#discussion_r1452966020


##
flink-kubernetes-operator-api/src/main/java/org/apache/flink/kubernetes/operator/api/status/CommonStatus.java:
##
@@ -101,4 +108,67 @@ public ResourceLifecycleState getLifecycleState() {
  * loop immediately. For example autoscaler overrides have changed and we 
need to apply them.
  */
 @JsonIgnore @Internal private boolean immediateReconciliationNeeded = 
false;
+
+public List getConditions() {
+switch (getLifecycleState()) {
+case CREATED:
+updateConditionIfNotExist(
+conditions,
+ConditionUtils.notReady(
+"The resource was created in Kubernetes but 
not yet handled by the operator"));
+break;
+case SUSPENDED:
+updateConditionIfNotExist(
+conditions,
+ConditionUtils.notReady("The resource (job) has been 
suspended"));
+break;
+case UPGRADING:
+updateConditionIfNotExist(
+conditions, ConditionUtils.notReady("The resource is 
being upgraded"));
+break;
+case DEPLOYED:
+updateConditionIfNotExist(
+conditions,
+ConditionUtils.ready(
+"The resource is deployed, but it’s not yet 
considered to be stable and might be rolled back in the future"));
+break;
+case ROLLING_BACK:
+updateConditionIfNotExist(
+conditions,
+ConditionUtils.notReady(
+"The resource is being rolled back to the last 
stable spec"));
+break;
+case ROLLED_BACK:
+updateConditionIfNotExist(
+conditions,
+ConditionUtils.ready("The resource is deployed with 
the last stable spec"));
+break;
+case FAILED:
+updateConditionIfNotExist(conditions, 
ConditionUtils.error("failed"));
+break;
+case STABLE:
+updateConditionIfNotExist(
+conditions,
+ConditionUtils.ready(
+"The resource deployment is considered to be 
stable and won’t be rolled back"));
+break;
+}
+
+return conditions;
+}
+
+private void updateConditionIfNotExist(List conditions, 
Condition newCondition) {
+if (conditions.isEmpty()) {
+conditions.add(newCondition);
+}
+if (conditions.stream()
+.noneMatch(condition -> 
condition.getType().equals(newCondition.getType( {
+conditions.add(newCondition);
+} else if (conditions.removeIf(
+condition ->
+
!(condition.getReason().equals(newCondition.getReason())
+&& 
condition.getMessage().equals(newCondition.getMessage() {
+conditions.add(newCondition);
+}

Review Comment:
   Sure, will add the docs to code. above logic is to make sure that rather to 
blindly replace any existing conditions with new one , check for existing 
condition with same type and replace only if the same condition type has 
different message. 



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-34105) Akka timeout happens in TPC-DS benchmarks

2024-01-15 Thread Zhu Zhu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhu Zhu updated FLINK-34105:

Priority: Critical  (was: Major)

> Akka timeout happens in TPC-DS benchmarks
> -
>
> Key: FLINK-34105
> URL: https://issues.apache.org/jira/browse/FLINK-34105
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.19.0
>Reporter: Zhu Zhu
>Priority: Critical
> Attachments: image-2024-01-16-13-59-45-556.png
>
>
> We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The 
> problem did not happen in 1.18.0.
> After bisecting, we find the problem was introduced in FLINK-33532.
>  !image-2024-01-16-13-59-45-556.png|width=800! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-34105) Akka timeout happens in TPC-DS benchmarks

2024-01-15 Thread Zhu Zhu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34105?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807052#comment-17807052
 ] 

Zhu Zhu edited comment on FLINK-34105 at 1/16/24 6:06 AM:
--

[~lsdy] [~guoyangze] would you help to take a look?

This indicates a problem that existing large Flink batch jobs may become 
unstable in 1.19.


was (Author: zhuzh):
[~lsdy] [~guoyangze] would you help to take a look?

> Akka timeout happens in TPC-DS benchmarks
> -
>
> Key: FLINK-34105
> URL: https://issues.apache.org/jira/browse/FLINK-34105
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.19.0
>Reporter: Zhu Zhu
>Priority: Major
> Attachments: image-2024-01-16-13-59-45-556.png
>
>
> We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The 
> problem did not happen in 1.18.0.
> After bisecting, we find the problem was introduced in FLINK-33532.
>  !image-2024-01-16-13-59-45-556.png|width=800! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34105) Akka timeout happens in TPC-DS benchmarks

2024-01-15 Thread Zhu Zhu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhu Zhu updated FLINK-34105:

Description: 
We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The problem 
did not happen in 1.18.0.
After bisecting, we find the problem was introduced in FLINK-33532.

 !image-2024-01-16-13-59-45-556.png|width=250! 

  was:
We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The problem 
did not happen in 1.18.0.
After bisecting, we find the problem was introduced in FLINK-33532.

 !image-2024-01-16-13-59-45-556.png! 


> Akka timeout happens in TPC-DS benchmarks
> -
>
> Key: FLINK-34105
> URL: https://issues.apache.org/jira/browse/FLINK-34105
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.19.0
>Reporter: Zhu Zhu
>Priority: Major
> Attachments: image-2024-01-16-13-59-45-556.png
>
>
> We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The 
> problem did not happen in 1.18.0.
> After bisecting, we find the problem was introduced in FLINK-33532.
>  !image-2024-01-16-13-59-45-556.png|width=250! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34105) Akka timeout happens in TPC-DS benchmarks

2024-01-15 Thread Zhu Zhu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhu Zhu updated FLINK-34105:

Description: 
We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The problem 
did not happen in 1.18.0.
After bisecting, we find the problem was introduced in FLINK-33532.

 !image-2024-01-16-13-59-45-556.png|width=800! 

  was:
We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The problem 
did not happen in 1.18.0.
After bisecting, we find the problem was introduced in FLINK-33532.

 !image-2024-01-16-13-59-45-556.png|width=600! 


> Akka timeout happens in TPC-DS benchmarks
> -
>
> Key: FLINK-34105
> URL: https://issues.apache.org/jira/browse/FLINK-34105
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.19.0
>Reporter: Zhu Zhu
>Priority: Major
> Attachments: image-2024-01-16-13-59-45-556.png
>
>
> We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The 
> problem did not happen in 1.18.0.
> After bisecting, we find the problem was introduced in FLINK-33532.
>  !image-2024-01-16-13-59-45-556.png|width=800! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34105) Akka timeout happens in TPC-DS benchmarks

2024-01-15 Thread Zhu Zhu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhu Zhu updated FLINK-34105:

Description: 
We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The problem 
did not happen in 1.18.0.
After bisecting, we find the problem was introduced in FLINK-33532.

 !image-2024-01-16-13-59-45-556.png|width=600! 

  was:
We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The problem 
did not happen in 1.18.0.
After bisecting, we find the problem was introduced in FLINK-33532.

 !image-2024-01-16-13-59-45-556.png|width=250! 


> Akka timeout happens in TPC-DS benchmarks
> -
>
> Key: FLINK-34105
> URL: https://issues.apache.org/jira/browse/FLINK-34105
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.19.0
>Reporter: Zhu Zhu
>Priority: Major
> Attachments: image-2024-01-16-13-59-45-556.png
>
>
> We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The 
> problem did not happen in 1.18.0.
> After bisecting, we find the problem was introduced in FLINK-33532.
>  !image-2024-01-16-13-59-45-556.png|width=600! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] [FLINK-34104][autoscaler] Improve the ScalingReport format of autoscaling [flink-kubernetes-operator]

2024-01-15 Thread via GitHub


1996fanrui opened a new pull request, #757:
URL: https://github.com/apache/flink-kubernetes-operator/pull/757

   ## What is the purpose of the change
   
   Currently, the scaling report format is  `Vertex ID %s | Parallelism %d -> 
%d | Processing capacity %.2f -> %.2f | Target data rate %.2f`.
   
   It has 2 disadvantages:
   
   1. When one job has multiple vertices, the report of all vertices are mixed 
together without any separator, here is an example:
   - `Scaling execution enabled, begin scaling vertices: Vertex ID 
ea632d67b7d595e5b851708ae9ad79d6 | Parallelism 2 -> 1 | Processing capacity 
800466.67 -> 320186.67 | Target data rate 715.10 Vertex ID 
bc764cd8ddf7a0cff126f51c16239658 | Parallelism 2 -> 1 | Processing capacity 
79252.08 -> 31700.83 | Target data rate 895.93 Vertex ID 
0a448493b4782967b150582570326227 | Parallelism 8 -> 16 | Processing capacity 
716.05 -> 1141.00 | Target data rate 715.54`
   - We can see the Vertex ID is the beginning of each vertex report, it 
doesn't have any separator with the last vertex.
   2. This format is non-standard, it's hard to deserialize.
   - When job enables the autoscaler and disable the scaling.
   - Flink platform maintainer wants to show the scaling report in WebUI, 
it's helpful to using the report result for flink users.
   - So easy to deserialize is useful for these flink platform.
   
   ## Brief change log
   
   Serializing the scaling report with json format, it's easy to read and 
deserialize.
   
   Here is the scaling report message after this PR:  ``
   
   ## Verifying this change
   
 - Added `AutoScalerEventHandlerTest` to test the ScalingReport format.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency):  no
 - The public API, i.e., is any changes to the `CustomResourceDescriptors`: 
no
 - Core observer or reconciler logic that is regularly executed: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-34105) Akka timeout happens in TPC-DS benchmarks

2024-01-15 Thread Zhu Zhu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34105?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807052#comment-17807052
 ] 

Zhu Zhu commented on FLINK-34105:
-

[~lsdy] [~guoyangze] would you help to take a look?

> Akka timeout happens in TPC-DS benchmarks
> -
>
> Key: FLINK-34105
> URL: https://issues.apache.org/jira/browse/FLINK-34105
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.19.0
>Reporter: Zhu Zhu
>Priority: Major
> Attachments: image-2024-01-16-13-59-45-556.png
>
>
> We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The 
> problem did not happen in 1.18.0.
> After bisecting, we find the problem was introduced in FLINK-33532.
>  !image-2024-01-16-13-59-45-556.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34104) Improve the ScalingReport format of autoscaling

2024-01-15 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34104?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-34104:
---
Labels: pull-request-available  (was: )

> Improve the ScalingReport format of autoscaling
> ---
>
> Key: FLINK-34104
> URL: https://issues.apache.org/jira/browse/FLINK-34104
> Project: Flink
>  Issue Type: Improvement
>  Components: Autoscaler
>Reporter: Rui Fan
>Assignee: Rui Fan
>Priority: Major
>  Labels: pull-request-available
> Fix For: kubernetes-operator-1.8.0
>
>
> Currently, the scaling report format is 
> {color:#6a8759} Vertex ID %s | Parallelism %d -> %d | Processing capacity 
> %.2f -> %.2f | Target data rate %.2f{color}
> {color:#172b4d}It has 2 disadvantages:{color}
>  # {color:#172b4d}When one job has multiple vertices, the report of all 
> vertices are mixed together without any separator{color}{color:#172b4d}, here 
> is an example:{color}
>  ** {color:#172b4d}Scaling execution enabled, begin scaling vertices: Vertex 
> ID ea632d67b7d595e5b851708ae9ad79d6 | Parallelism 2 -> 1 | Processing 
> capacity 800466.67 -> 320186.67 | Target data rate 715.10 Vertex ID 
> bc764cd8ddf7a0cff126f51c16239658 | Parallelism 2 -> 1 | Processing capacity 
> 79252.08 -> 31700.83 | Target data rate 895.93 Vertex ID 
> 0a448493b4782967b150582570326227 | Parallelism 8 -> 16 | Processing capacity 
> 716.05 -> 1141.00 | Target data rate 715.54{color}
>  ** {color:#172b4d}We can see the Vertex ID is the beginning of each vertex 
> report, it doesn't have any {color}{color:#172b4d}separator with the last 
> vertex.
> {color}
>  # {color:#172b4d}This format is non-standard{color}{color:#172b4d}, it's 
> hard to deserialize.{color}
>  ** {color:#172b4d}When job enables the autoscaler and disable the 
> scaling.{color}
>  ** {color:#172b4d}Flink platform maintainer wants to show the scaling report 
> in WebUI, it's helpful to using the report result for flink users.{color}
>  ** {color:#172b4d}So easy to deserialize is useful for these flink 
> platform.{color}
> h2. {color:#172b4d}Solution:{color}
> {color:#172b4d}Serializing the scaling report with json format, it's easy to 
> read and deserialize.{color}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34105) Akka timeout happens in TPC-DS benchmarks

2024-01-15 Thread Zhu Zhu (Jira)
Zhu Zhu created FLINK-34105:
---

 Summary: Akka timeout happens in TPC-DS benchmarks
 Key: FLINK-34105
 URL: https://issues.apache.org/jira/browse/FLINK-34105
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Coordination
Affects Versions: 1.19.0
Reporter: Zhu Zhu
 Attachments: image-2024-01-16-13-59-45-556.png

We noticed akka timeout happens in 10TB TPC-DS benchmarks in 1.19. The problem 
did not happen in 1.18.0.
After bisecting, we find the problem was introduced in FLINK-33532.

 !image-2024-01-16-13-59-45-556.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33973] Add new interfaces for SinkV2 to synchronize the API with the SourceV2 API [flink]

2024-01-15 Thread via GitHub


pvary commented on PR #24080:
URL: https://github.com/apache/flink/pull/24080#issuecomment-1893114512

   @gyfora: Please merge. We have a green test run. Thx


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-34104) Improve the ScalingReport format of autoscaling

2024-01-15 Thread Rui Fan (Jira)
Rui Fan created FLINK-34104:
---

 Summary: Improve the ScalingReport format of autoscaling
 Key: FLINK-34104
 URL: https://issues.apache.org/jira/browse/FLINK-34104
 Project: Flink
  Issue Type: Improvement
  Components: Autoscaler
Reporter: Rui Fan
Assignee: Rui Fan
 Fix For: kubernetes-operator-1.8.0


Currently, the scaling report format is 

{color:#6a8759} Vertex ID %s | Parallelism %d -> %d | Processing capacity %.2f 
-> %.2f | Target data rate %.2f{color}

{color:#172b4d}It has 2 disadvantages:{color}
 # {color:#172b4d}When one job has multiple vertices, the report of all 
vertices are mixed together without any separator{color}{color:#172b4d}, here 
is an example:{color}
 ** {color:#172b4d}Scaling execution enabled, begin scaling vertices: Vertex ID 
ea632d67b7d595e5b851708ae9ad79d6 | Parallelism 2 -> 1 | Processing capacity 
800466.67 -> 320186.67 | Target data rate 715.10 Vertex ID 
bc764cd8ddf7a0cff126f51c16239658 | Parallelism 2 -> 1 | Processing capacity 
79252.08 -> 31700.83 | Target data rate 895.93 Vertex ID 
0a448493b4782967b150582570326227 | Parallelism 8 -> 16 | Processing capacity 
716.05 -> 1141.00 | Target data rate 715.54{color}
 ** {color:#172b4d}We can see the Vertex ID is the beginning of each vertex 
report, it doesn't have any {color}{color:#172b4d}separator with the last 
vertex.
{color}
 # {color:#172b4d}This format is non-standard{color}{color:#172b4d}, it's hard 
to deserialize.{color}
 ** {color:#172b4d}When job enables the autoscaler and disable the 
scaling.{color}
 ** {color:#172b4d}Flink platform maintainer wants to show the scaling report 
in WebUI, it's helpful to using the report result for flink users.{color}
 ** {color:#172b4d}So easy to deserialize is useful for these flink 
platform.{color}

h2. {color:#172b4d}Solution:{color}

{color:#172b4d}Serializing the scaling report with json format, it's easy to 
read and deserialize.{color}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34078][state]Move InternalKeyContext classes from o.a.f.runtim… [flink]

2024-01-15 Thread via GitHub


masteryhx closed pull request #24093: [FLINK-34078][state]Move 
InternalKeyContext classes from o.a.f.runtim…
URL: https://github.com/apache/flink/pull/24093


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-34103) AsyncIO example failed to run as DataGen Connector is not bundled

2024-01-15 Thread Zhanghao Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34103?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhanghao Chen updated FLINK-34103:
--
External issue ID:   (was: FLINK-32821)

> AsyncIO example failed to run as DataGen Connector is not bundled
> -
>
> Key: FLINK-34103
> URL: https://issues.apache.org/jira/browse/FLINK-34103
> Project: Flink
>  Issue Type: Bug
>  Components: Examples
>Affects Versions: 1.18.0
>Reporter: Zhanghao Chen
>Priority: Major
>
> From the comments of FLINK-32821:
> root@73186f600374:/opt/flink# bin/flink run 
> /volume/flink-examples-streaming-1.18.0-AsyncIO.jar
> WARNING: Unknown module: jdk.compiler specified to --add-exports
> WARNING: Unknown module: jdk.compiler specified to --add-exports
> WARNING: Unknown module: jdk.compiler specified to --add-exports
> WARNING: Unknown module: jdk.compiler specified to --add-exports
> WARNING: Unknown module: jdk.compiler specified to --add-exports
> java.lang.NoClassDefFoundError: 
> org/apache/flink/connector/datagen/source/DataGeneratorSource
> at 
> org.apache.flink.streaming.examples.async.AsyncIOExample.main(AsyncIOExample.java:82)
> at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown 
> Source)
> at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown 
> Source)
> at java.base/java.lang.reflect.Method.invoke(Unknown Source)
> at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355)
> at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222)
> at org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:105)
> at 
> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:851)
> at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:245)
> at org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1095)
> at 
> org.apache.flink.client.cli.CliFrontend.lambda$mainInternal$9(CliFrontend.java:1189)
> at 
> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
> at org.apache.flink.client.cli.CliFrontend.mainInternal(CliFrontend.java:1189)
> at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1157)
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.flink.connector.datagen.source.DataGeneratorSource
> at java.base/java.net.URLClassLoader.findClass(Unknown Source)
> at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
> at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:67)
> at 
> org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65)
> at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:51)
> at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
> ... 15 more



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34103) AsyncIO example failed to run as DataGen Connector is not bundled

2024-01-15 Thread Zhanghao Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34103?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhanghao Chen updated FLINK-34103:
--
External issue ID: FLINK-32821

> AsyncIO example failed to run as DataGen Connector is not bundled
> -
>
> Key: FLINK-34103
> URL: https://issues.apache.org/jira/browse/FLINK-34103
> Project: Flink
>  Issue Type: Bug
>  Components: Examples
>Affects Versions: 1.18.0
>Reporter: Zhanghao Chen
>Priority: Major
>
> From the comments of FLINK-32821:
> root@73186f600374:/opt/flink# bin/flink run 
> /volume/flink-examples-streaming-1.18.0-AsyncIO.jar
> WARNING: Unknown module: jdk.compiler specified to --add-exports
> WARNING: Unknown module: jdk.compiler specified to --add-exports
> WARNING: Unknown module: jdk.compiler specified to --add-exports
> WARNING: Unknown module: jdk.compiler specified to --add-exports
> WARNING: Unknown module: jdk.compiler specified to --add-exports
> java.lang.NoClassDefFoundError: 
> org/apache/flink/connector/datagen/source/DataGeneratorSource
> at 
> org.apache.flink.streaming.examples.async.AsyncIOExample.main(AsyncIOExample.java:82)
> at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown 
> Source)
> at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown 
> Source)
> at java.base/java.lang.reflect.Method.invoke(Unknown Source)
> at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355)
> at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222)
> at org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:105)
> at 
> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:851)
> at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:245)
> at org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1095)
> at 
> org.apache.flink.client.cli.CliFrontend.lambda$mainInternal$9(CliFrontend.java:1189)
> at 
> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
> at org.apache.flink.client.cli.CliFrontend.mainInternal(CliFrontend.java:1189)
> at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1157)
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.flink.connector.datagen.source.DataGeneratorSource
> at java.base/java.net.URLClassLoader.findClass(Unknown Source)
> at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
> at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:67)
> at 
> org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65)
> at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:51)
> at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
> ... 15 more



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-32821) Streaming examples failed to execute due to error in packaging

2024-01-15 Thread Zhanghao Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-32821?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807047#comment-17807047
 ] 

Zhanghao Chen commented on FLINK-32821:
---

FYI [~qinjunjerry] , I've created [FLINK-34103] AsyncIO example failed to run 
as DataGen Connector is not bundled - ASF JIRA (apache.org) to track this issue.

> Streaming examples failed to execute due to error in packaging
> --
>
> Key: FLINK-32821
> URL: https://issues.apache.org/jira/browse/FLINK-32821
> Project: Flink
>  Issue Type: Bug
>  Components: Examples
>Affects Versions: 1.18.0
>Reporter: Zhanghao Chen
>Assignee: Zhanghao Chen
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.18.0
>
>
> 6 out of the 8 streaming examples failed to run:
>  * Iteration & SessionWindowing & SocketWindowWordCount & WindowJoin failed 
> to run due to java.lang.NoClassDefFoundError: 
> org/apache/flink/streaming/examples/utils/ParameterTool
>  * MatrixVectorMul & TopSpeedWindowing failed to run due to: Caused by: 
> java.lang.ClassNotFoundException: 
> org.apache.flink.connector.datagen.source.GeneratorFunction
> The NoClassDefFoundError with ParameterTool is introduced by FLINK-32558 
> Properly deprecate DataSet API - ASF JIRA (apache.org), and we'd better 
> resolve FLINK-32820 ParameterTool is mistakenly marked as deprecated - ASF 
> JIRA (apache.org) first before we come to a fix for this problem.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34103) AsyncIO example failed to run as DataGen Connector is not bundled

2024-01-15 Thread Zhanghao Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34103?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhanghao Chen updated FLINK-34103:
--
Description: 
>From the comments of FLINK-32821:


root@73186f600374:/opt/flink# bin/flink run 
/volume/flink-examples-streaming-1.18.0-AsyncIO.jar
WARNING: Unknown module: jdk.compiler specified to --add-exports
WARNING: Unknown module: jdk.compiler specified to --add-exports
WARNING: Unknown module: jdk.compiler specified to --add-exports
WARNING: Unknown module: jdk.compiler specified to --add-exports
WARNING: Unknown module: jdk.compiler specified to --add-exports
java.lang.NoClassDefFoundError: 
org/apache/flink/connector/datagen/source/DataGeneratorSource
at 
org.apache.flink.streaming.examples.async.AsyncIOExample.main(AsyncIOExample.java:82)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native 
Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown 
Source)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown 
Source)
at java.base/java.lang.reflect.Method.invoke(Unknown Source)
at 
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355)
at 
org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222)
at org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:105)
at org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:851)
at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:245)
at org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1095)
at 
org.apache.flink.client.cli.CliFrontend.lambda$mainInternal$9(CliFrontend.java:1189)
at 
org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
at org.apache.flink.client.cli.CliFrontend.mainInternal(CliFrontend.java:1189)
at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1157)
Caused by: java.lang.ClassNotFoundException: 
org.apache.flink.connector.datagen.source.DataGeneratorSource
at java.base/java.net.URLClassLoader.findClass(Unknown Source)
at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
at 
org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:67)
at 
org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65)
at 
org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:51)
at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
... 15 more

  was:
>From the comments of 
>[FLINK-32821|https://issues.apache.org/jira/browse/FLINK-32821]:
root@73186f600374:/opt/flink# bin/flink run 
/volume/flink-examples-streaming-1.18.0-AsyncIO.jar
WARNING: Unknown module: jdk.compiler specified to --add-exports
WARNING: Unknown module: jdk.compiler specified to --add-exports
WARNING: Unknown module: jdk.compiler specified to --add-exports
WARNING: Unknown module: jdk.compiler specified to --add-exports
WARNING: Unknown module: jdk.compiler specified to --add-exports
java.lang.NoClassDefFoundError: 
org/apache/flink/connector/datagen/source/DataGeneratorSource
at 
org.apache.flink.streaming.examples.async.AsyncIOExample.main(AsyncIOExample.java:82)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at 
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown 
Source)
at java.base/java.lang.reflect.Method.invoke(Unknown Source)
at 
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355)
at 
org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222)
at 
org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:105)
at 
org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:851)
at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:245)
at 
org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1095)
at 
org.apache.flink.client.cli.CliFrontend.lambda$mainInternal$9(CliFrontend.java:1189)
at 
org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
at 
org.apache.flink.client.cli.CliFrontend.mainInternal(CliFrontend.java:1189)
at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1157)
Caused by: java.lang.ClassNotFoundException: 
org.apache.flink.connector.datagen.source.DataGeneratorSource
at java.base/java.net.URLClassLoader.findClass(Unknown Source)
at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
at 

[jira] [Created] (FLINK-34103) AsyncIO example failed to run as DataGen Connector is not bundled

2024-01-15 Thread Zhanghao Chen (Jira)
Zhanghao Chen created FLINK-34103:
-

 Summary: AsyncIO example failed to run as DataGen Connector is not 
bundled
 Key: FLINK-34103
 URL: https://issues.apache.org/jira/browse/FLINK-34103
 Project: Flink
  Issue Type: Bug
  Components: Examples
Affects Versions: 1.18.0
Reporter: Zhanghao Chen


>From the comments of 
>[FLINK-32821|https://issues.apache.org/jira/browse/FLINK-32821]:
root@73186f600374:/opt/flink# bin/flink run 
/volume/flink-examples-streaming-1.18.0-AsyncIO.jar
WARNING: Unknown module: jdk.compiler specified to --add-exports
WARNING: Unknown module: jdk.compiler specified to --add-exports
WARNING: Unknown module: jdk.compiler specified to --add-exports
WARNING: Unknown module: jdk.compiler specified to --add-exports
WARNING: Unknown module: jdk.compiler specified to --add-exports
java.lang.NoClassDefFoundError: 
org/apache/flink/connector/datagen/source/DataGeneratorSource
at 
org.apache.flink.streaming.examples.async.AsyncIOExample.main(AsyncIOExample.java:82)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at 
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown 
Source)
at java.base/java.lang.reflect.Method.invoke(Unknown Source)
at 
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355)
at 
org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222)
at 
org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:105)
at 
org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:851)
at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:245)
at 
org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1095)
at 
org.apache.flink.client.cli.CliFrontend.lambda$mainInternal$9(CliFrontend.java:1189)
at 
org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
at 
org.apache.flink.client.cli.CliFrontend.mainInternal(CliFrontend.java:1189)
at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1157)
Caused by: java.lang.ClassNotFoundException: 
org.apache.flink.connector.datagen.source.DataGeneratorSource
at java.base/java.net.URLClassLoader.findClass(Unknown Source)
at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
at 
org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:67)
at 
org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65)
at 
org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:51)
at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
... 15 more



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34052) Missing TopSpeedWindowing and SessionWindowing JARs in Flink Maven Repository

2024-01-15 Thread Zhanghao Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807045#comment-17807045
 ] 

Zhanghao Chen commented on FLINK-34052:
---

Thanks for reporting this, I'll take a look. [~huweihua] could you assign this 
to me? 

> Missing TopSpeedWindowing and SessionWindowing JARs in Flink Maven Repository
> -
>
> Key: FLINK-34052
> URL: https://issues.apache.org/jira/browse/FLINK-34052
> Project: Flink
>  Issue Type: Bug
>  Components: Build System, Examples
>Affects Versions: 1.18.0
>Reporter: Junrui Li
>Priority: Major
>
> As a result of the changes implemented in FLINK-32821, the build process no 
> longer produces artifacts with the names 
> flink-examples-streaming-1.x-TopSpeedWindowing.jar and 
> flink-examples-streaming-1.x-SessionWindowing.jar. This has led to the 
> absence of these specific JAR files in the Maven repository 
> (https://repo1.maven.org/maven2/org/apache/flink/flink-examples-streaming/1.18.0/).
> These artifacts were previously available and may still be expected by users 
> as part of their application dependencies. Their removal could potentially 
> break existing build pipelines and applications that depend on these example 
> JARs.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33221][core][config] Add config options for administrator JVM options [flink]

2024-01-15 Thread via GitHub


X-czh commented on PR #24098:
URL: https://github.com/apache/flink/pull/24098#issuecomment-1893096096

   Hi @huwh @1996fanrui, do you have time to help review this?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33803] Set observedGeneration at end of reconciliation [flink-kubernetes-operator]

2024-01-15 Thread via GitHub


justin-chen commented on code in PR #755:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/755#discussion_r1452942179


##
flink-kubernetes-operator-api/src/main/java/org/apache/flink/kubernetes/operator/api/status/FlinkDeploymentStatus.java:
##
@@ -55,4 +55,7 @@ public class FlinkDeploymentStatus extends 
CommonStatus {
 
 /** Information about the TaskManagers for the scale subresource. */
 private TaskManagerInfo taskManager;
+
+/** Last observed generation of the FlinkDeployment. */
+private Long observedGeneration;

Review Comment:
   Thanks for the guidance - I've added `observedGeneration` to `CommonStatus` 
class and set it alongside the `lastReconciledSpec`. I've left the `generation` 
within the `lastReconciledSpec` intact in case there are existing client 
dependencies on it. Let me know about any additional changes to be made



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-34102) Invalid configuration when using 'env.log.max' on yarn application mode

2024-01-15 Thread RocMarshal (Jira)
RocMarshal created FLINK-34102:
--

 Summary: Invalid configuration when using 'env.log.max' on yarn 
application mode
 Key: FLINK-34102
 URL: https://issues.apache.org/jira/browse/FLINK-34102
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Configuration
Affects Versions: 1.19.0
Reporter: RocMarshal






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33988][configuration] Fix the invalid configuration when using initialized root logger level on yarn deployment mode [flink]

2024-01-15 Thread via GitHub


1996fanrui commented on PR #24042:
URL: https://github.com/apache/flink/pull/24042#issuecomment-1893069359

   > I made a fix like the env-value passing-in at 
https://github.com/RocMarshal/flink/commit/624c199e756ed5f04f849e8e53033bc1ef258f93
   > Would we merge the additional fix into the current pr ?
   > Please let me know what's your opinion~ :)
   
   Same PR is fine for me, but separate JIRA and commits are necessary.
   
   Separate JIRA and commits are easy to follow up and cherry pick for other 
users.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33768] Support dynamic source parallelism inference for batch jobs [flink]

2024-01-15 Thread via GitHub


SinBex commented on PR #24087:
URL: https://github.com/apache/flink/pull/24087#issuecomment-1893062734

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-25054) Improve exception message for unsupported hashLength for SHA2 function

2024-01-15 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-25054?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-25054:
---
Labels: pull-request-available  (was: )

> Improve exception message for unsupported hashLength for SHA2 function
> --
>
> Key: FLINK-25054
> URL: https://issues.apache.org/jira/browse/FLINK-25054
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / API
>Affects Versions: 1.12.3
>Reporter: DingGeGe
>Assignee: Shuai Xu
>Priority: Major
>  Labels: pull-request-available
> Attachments: image-2021-11-25-16-59-56-699.png
>
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> 【exception sql】
> SELECT
> SHA2(, 128)
> FROM
>  
> 【effect】
> when sql is long , it`s hard to clear where is the problem on this issue
> 【reason】
> build-in function SHA2, hashLength do not support “128”, but I could not 
> understand from
> 【Exception log】
> !image-2021-11-25-16-59-56-699.png!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-25054][table-planner] Extend the exception message for SHA2 function [flink]

2024-01-15 Thread via GitHub


flinkbot commented on PR #24099:
URL: https://github.com/apache/flink/pull/24099#issuecomment-1893037933

   
   ## CI report:
   
   * 6525ba25342d12481b64e64fcc66ba6e57f0480f UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33988][configuration] Fix the invalid configuration when using initialized root logger level on yarn deployment mode [flink]

2024-01-15 Thread via GitHub


RocMarshal commented on PR #24042:
URL: https://github.com/apache/flink/pull/24042#issuecomment-1893035604

   Thanks @X-czh for the double-check and @1996fanrui review.
   
   I made a fix like the env-value passing-in  at 
https://github.com/RocMarshal/flink/commit/624c199e756ed5f04f849e8e53033bc1ef258f93
   
   Would we merge the additional fix into the current pr ?
   Please let me know what's your opinion~ :)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[PR] Extend the exception message for SHA2 function [flink]

2024-01-15 Thread via GitHub


xishuaidelin opened a new pull request, #24099:
URL: https://github.com/apache/flink/pull/24099

   
   
   ## What is the purpose of the change
   
   *This pull request extends the exception message for sha2 function.*
   
   
   ## Brief change log
   
   modify the function addReusableSha2MessageDigest.
   
   ## Verifying this change
   
   N/A
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): ( no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: ( no)
 - The serializers: (no )
 - The runtime per-record code paths (performance sensitive): (no )
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (no )
 - The S3 file system connector: (no )
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34087][tests][JUnit5 Migration] Migarate to junit5 of flink-dist module [flink]

2024-01-15 Thread via GitHub


Jiabao-Sun commented on PR #24092:
URL: https://github.com/apache/flink/pull/24092#issuecomment-1893033205

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-33856) Add metrics to monitor the interaction performance between task and external storage system in the process of checkpoint making

2024-01-15 Thread Jufang He (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33856?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807019#comment-17807019
 ] 

Jufang He edited comment on FLINK-33856 at 1/16/24 3:51 AM:


ok, I'll get to work on it. [~pnowojski] 


was (Author: JIRAUSER302059):
ok, I'll get to work on it.

> Add metrics to monitor the interaction performance between task and external 
> storage system in the process of checkpoint making
> ---
>
> Key: FLINK-33856
> URL: https://issues.apache.org/jira/browse/FLINK-33856
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Checkpointing
>Affects Versions: 1.18.0
>Reporter: Jufang He
>Assignee: Jufang He
>Priority: Major
>  Labels: pull-request-available
>
> When Flink makes a checkpoint, the interaction performance with the external 
> file system has a great impact on the overall time-consuming. Therefore, it 
> is easy to observe the bottleneck point by adding performance indicators when 
> the task interacts with the external file storage system. These include: the 
> rate of file write , the latency to write the file, the latency to close the 
> file.
> In flink side add the above metrics has the following advantages: convenient 
> statistical different task E2E time-consuming; do not need to distinguish the 
> type of external storage system, can be unified in the 
> FsCheckpointStreamFactory.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33856) Add metrics to monitor the interaction performance between task and external storage system in the process of checkpoint making

2024-01-15 Thread Jufang He (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33856?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807019#comment-17807019
 ] 

Jufang He commented on FLINK-33856:
---

ok, I'll get to work on it.

> Add metrics to monitor the interaction performance between task and external 
> storage system in the process of checkpoint making
> ---
>
> Key: FLINK-33856
> URL: https://issues.apache.org/jira/browse/FLINK-33856
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Checkpointing
>Affects Versions: 1.18.0
>Reporter: Jufang He
>Assignee: Jufang He
>Priority: Major
>  Labels: pull-request-available
>
> When Flink makes a checkpoint, the interaction performance with the external 
> file system has a great impact on the overall time-consuming. Therefore, it 
> is easy to observe the bottleneck point by adding performance indicators when 
> the task interacts with the external file storage system. These include: the 
> rate of file write , the latency to write the file, the latency to close the 
> file.
> In flink side add the above metrics has the following advantages: convenient 
> statistical different task E2E time-consuming; do not need to distinguish the 
> type of external storage system, can be unified in the 
> FsCheckpointStreamFactory.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-34007) Flink Job stuck in suspend state after losing leadership in HA Mode

2024-01-15 Thread Zhenqiu Huang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34007?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807017#comment-17807017
 ] 

Zhenqiu Huang edited comment on FLINK-34007 at 1/16/24 3:40 AM:


[~mapohl] [~wangyang0918]
I am intensively testing flink 1.18. Within two days, there are users reported 
the job manager stuck issue in 1.17 and 1.16. 1.18 and 1.17 job instances are 
running in the same cluster. 1.16 is in different cluster.

I attached another LeaderElector-Debug.json file that contains debug log of a 
flink 1.18 app. The issue happened several times:
1. due to the configmap not accessible from api sever then renew timeout 
exceeded. 
2. a failure on patch on a updated configmap


The interesting part of the behavior of last several days is that job manager 
was not stuck but exit directly. Then, new job manager pod started correctly 
that is why new leader is selected in the log above. Hopefully, it is useful 
for your diagnosis.


[~wangyang0918]
>From my initial observation (before creating the jira), the leader annotation 
>update stopped when job manager was stuck. 







was (Author: zhenqiuhuang):
[~mapohl] [~wangyang0918]
I am intensively testing flink 1.18. Within two days, there are users reported 
the job manager stuck issue in 1.17 and 1.16. 1.18 and 1.17 job instances are 
running in the same cluster. 1.16 is in different cluster.

I attached another LeaderElector-Debug.json file that contains debug log of a 
flink 1.18 app. The issue happened several times:
1. due to the configmap not accessible from api sever then renew timeout 
exceeded 
2. a failure on patch on a updated configmap

The interesting part of the behavior of last several days is that job manager 
was not stuck but exit directly. Then, new job manager pod started correctly 
that is why new leader is selected in the log above. Hopefully, it is useful 
for your diagnosis.






> Flink Job stuck in suspend state after losing leadership in HA Mode
> ---
>
> Key: FLINK-34007
> URL: https://issues.apache.org/jira/browse/FLINK-34007
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.16.3, 1.17.2, 1.18.1, 1.18.2
>Reporter: Zhenqiu Huang
>Priority: Major
> Attachments: Debug.log, LeaderElector-Debug.json, job-manager.log
>
>
> The observation is that Job manager goes to suspend state with a failed 
> container not able to register itself to resource manager after timeout.
> JM Log, see attached
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34101) use flink ui upload jar error, message is 'Exactly 1 file must be sent, received 0'

2024-01-15 Thread blackpighe (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34101?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807018#comment-17807018
 ] 

blackpighe commented on FLINK-34101:


 

!image-2024-01-16-11-37-26-343.png!

this is copy curl request by web console:

 

curl "http://localhost:8083/proxy/ingress/ui/1/224/jars/upload; ^
  -H "Accept: application/json, text/plain, */*" ^
  -H "Accept-Language: zh-CN,zh;q=0.9" ^
  -H "Cache-Control: no-cache" ^
  -H "Connection: keep-alive" ^
  -H "Content-Type: multipart/form-data; 
boundary=WebKitFormBoundaryl6QGLTdn1N5xv5zc" ^
  -H "Origin: http://localhost:8083; ^
  -H "Pragma: no-cache" ^
  -H "Referer: http://localhost:8083/proxy/ingress/ui/1/224/; ^
  -H "Sec-Fetch-Dest: empty" ^
  -H "Sec-Fetch-Mode: cors" ^
  -H "Sec-Fetch-Site: same-origin" ^
  -H "User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 
(KHTML, like Gecko) Chrome/120.0.0.0 Safari/537.36" ^
  -H "sec-ch-ua: ^\^"Not_A Brand^\^";v=^\^"8^\^", 
^\^"Chromium^\^";v=^\^"120^\^", ^\^"Google Chrome^\^";v=^\^"120^\^"" ^
  -H "sec-ch-ua-mobile: ?0" ^
  -H "sec-ch-ua-platform: ^\^"Windows^\^"" ^
  --data-raw ^"--WebKitFormBoundaryl6QGLTdn1N5xv5zc^

Content-Disposition: form-data; name=^\^"jarfile^\^"; 
filename=^\^"faas-flink-1.14.3_2.12-20481024-SNAPSHOT-session.jar^\^"^

Content-Type: application/octet-stream^

^

^

--WebKitFormBoundaryl6QGLTdn1N5xv5zc--^

^" ^
  --compressed

> use flink ui upload jar error, message is 'Exactly 1 file must be sent, 
> received 0'
> ---
>
> Key: FLINK-34101
> URL: https://issues.apache.org/jira/browse/FLINK-34101
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / REST, Runtime / Web Frontend
>Affects Versions: 1.17.2
>Reporter: blackpighe
>Priority: Major
> Fix For: 1.17.2
>
> Attachments: image-2024-01-16-11-32-50-015.png, 
> image-2024-01-16-11-35-21-582.png, image-2024-01-16-11-37-26-343.png
>
>
> use flink ui upload jar, Something went wrong.
>  
> !image-2024-01-16-11-32-50-015.png!
> message is :
> Server Response:org.apache.flink.runtime.rest.handler.RestHandlerException: 
> Exactly 1 file must be sent, received 0. at 
> org.apache.flink.runtime.webmonitor.handlers.JarUploadHandler.handleRequest(JarUploadHandler.java:80)
>  at 
> org.apache.flink.runtime.rest.handler.AbstractRestHandler.respondToRequest(AbstractRestHandler.java:83)
>  at 
> org.apache.flink.runtime.rest.handler.AbstractHandler.respondAsLeader(AbstractHandler.java:196)
>  at 
> org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.lambda$channelRead0$0(LeaderRetrievalHandler.java:83)
>  at java.util.Optional.ifPresent(Optional.java:159) at 
> org.apache.flink.util.OptionalConsumer.ifPresent(OptionalConsumer.java:45) at 
> org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.channelRead0(LeaderRetrievalHandler.java:80)
>  at 
> org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.channelRead0(LeaderRetrievalHandler.java:49)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
>  at 
> org.apache.flink.runtime.rest.handler.router.RouterHandler.routed(RouterHandler.java:115)
>  at 
> org.apache.flink.runtime.rest.handler.router.RouterHandler.channelRead0(RouterHandler.java:94)
>  at 
> org.apache.flink.runtime.rest.handler.router.RouterHandler.channelRead0(RouterHandler.java:55)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
>  at 
> org.apache.flink.shaded.netty4.io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>  at 
> 

[jira] [Updated] (FLINK-34101) use flink ui upload jar error, message is 'Exactly 1 file must be sent, received 0'

2024-01-15 Thread blackpighe (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34101?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

blackpighe updated FLINK-34101:
---
Attachment: image-2024-01-16-11-37-26-343.png

> use flink ui upload jar error, message is 'Exactly 1 file must be sent, 
> received 0'
> ---
>
> Key: FLINK-34101
> URL: https://issues.apache.org/jira/browse/FLINK-34101
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / REST, Runtime / Web Frontend
>Affects Versions: 1.17.2
>Reporter: blackpighe
>Priority: Major
> Fix For: 1.17.2
>
> Attachments: image-2024-01-16-11-32-50-015.png, 
> image-2024-01-16-11-35-21-582.png, image-2024-01-16-11-37-26-343.png
>
>
> use flink ui upload jar, Something went wrong.
>  
> !image-2024-01-16-11-32-50-015.png!
> message is :
> Server Response:org.apache.flink.runtime.rest.handler.RestHandlerException: 
> Exactly 1 file must be sent, received 0. at 
> org.apache.flink.runtime.webmonitor.handlers.JarUploadHandler.handleRequest(JarUploadHandler.java:80)
>  at 
> org.apache.flink.runtime.rest.handler.AbstractRestHandler.respondToRequest(AbstractRestHandler.java:83)
>  at 
> org.apache.flink.runtime.rest.handler.AbstractHandler.respondAsLeader(AbstractHandler.java:196)
>  at 
> org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.lambda$channelRead0$0(LeaderRetrievalHandler.java:83)
>  at java.util.Optional.ifPresent(Optional.java:159) at 
> org.apache.flink.util.OptionalConsumer.ifPresent(OptionalConsumer.java:45) at 
> org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.channelRead0(LeaderRetrievalHandler.java:80)
>  at 
> org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.channelRead0(LeaderRetrievalHandler.java:49)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
>  at 
> org.apache.flink.runtime.rest.handler.router.RouterHandler.routed(RouterHandler.java:115)
>  at 
> org.apache.flink.runtime.rest.handler.router.RouterHandler.channelRead0(RouterHandler.java:94)
>  at 
> org.apache.flink.runtime.rest.handler.router.RouterHandler.channelRead0(RouterHandler.java:55)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
>  at 
> org.apache.flink.shaded.netty4.io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
>  at 
> org.apache.flink.runtime.rest.FileUploadHandler.channelRead0(FileUploadHandler.java:203)
>  at 
> org.apache.flink.runtime.rest.FileUploadHandler.channelRead0(FileUploadHandler.java:69)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler$DelegatingChannelHandlerContext.fireChannelRead(CombinedChannelDuplexHandler.java:436)
>  at 
> 

[jira] [Created] (FLINK-34101) use flink ui upload jar error, message is 'Exactly 1 file must be sent, received 0'

2024-01-15 Thread blackpighe (Jira)
blackpighe created FLINK-34101:
--

 Summary: use flink ui upload jar error, message is 'Exactly 1 file 
must be sent, received 0'
 Key: FLINK-34101
 URL: https://issues.apache.org/jira/browse/FLINK-34101
 Project: Flink
  Issue Type: Bug
  Components: Runtime / REST, Runtime / Web Frontend
Affects Versions: 1.17.2
Reporter: blackpighe
 Fix For: 1.17.2
 Attachments: image-2024-01-16-11-32-50-015.png, 
image-2024-01-16-11-35-21-582.png

use flink ui upload jar, Something went wrong.

 

!image-2024-01-16-11-32-50-015.png!

message is :

Server Response:org.apache.flink.runtime.rest.handler.RestHandlerException: 
Exactly 1 file must be sent, received 0. at 
org.apache.flink.runtime.webmonitor.handlers.JarUploadHandler.handleRequest(JarUploadHandler.java:80)
 at 
org.apache.flink.runtime.rest.handler.AbstractRestHandler.respondToRequest(AbstractRestHandler.java:83)
 at 
org.apache.flink.runtime.rest.handler.AbstractHandler.respondAsLeader(AbstractHandler.java:196)
 at 
org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.lambda$channelRead0$0(LeaderRetrievalHandler.java:83)
 at java.util.Optional.ifPresent(Optional.java:159) at 
org.apache.flink.util.OptionalConsumer.ifPresent(OptionalConsumer.java:45) at 
org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.channelRead0(LeaderRetrievalHandler.java:80)
 at 
org.apache.flink.runtime.rest.handler.LeaderRetrievalHandler.channelRead0(LeaderRetrievalHandler.java:49)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
 at 
org.apache.flink.runtime.rest.handler.router.RouterHandler.routed(RouterHandler.java:115)
 at 
org.apache.flink.runtime.rest.handler.router.RouterHandler.channelRead0(RouterHandler.java:94)
 at 
org.apache.flink.runtime.rest.handler.router.RouterHandler.channelRead0(RouterHandler.java:55)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
 at 
org.apache.flink.shaded.netty4.io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
 at 
org.apache.flink.runtime.rest.FileUploadHandler.channelRead0(FileUploadHandler.java:203)
 at 
org.apache.flink.runtime.rest.FileUploadHandler.channelRead0(FileUploadHandler.java:69)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler$DelegatingChannelHandlerContext.fireChannelRead(CombinedChannelDuplexHandler.java:436)
 at 
org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336)
 at 
org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler.channelRead(CombinedChannelDuplexHandler.java:251)
 at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
 at 

[jira] [Comment Edited] (FLINK-34007) Flink Job stuck in suspend state after losing leadership in HA Mode

2024-01-15 Thread Zhenqiu Huang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34007?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807017#comment-17807017
 ] 

Zhenqiu Huang edited comment on FLINK-34007 at 1/16/24 3:34 AM:


[~mapohl] [~wangyang0918]
I am intensively testing flink 1.18. Within two days, there are users reported 
the job manager stuck issue in 1.17 and 1.16. 1.18 and 1.17 job instances are 
running in the same cluster. 1.16 is in different cluster.

I attached another LeaderElector-Debug.json file that contains debug log of a 
flink 1.18 app. The issue happened several times:
1. due to the configmap not accessible from api sever then renew timeout 
exceeded 
2. a failure on patch on a updated configmap

The interesting part of the behavior of last several days is that job manager 
was not stuck but exit directly. Then, new job manager pod started correctly 
that is why new leader is selected in the log above. Hopefully, it is useful 
for your diagnosis.







was (Author: zhenqiuhuang):
[~mapohl]
I am intensively testing flink 1.18. Within two days, there are users reported 
the job manager stuck issue in 1.17 and 1.16. 1.18 and 1.17 job instances are 
running in the same cluster. 1.16 is in different cluster.

I attached another LeaderElector-Debug.json file that contains debug log of a 
flink 1.18 app. The issue happened several times:
1. due to the configmap not accessible from api sever then renew timeout 
exceeded 
2. a failure on patch on a updated configmap

The interesting part of the behavior of last several days is that job manager 
was not stuck but exit directly. Then, new job manager pod started correctly 
that is why new leader is selected in the log above. Hopefully, it is useful 
for your diagnosis.






> Flink Job stuck in suspend state after losing leadership in HA Mode
> ---
>
> Key: FLINK-34007
> URL: https://issues.apache.org/jira/browse/FLINK-34007
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.16.3, 1.17.2, 1.18.1, 1.18.2
>Reporter: Zhenqiu Huang
>Priority: Major
> Attachments: Debug.log, LeaderElector-Debug.json, job-manager.log
>
>
> The observation is that Job manager goes to suspend state with a failed 
> container not able to register itself to resource manager after timeout.
> JM Log, see attached
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34007) Flink Job stuck in suspend state after losing leadership in HA Mode

2024-01-15 Thread Zhenqiu Huang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34007?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807017#comment-17807017
 ] 

Zhenqiu Huang commented on FLINK-34007:
---

[~mapohl]
I am intensively testing flink 1.18. Within two days, there are users reported 
the job manager stuck issue in 1.17 and 1.16. 1.18 and 1.17 job instances are 
running in the same cluster. 1.16 is in different cluster.

I attached another LeaderElector-Debug.json file that contains debug log of a 
flink 1.18 app. The issue happened several times:
1. due to the configmap not accessible from api sever then renew timeout 
exceeded 
2. a failure on patch on a updated configmap

The interesting part of the behavior of last several days is that job manager 
was not stuck but exit directly. Then, new job manager pod started correctly 
that is why new leader is selected in the log above. Hopefully, it is useful 
for your diagnosis.






> Flink Job stuck in suspend state after losing leadership in HA Mode
> ---
>
> Key: FLINK-34007
> URL: https://issues.apache.org/jira/browse/FLINK-34007
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.16.3, 1.17.2, 1.18.1, 1.18.2
>Reporter: Zhenqiu Huang
>Priority: Major
> Attachments: Debug.log, LeaderElector-Debug.json, job-manager.log
>
>
> The observation is that Job manager goes to suspend state with a failed 
> container not able to register itself to resource manager after timeout.
> JM Log, see attached
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-21949][table] Support ARRAY_AGG aggregate function [flink]

2024-01-15 Thread via GitHub


Jiabao-Sun commented on code in PR #23411:
URL: https://github.com/apache/flink/pull/23411#discussion_r1452891261


##
flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/functions/bridging/BridgingSqlAggFunction.java:
##
@@ -63,6 +64,7 @@ public final class BridgingSqlAggFunction extends 
SqlAggFunction {
 private final List paramTypes;
 
 private BridgingSqlAggFunction(
+FlinkContext context,

Review Comment:
   Yes, I'll revert that changes.
   Thanks.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-34007) Flink Job stuck in suspend state after losing leadership in HA Mode

2024-01-15 Thread Zhenqiu Huang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34007?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhenqiu Huang updated FLINK-34007:
--
Attachment: LeaderElector-Debug.json

> Flink Job stuck in suspend state after losing leadership in HA Mode
> ---
>
> Key: FLINK-34007
> URL: https://issues.apache.org/jira/browse/FLINK-34007
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.16.3, 1.17.2, 1.18.1, 1.18.2
>Reporter: Zhenqiu Huang
>Priority: Major
> Attachments: Debug.log, LeaderElector-Debug.json, job-manager.log
>
>
> The observation is that Job manager goes to suspend state with a failed 
> container not able to register itself to resource manager after timeout.
> JM Log, see attached
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34100) Support session window table function without pulling up with window agg

2024-01-15 Thread xuyang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34100?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

xuyang updated FLINK-34100:
---
Description: This subtask resolves the session support in 
ExecWindowTableFunction. And then the session window can support window agg 
with WindowAttachedWindowingStrategy.  (was: This subtask resolves the session 
window agg with WindowAttachedWindowingStrategy.)

> Support session window table function without pulling up with window agg 
> -
>
> Key: FLINK-34100
> URL: https://issues.apache.org/jira/browse/FLINK-34100
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Runtime
>Reporter: xuyang
>Priority: Major
>
> This subtask resolves the session support in ExecWindowTableFunction. And 
> then the session window can support window agg with 
> WindowAttachedWindowingStrategy.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34048) Support session window agg in table runtime instead of using legacy group window agg op

2024-01-15 Thread xuyang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34048?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

xuyang updated FLINK-34048:
---
Description: 
We not only need FLINK-24024  to support session window agg in planner, but 
also need to support it in runtime.

This subtask only resolves the session window agg with 
TimeAttributeWindowingStrategy.

  was:We not only need FLINK-24024  to support session window agg in planner, 
but also need to support it in runtime.


> Support session window agg in table runtime instead of using legacy group 
> window agg op
> ---
>
> Key: FLINK-34048
> URL: https://issues.apache.org/jira/browse/FLINK-34048
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Runtime
>Affects Versions: 1.19.0
>Reporter: xuyang
>Priority: Major
>
> We not only need FLINK-24024  to support session window agg in planner, but 
> also need to support it in runtime.
> This subtask only resolves the session window agg with 
> TimeAttributeWindowingStrategy.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34100) Support session window table function without pulling up with window agg

2024-01-15 Thread xuyang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34100?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

xuyang updated FLINK-34100:
---
Description: This subtask resolves the session window agg with 
WindowAttachedWindowingStrategy.

> Support session window table function without pulling up with window agg 
> -
>
> Key: FLINK-34100
> URL: https://issues.apache.org/jira/browse/FLINK-34100
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Runtime
>Reporter: xuyang
>Priority: Major
>
> This subtask resolves the session window agg with 
> WindowAttachedWindowingStrategy.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34070][test] Adds dedicated tests in MiniClusterITCase for scenarios where there are not enough slots available. [flink]

2024-01-15 Thread via GitHub


jiangxin369 commented on code in PR #24082:
URL: https://github.com/apache/flink/pull/24082#discussion_r1452878415


##
tools/azure-pipelines/build-apache-repo.yml:
##
@@ -73,6 +73,17 @@ stages:
   run_end_to_end: false
   container: flink-build-container
   jdk: 8
+  - template: jobs-template.yml
+parameters:
+  stage_name: ci-adaptive-scheduler
+  test_pool_definition:
+name: Default
+  e2e_pool_definition:
+vmImage: 'ubuntu-20.04'
+  environment: PROFILE="-Penable-adaptive-scheduler"

Review Comment:
   Shall we add the `-Dflink.hadoop.version=2.10.2` in the profile to keep 
consistency with other stages? We can also remove the profile from other stages 
If `2.10.2` is the default value.



##
flink-runtime/src/test/java/org/apache/flink/runtime/minicluster/MiniClusterITCase.java:
##
@@ -149,7 +174,15 @@ private void runHandleJobsWhenNotEnoughSlots(final 
JobGraph jobGraph) throws Exc
 try (final MiniCluster miniCluster = new MiniCluster(cfg)) {
 miniCluster.start();
 
-miniCluster.executeJobBlocking(jobGraph);
+assertThatThrownBy(() -> miniCluster.executeJobBlocking(jobGraph))
+.isInstanceOf(JobExecutionException.class)
+.hasMessageContaining("Job execution failed")
+.extracting(Throwable::getCause)
+.extracting(FlinkAssertions::chainOfCauses, 
FlinkAssertions.STREAM_THROWABLE)
+.anySatisfy(
+cause ->
+assertThat(cause)
+
.isInstanceOf(NoResourceAvailableException.class));

Review Comment:
   Would it be more explicit if we assert the TimeoutException is thrown when 
running `testHandlingNotEnoughSlotsThroughTimeout` with the default scheduler?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-34100) Support session window table function without pulling up with window agg

2024-01-15 Thread xuyang (Jira)
xuyang created FLINK-34100:
--

 Summary: Support session window table function without pulling up 
with window agg 
 Key: FLINK-34100
 URL: https://issues.apache.org/jira/browse/FLINK-34100
 Project: Flink
  Issue Type: Sub-task
  Components: Table SQL / Runtime
Reporter: xuyang






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33988][configuration] Fix the invalid configuration when using initialized root logger level on yarn deployment mode [flink]

2024-01-15 Thread via GitHub


X-czh commented on PR #24042:
URL: https://github.com/apache/flink/pull/24042#issuecomment-1892997269

   Hi @1996fanrui, I've verified that the current release of Flink supports 
modifying the logger level in both Native K8s and Standalone mode. Since YARN 
prepares the container launching context in a special way, we'll need to do 
some special adaptation anyway. LGTM


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33988) Invalid configuration when using initialized root logger level on yarn application mode

2024-01-15 Thread Zhanghao Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33988?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807014#comment-17807014
 ] 

Zhanghao Chen commented on FLINK-33988:
---

[~RocMarshal] Thanks for reporting this. You are right, the impl of FLINK-33166 
won't work under YARN Application mode. The impl of FLINK-33166 replies on the 
preparing the starting context in pure shell code (Standalone & Native K8s 
follows this track). However, YARN application mode prepares the starting 
context in Java code 
(flink-yarn/YarnClusterDescriptor#setupApplicationMasterContainer & 
flink-yarn/Utils#createTaskExecutorContext, the logging-related config is 
assembled in YarnLogConfigUtil). It seems thatthe option {{env.log.max}} does 
not take effect under YARN as well.

> Invalid configuration when using initialized root logger level on yarn 
> application mode
> ---
>
> Key: FLINK-33988
> URL: https://issues.apache.org/jira/browse/FLINK-33988
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Configuration
>Affects Versions: 1.19.0
>Reporter: RocMarshal
>Assignee: RocMarshal
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0
>
>
> relevant https://issues.apache.org/jira/browse/FLINK-33166
> When I set env. log. level=DEBUG and start the flink job by yarn application 
> mode, the logs of TM and JM are still INFO.
> Preliminary inference is that it is ROOT_ LOG_ LEVEL value transmission link 
> is not complete enough.
> So I used the following configuration:
> containerized. taskmanager. env. ROOT_ LOG_ LEVEL=DEBUG
> containerized. master. env. ROOT_ LOG_ LEVEL=DEBUG
>  
> When starting the job by yarn application mode, TM and JM can output DEBUG 
> level logs.
>  
> Repair ideas:
> Fill the value of *env. log. level* into the Flink configuration by 
> *containerized. xxx. env. ROOT_ LOG_ LEVEL* before obtaining the environment 
> variable for the container



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33929][jdbc-connector] Support JDBC String field read Fragment read [flink-connector-jdbc]

2024-01-15 Thread via GitHub


zhilinli123 commented on PR #87:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/87#issuecomment-1892982217

   > > @zhilinli123 I still don't understand the Jira ticket, so I can't really 
review it. Perhaps @snuyanzin or @eskabetxe understand the goal of this PR.
   > 
   > For what I can assume, @zhilinli123 are trying to use scan.partition on a 
table that is hash partitioned by a string field..
   > 
   > My understanding of scan.partition is that only divides the query on 
multiples query based on params of scan.partition but is not using partitioning 
of table, it only adds "field BETWEEN ? AND ?" in query, and for that it only 
allow numeric, date, or timestamp column..
   
   You're right, but slicing string hash md5 doesn't need to be done BETWEEN? 
AND ?  It's much more efficient to check this shard directly
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Assigned] (FLINK-29802) ChangelogStateBackend supports native savepoint

2024-01-15 Thread Hangxiang Yu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29802?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hangxiang Yu reassigned FLINK-29802:


Assignee: Hangxiang Yu

> ChangelogStateBackend supports native savepoint
> ---
>
> Key: FLINK-29802
> URL: https://issues.apache.org/jira/browse/FLINK-29802
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Reporter: Hangxiang Yu
>Assignee: Hangxiang Yu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-34078) Move InternalKeyContext classes from o.a.f.runtime.state.heap to o.a.f.runtime.state package

2024-01-15 Thread Hangxiang Yu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34078?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hangxiang Yu resolved FLINK-34078.
--
Fix Version/s: 1.19.0
   Resolution: Fixed

merged 6bdb4f75 into master

> Move InternalKeyContext classes from o.a.f.runtime.state.heap to 
> o.a.f.runtime.state package
> 
>
> Key: FLINK-34078
> URL: https://issues.apache.org/jira/browse/FLINK-34078
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Reporter: Jinzhong Li
>Assignee: Jinzhong Li
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.19.0
>
> Attachments: image-2024-01-15-12-57-12-667.png
>
>
> h3. Motication:
> When Rocksdb statebackend throws a keyGroup check illegal exception, 
> the exception stack contains the heap stateBackend scoped class, which looks 
> so strange to user.
> !image-2024-01-15-12-57-12-667.png|width=555,height=68!
> h3. Proposed changes:
> InternalKeyContext and InternalKeyContextImpl are commonly used by all state 
> backends (heap/rocksdb/changelog), they should be moved from 
> org.apache.flink.runtime.state.heap package to org.apache.flink.runtime.state 
> package.
> h3. Compatibility:
> InternalKeyContext is annotated with @Internal, so this change has no 
> compatibility issues.
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33689][table-runtime] Fix JsonObjectAggFunction can't retract records when enabling LocalGlobal. [flink]

2024-01-15 Thread via GitHub


xishuaidelin commented on code in PR #23827:
URL: https://github.com/apache/flink/pull/23827#discussion_r1452864030


##
flink-table/flink-table-planner/src/test/scala/org/apache/flink/table/planner/runtime/stream/sql/AggregateITCase.scala:
##
@@ -1916,4 +1916,44 @@ class AggregateITCase(aggMode: AggMode, miniBatch: 
MiniBatchMode, backend: State
   )
 assertThat(sink.getRetractResults.sorted).isEqualTo(expected.sorted)
   }
+
+  @TestTemplate
+  def testGroupJsonObjectAggWithRetract(): Unit = {
+val data = new mutable.MutableList[(Long, String, Long)]
+data.+=((2L, "Hallo", 2L))

Review Comment:
   Hi lincoln, Thanks for your response. I have added the tests with nested agg 
and simplify the code.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33971]Specifies whether to use HBase table that supports dynamic columns. [flink-connector-hbase]

2024-01-15 Thread via GitHub


MOBIN-F commented on code in PR #36:
URL: 
https://github.com/apache/flink-connector-hbase/pull/36#discussion_r1452863933


##
flink-connector-hbase-base/src/main/java/org/apache/flink/connector/hbase/sink/RowDataToMutationConverter.java:
##
@@ -43,23 +43,26 @@ public class RowDataToMutationConverter implements 
HBaseMutationConverter DYNAMIC_TABLE =
+ConfigOptions.key("dynamic.table")

Review Comment:
   yep



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33689][table-runtime] Fix JsonObjectAggFunction can't retract records when enabling LocalGlobal. [flink]

2024-01-15 Thread via GitHub


xishuaidelin commented on code in PR #23827:
URL: https://github.com/apache/flink/pull/23827#discussion_r1452862528


##
flink-table/flink-table-runtime/src/main/java/org/apache/flink/table/runtime/functions/aggregate/JsonObjectAggFunction.java:
##
@@ -128,6 +137,15 @@ public void merge(Accumulator acc, Iterable 
others) throws Exceptio
 assertKeyNotPresent(acc, key);
 acc.map.put(key, other.map.get(key));
 }
+for (final StringData key : other.retractMap.keys()) {

Review Comment:
   Hi xuyang, Thanks for your comments. if +U is eliminated by -U, It means two 
messages has same key and value. It is an expected behaviour. However, you 
reminded me that the implementation here needs to be changed. Retraction is 
supposed to be done first in merge function.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33971]Specifies whether to use HBase table that supports dynamic columns. [flink-connector-hbase]

2024-01-15 Thread via GitHub


MOBIN-F commented on code in PR #36:
URL: 
https://github.com/apache/flink-connector-hbase/pull/36#discussion_r1452862759


##
flink-connector-hbase-base/src/main/java/org/apache/flink/connector/hbase/util/HBaseSerde.java:
##


Review Comment:
   I agree with this point of view. If you have no objection to the 
dynamic-table parameter, I am willing to do this part of the refactoring work.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-34099) CheckpointIntervalDuringBacklogITCase.testNoCheckpointDuringBacklog is unstable on AZP

2024-01-15 Thread Yu Chen (Jira)
Yu Chen created FLINK-34099:
---

 Summary: 
CheckpointIntervalDuringBacklogITCase.testNoCheckpointDuringBacklog is unstable 
on AZP
 Key: FLINK-34099
 URL: https://issues.apache.org/jira/browse/FLINK-34099
 Project: Flink
  Issue Type: Bug
Affects Versions: 1.19.0
Reporter: Yu Chen


This build [Pipelines - Run 20240115.30 logs 
(azure.com)|https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=56403=logs=5c8e7682-d68f-54d1-16a2-a09310218a49=86f654fa-ab48-5c1a-25f4-7e7f6afb9bba]
 fails as 
{code:java}
Jan 15 18:29:51 18:29:51.938 [ERROR] 
org.apache.flink.test.checkpointing.CheckpointIntervalDuringBacklogITCase.testNoCheckpointDuringBacklog
 -- Time elapsed: 2.022 s <<< FAILURE!
Jan 15 18:29:51 org.opentest4j.AssertionFailedError: 
Jan 15 18:29:51 
Jan 15 18:29:51 expected: 0
Jan 15 18:29:51  but was: 1
Jan 15 18:29:51 at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
Jan 15 18:29:51 at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
Jan 15 18:29:51 at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
Jan 15 18:29:51 at 
org.apache.flink.test.checkpointing.CheckpointIntervalDuringBacklogITCase.testNoCheckpointDuringBacklog(CheckpointIntervalDuringBacklogITCase.java:141)
Jan 15 18:29:51 at java.lang.reflect.Method.invoke(Method.java:498)
 {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33971]Specifies whether to use HBase table that supports dynamic columns. [flink-connector-hbase]

2024-01-15 Thread via GitHub


MOBIN-F commented on PR #36:
URL: 
https://github.com/apache/flink-connector-hbase/pull/36#issuecomment-1892973838

   > > the main question is that is it a good idea to introduce 
platform-specific features and flags to the IMO common upstream 
connector.主要问题是,将特定于平台的功能和标志引入 IMO 通用上游连接器是否是一个好主意。
   > 
   > If this isn't something that isn't supported in open source HBase itself, 
then at least we should first have a discussion on this topic in the Dev 
mailing list.如果开源 HBase 本身不支持这一点,那么至少我们应该首先在 Dev 邮件列表中讨论这个话题。
   
   @MartijnVisser @ferenc-csaky 
   open source hbase is supported. We have used this function internally 
online. The version is also open source hbase and has nothing to do with the 
cloud platform.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-29802][state] Changelog supports native savepoint [flink]

2024-01-15 Thread via GitHub


masteryhx commented on PR #22744:
URL: https://github.com/apache/flink/pull/22744#issuecomment-1892973498

   @rkhachatryan Could you help to take a review again ?
   cc @Zakelly You may also be interested in this.
   Since Changelog doesn't support NO_CLAIM mode yet, we could let the native 
savepoint usable so that users could have a quick backup of snapshot to other 
jobs (see 
[comment](https://issues.apache.org/jira/browse/FLINK-25322?focusedCommentId=17738073=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-17738073))


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-33668) Decoupling Shuffle network memory and job topology

2024-01-15 Thread Jiang Xin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33668?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jiang Xin updated FLINK-33668:
--
Description: 
With FLINK-30469  and FLINK-31643, we have decoupled the shuffle network memory 
and the parallelism of tasks by limiting the number of buffers for each 
InputGate and ResultPartition in Hybrid Shuffle. However, when a task contains 
multiple ResultPartitions or many shuffle tasks are running simultaneously on 
the same TaskManager, "Insufficient number of network buffers" errors would 
still occur. This usually happens when Slot Sharing Group is enabled or a 
TaskManager contains multiple slots.

We want to make sure that the TaskManager does not encounter "Insufficient 
number of network buffers" even if there are dozens of InputGates and 
ResultPartitions running on the same TaskManager simultaneously. I have given 
this some thought, and here is my rough proposal.

1. InputGate or ResultPartition only apply for buffers from LocalBufferPool, 
which means that InputGate will no longer ask for exclusive buffers from 
NetworkBufferPool directly.
2. When creating a LocalBufferPool, we need to specify the maximum, minimum, 
and expected number of buffers. Whenever a new LBP is created or destroyed, a 
redistribution will occur, to adjust the buffer count of all LocalBufferPools 
using the expected value as a weight and between the minimum and maximum 
values. According to the test, the minimum value can be set to 4 to make the 
Flink Job work despite the possibility of lower performance. With this minimum 
value, a task with 20 shuffle edges needs only 5MB of memory to avoid 
"insufficient network buffer" error.
3. During runtime, InputGate and ResultPartition both calculate the number of 
buffers used by their internal data structures based on the pool size of their 
corresponding LocalBufferPool, such as the exclusive buffer queue of InputGate 
and BufferAccumulator of ResultPartition.

  was:
With FLINK-30469  and FLINK-31643, we have decoupled the shuffle network memory 
and the parallelism of tasks by limiting the number of buffers for each 
InputGate and ResultPartition in Hybrid Shuffle. However, when too many shuffle 
tasks are running simultaneously on the same TaskManager, "Insufficient number 
of network buffers" errors would still occur. This usually happens when Slot 
Sharing Group is enabled or a TaskManager contains multiple slots.

We want to make sure that the TaskManager does not encounter "Insufficient 
number of network buffers" even if there are dozens of InputGates and 
ResultPartitions running on the same TaskManager simultaneously. I have given 
this some thought, and here is my rough proposal.

1. InputGate or ResultPartition only apply for buffers from LocalBufferPool, 
which means that InputGate will no longer ask for exclusive buffers from 
NetworkBufferPool directly.
2. When creating a LocalBufferPool, we need to specify the maximum, minimum, 
and expected number of buffers. Whenever a new LBP is created or destroyed, a 
redistribution will occur, to adjust the buffer count of all LocalBufferPools 
using the expected value as a weight and between the minimum and maximum 
values. According to the test, the minimum value can be set to 4 to make the 
Flink Job work despite the possibility of lower performance. With this minimum 
value, a task with 20 shuffle edges needs only 5MB of memory to avoid 
"insufficient network buffer" error.
3. During runtime, InputGate and ResultPartition both calculate the number of 
buffers used by their internal data structures based on the pool size of their 
corresponding LocalBufferPool, such as the exclusive buffer queue of InputGate 
and BufferAccumulator of ResultPartition.


> Decoupling Shuffle network memory and job topology
> --
>
> Key: FLINK-33668
> URL: https://issues.apache.org/jira/browse/FLINK-33668
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Network
>Reporter: Jiang Xin
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0
>
>
> With FLINK-30469  and FLINK-31643, we have decoupled the shuffle network 
> memory and the parallelism of tasks by limiting the number of buffers for 
> each InputGate and ResultPartition in Hybrid Shuffle. However, when a task 
> contains multiple ResultPartitions or many shuffle tasks are running 
> simultaneously on the same TaskManager, "Insufficient number of network 
> buffers" errors would still occur. This usually happens when Slot Sharing 
> Group is enabled or a TaskManager contains multiple slots.
> We want to make sure that the TaskManager does not encounter "Insufficient 
> number of network buffers" even if there are dozens of InputGates and 
> ResultPartitions running on the same TaskManager 

Re: [PR] [FLINK-34078][state]Move InternalKeyContext classes from o.a.f.runtim… [flink]

2024-01-15 Thread via GitHub


ljz2051 commented on PR #24093:
URL: https://github.com/apache/flink/pull/24093#issuecomment-1892971770

   @masteryhx Would you please help to review this pr?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-34007) Flink Job stuck in suspend state after losing leadership in HA Mode

2024-01-15 Thread Yang Wang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34007?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807009#comment-17807009
 ] 

Yang Wang commented on FLINK-34007:
---

{quote}At least based on the reports of this Jira issue, there must have been 
an incident (which caused the lease to not be renewed)
{quote}
I am afraid we could not get this conclusion before we have the K8s APIServer 
audit logs to verify that the lease annotation did not get renewed. Because it 
could also happen that the lease annotation get renewed normally while the 
onStartLeading callback is not executed somehow. 

 
{quote}Therefore, the issue should exist in the entire version range [5.12.3, 
6.6.2].
{quote}
If this issue only happened in the Flink 1.18, then it should be related with 
the fabric8 K8s client 6.6.2 behavior change. Otherwise, we still have not find 
the root cause.

 

You are right. The slight difference in the revocation protocol in the 
[FLIP-285|https://cwiki.apache.org/confluence/display/FLINK/FLIP-285%3A+Refactoring+LeaderElection+to+make+Flink+support+multi-component+leader+election+out-of-the-box]
 changes about clear the leader information in ConfigMap is not related with 
this issue.

 

BTW, if we know how to reproduce this issue, it will be easier to find the root 
cause. Because we might also need the K8s APIServer audit log to do some deep 
analysis.

> Flink Job stuck in suspend state after losing leadership in HA Mode
> ---
>
> Key: FLINK-34007
> URL: https://issues.apache.org/jira/browse/FLINK-34007
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.16.3, 1.17.2, 1.18.1, 1.18.2
>Reporter: Zhenqiu Huang
>Priority: Major
> Attachments: Debug.log, job-manager.log
>
>
> The observation is that Job manager goes to suspend state with a failed 
> container not able to register itself to resource manager after timeout.
> JM Log, see attached
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34089] Verify that the subscribed topics match the assigned topics [flink-connector-kafka]

2024-01-15 Thread via GitHub


Tan-JiaLiang commented on PR #77:
URL: 
https://github.com/apache/flink-connector-kafka/pull/77#issuecomment-1892967906

   @tzulitai Hi gordon, can you take a look when you are free?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33950) Update max aggregate functions to new type system

2024-01-15 Thread Jacky Lau (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807008#comment-17807008
 ] 

Jacky Lau commented on FLINK-33950:
---

hi [~martijnvisser] [~Zakelly] [~dwysakowicz] , i found it will have 69 test 
failed 
https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_apis/build/builds/55946/logs/104.
 and if we change the plan to make the test passed . But some test we should 
not change, because some unit test is for construct a hash agg to test some 
other rules, if we change it to sort agg, the test will not take effective.

> Update max aggregate functions to new type system
> -
>
> Key: FLINK-33950
> URL: https://issues.apache.org/jira/browse/FLINK-33950
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Affects Versions: 1.19.0
>Reporter: Jacky Lau
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33386][runtime] Support tasks balancing at slot level for Default Scheduler [flink]

2024-01-15 Thread via GitHub


KarmaGYZ commented on PR #23635:
URL: https://github.com/apache/flink/pull/23635#issuecomment-1892964860

   > @RocMarshal Just be curious about the progress, does this PR still wait 
for some comments to be addressed before it could be merged?
   
   This PR is in progress now. We plan to merge it after the complete Task 
Balancing feature is implemented.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33819] support set CompressionType for RocksDBStateBackend [flink]

2024-01-15 Thread via GitHub


masteryhx commented on code in PR #24072:
URL: https://github.com/apache/flink/pull/24072#discussion_r1452854496


##
flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBConfigurableOptions.java:
##
@@ -139,6 +149,25 @@ public class RocksDBConfigurableOptions implements 
Serializable {
 NONE.name(),
 LEVEL.name()));
 
+public static final ConfigOption COMPRESSION_TYPE =

Review Comment:
   +1 to the idea of @pnowojski 



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-34098) Not enough Azure Pipeline CI runners available?

2024-01-15 Thread lincoln lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34098?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807004#comment-17807004
 ] 

lincoln lee commented on FLINK-34098:
-

[~jingge] Can you help check it?

> Not enough Azure Pipeline CI runners available?
> ---
>
> Key: FLINK-34098
> URL: https://issues.apache.org/jira/browse/FLINK-34098
> Project: Flink
>  Issue Type: Bug
>  Components: Build System / CI
>Affects Versions: 1.18.0, 1.17.2, 1.19.0
>Reporter: Matthias Pohl
>Priority: Critical
>
> CI takes longer than usually. There might be an issue with the amount of 
> runners (on the Alibaba VMs)?



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-32659] Fix the connection leak due to exception [flink-connector-jdbc]

2024-01-15 Thread via GitHub


Aitozi closed pull request #71: [FLINK-32659] Fix the connection leak due to 
exception
URL: https://github.com/apache/flink-connector-jdbc/pull/71


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-34088) fix the problems with special table name characters of postgres and oracle and sqlserver.

2024-01-15 Thread blackpighe (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34088?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807002#comment-17807002
 ] 

blackpighe commented on FLINK-34088:


This problem has persisted in other releases as well, but take the new 
project's 3.1.1 as an example

> fix the problems with special table name characters of postgres and oracle  
> and sqlserver.
> --
>
> Key: FLINK-34088
> URL: https://issues.apache.org/jira/browse/FLINK-34088
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC
>Affects Versions: jdbc-3.1.1
>Reporter: blackpighe
>Priority: Major
> Fix For: jdbc-3.1.1
>
> Attachments: image-2024-01-15-17-10-57-586.png, 
> image-2024-01-15-17-11-11-136.png, image-2024-01-15-17-13-16-448.png
>
>   Original Estimate: 48h
>  Remaining Estimate: 48h
>
>  
> !image-2024-01-15-17-11-11-136.png!  
> [https://github.com/apache/flink-connector-jdbc/]
> The new version of flink jdbc is stand-alone and does not have an open issue 
> entry.
>  
> The FLink-JDBC-connector does not handle special table names for oracle and 
> postgres and sqlserver, when the table name is a special character such as 
> default. jdbc will report an error. hopefully giving me a chance to fix the 
> problem, I can provide a commit
>  
>  
> !image-2024-01-15-17-13-16-448.png!
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34088) fix the problems with special table name characters of postgres and oracle and sqlserver.

2024-01-15 Thread blackpighe (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34088?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17807001#comment-17807001
 ] 

blackpighe commented on FLINK-34088:


i edit and update the issue, by use english

> fix the problems with special table name characters of postgres and oracle  
> and sqlserver.
> --
>
> Key: FLINK-34088
> URL: https://issues.apache.org/jira/browse/FLINK-34088
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC
>Affects Versions: jdbc-3.1.1
>Reporter: blackpighe
>Priority: Major
> Fix For: jdbc-3.1.1
>
> Attachments: image-2024-01-15-17-10-57-586.png, 
> image-2024-01-15-17-11-11-136.png, image-2024-01-15-17-13-16-448.png
>
>   Original Estimate: 48h
>  Remaining Estimate: 48h
>
>  
> !image-2024-01-15-17-11-11-136.png!  
> [https://github.com/apache/flink-connector-jdbc/]
> The new version of flink jdbc is stand-alone and does not have an open issue 
> entry.
>  
> The FLink-JDBC-connector does not handle special table names for oracle and 
> postgres and sqlserver, when the table name is a special character such as 
> default. jdbc will report an error. hopefully giving me a chance to fix the 
> problem, I can provide a commit
>  
>  
> !image-2024-01-15-17-13-16-448.png!
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34088) fix the problems with special table name characters of postgres and oracle and sqlserver.

2024-01-15 Thread blackpighe (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34088?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

blackpighe updated FLINK-34088:
---
Fix Version/s: jdbc-3.1.1
Affects Version/s: jdbc-3.1.1
   (was: 1.17.2)
  Description: 
 

!image-2024-01-15-17-11-11-136.png!  

[https://github.com/apache/flink-connector-jdbc/]

The new version of flink jdbc is stand-alone and does not have an open issue 
entry.

 

The FLink-JDBC-connector does not handle special table names for oracle and 
postgres and sqlserver, when the table name is a special character such as 
default. jdbc will report an error. hopefully giving me a chance to fix the 
problem, I can provide a commit

 

 

!image-2024-01-15-17-13-16-448.png!

 

  was:
 

!image-2024-01-15-17-11-11-136.png!  

jdbc使用特殊表名需要进行转义,发现1.17版本,jdbc 
connector单独移出去成为了一个新项目,我希望给一个commit的机会,来修复这个问题,现在https://github.com/apache/flink-connector-jdbc/这个项目没有create
 issue的入口

!image-2024-01-15-17-13-16-448.png!

 


> fix the problems with special table name characters of postgres and oracle  
> and sqlserver.
> --
>
> Key: FLINK-34088
> URL: https://issues.apache.org/jira/browse/FLINK-34088
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC
>Affects Versions: jdbc-3.1.1
>Reporter: blackpighe
>Priority: Major
> Fix For: jdbc-3.1.1
>
> Attachments: image-2024-01-15-17-10-57-586.png, 
> image-2024-01-15-17-11-11-136.png, image-2024-01-15-17-13-16-448.png
>
>   Original Estimate: 48h
>  Remaining Estimate: 48h
>
>  
> !image-2024-01-15-17-11-11-136.png!  
> [https://github.com/apache/flink-connector-jdbc/]
> The new version of flink jdbc is stand-alone and does not have an open issue 
> entry.
>  
> The FLink-JDBC-connector does not handle special table names for oracle and 
> postgres and sqlserver, when the table name is a special character such as 
> default. jdbc will report an error. hopefully giving me a chance to fix the 
> problem, I can provide a commit
>  
>  
> !image-2024-01-15-17-13-16-448.png!
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


  1   2   3   4   >