[GitHub] [flink] Aitozi commented on pull request #21174: [FLINK-29557] Fix the SinkOperator with OutputFormatFunction is ignor…

2022-10-27 Thread GitBox


Aitozi commented on PR #21174:
URL: https://github.com/apache/flink/pull/21174#issuecomment-1294497007

   Maybe it's caused by this: https://issues.apache.org/jira/browse/FLINK-29787


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] Aitozi commented on pull request #21174: [FLINK-29557] Fix the SinkOperator with OutputFormatFunction is ignor…

2022-10-27 Thread GitBox


Aitozi commented on PR #21174:
URL: https://github.com/apache/flink/pull/21174#issuecomment-1294495106

   Hi @gaoyunhaii , can you help me understand the CI failure? 
   
   > [ERROR] Failed to execute goal 
io.github.zentol.japicmp:japicmp-maven-plugin:0.16.0_m325:cmp (default) on 
project flink-core: There is at least one incompatibility: 
org.apache.flink.api.connector.source.SourceReader.pauseOrResumeSplits(java.util.Collection,java.util.Collection):METHOD_NEW_DEFAULT
 -> [Help 1]
   
   The reported method seems not related to this PR, a little confused for the 
result 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-26390) InputStream is not closed at CliClient#executeFile

2022-10-27 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-26390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-26390:
--
Fix Version/s: 1.17.0

> InputStream is not closed at CliClient#executeFile
> --
>
> Key: FLINK-26390
> URL: https://issues.apache.org/jira/browse/FLINK-26390
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Table SQL / Client
>Affects Versions: 1.15.0
>Reporter: Sergey Nuyanzin
>Assignee: Sergey Nuyanzin
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
> Fix For: 1.17.0
>
>
> There is {{InputStream}} created however not closed
> https://github.com/apache/flink/blob/e2463ddace26601c6442dbaa68c198a27c9cfbcf/flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/CliClient.java#L349-L364
> Another thing, which is easy to fix: it seems there is no need to create 
> completer object for {{CliClient#executeFile}} since it makes sense only in 
> interactive mode 
> It seems the reason is https://github.com/apache/flink/pull/18363/files



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29748) Expose the optimize phase in the connector context

2022-10-27 Thread Aitozi (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29748?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625440#comment-17625440
 ] 

Aitozi commented on FLINK-29748:


As I know, the \{{getScanRuntimeProvider}} will be invoked serval times during 
the optimize, eg:
 * 
org.apache.flink.table.planner.connectors.DynamicSourceUtils#validateScanSource
 * 
org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecTableSourceScan#translateToPlanInternal

So, to the connector, it do not know whether the optimization is finished.

> Expose the optimize phase in the connector context
> --
>
> Key: FLINK-29748
> URL: https://issues.apache.org/jira/browse/FLINK-29748
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner, Table SQL / Runtime
>Reporter: Aitozi
>Priority: Minor
>
> Currently, in the connector it can not know whether the whole optimize is 
> finished.
> When the optimize finished, the all information is static, eg: the reading 
> partitions. If I want to validate the final optimized result (like whether 
> the reading partition is too much or empty), it needs the context of what is 
> the current phase. I think the {{ScanContext}} is ok to expose this 
> information. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-26390) InputStream is not closed at CliClient#executeFile

2022-10-27 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-26390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl resolved FLINK-26390.
---
Resolution: Fixed

master: 74c0ace01b61f18cda5c13abf0b940fe81f13e8a

> InputStream is not closed at CliClient#executeFile
> --
>
> Key: FLINK-26390
> URL: https://issues.apache.org/jira/browse/FLINK-26390
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Table SQL / Client
>Affects Versions: 1.15.0
>Reporter: Sergey Nuyanzin
>Assignee: Sergey Nuyanzin
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> There is {{InputStream}} created however not closed
> https://github.com/apache/flink/blob/e2463ddace26601c6442dbaa68c198a27c9cfbcf/flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/CliClient.java#L349-L364
> Another thing, which is easy to fix: it seems there is no need to create 
> completer object for {{CliClient#executeFile}} since it makes sense only in 
> interactive mode 
> It seems the reason is https://github.com/apache/flink/pull/18363/files



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] XComp merged pull request #18927: [FLINK-26390] Use try resourses for CliClient#executeFile

2022-10-27 Thread GitBox


XComp merged PR #18927:
URL: https://github.com/apache/flink/pull/18927


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-ml] lindong28 commented on a diff in pull request #167: [FLINK-29786] VarianceThresholdSelector Uses FeaturesCol Incorrectly

2022-10-27 Thread GitBox


lindong28 commented on code in PR #167:
URL: https://github.com/apache/flink-ml/pull/167#discussion_r1007646156


##
flink-ml-lib/src/main/java/org/apache/flink/ml/common/param/HasFeaturesCol.java:
##
@@ -24,6 +24,14 @@
 import org.apache.flink.ml.param.WithParams;
 
 /** Interface for the shared featuresCol param. */
+
+/**
+ * Interface for the shared featuresCol param.

Review Comment:
   Should we remove the redundant description above?
   
   And it might be simpler and softer to use the following doc:
   
   ```
   /**
* Interface for the shared featuresCol param.
*
* {@link HasFeaturesCol} is typically used for {@link Stage}s that 
implement {@link HasLabelCol}. It is preferred
* to use {@link HasInputCol} for other cases.
*/
   ```
   
   Please also update the corresponding python doc if we make the change here.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-ml] yunfengzhou-hub commented on a diff in pull request #166: [FLINK-29598] Add Estimator and Transformer for Imputer

2022-10-27 Thread GitBox


yunfengzhou-hub commented on code in PR #166:
URL: https://github.com/apache/flink-ml/pull/166#discussion_r1007561913


##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/imputer/ImputerParams.java:
##
@@ -0,0 +1,59 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.feature.imputer;
+
+import org.apache.flink.ml.param.Param;
+import org.apache.flink.ml.param.ParamValidators;
+import org.apache.flink.ml.param.StringParam;
+
+/**
+ * Params of {@link Imputer}.
+ *
+ * @param  The class type of this instance.
+ */
+public interface ImputerParams extends ImputerModelParams {

Review Comment:
   Spark's `Imputer` also has a `relativeError` parameter. Should we also add 
this parameter to Flink ML?



##
flink-ml-core/src/main/java/org/apache/flink/ml/param/DoubleParam.java:
##
@@ -32,4 +34,12 @@ public DoubleParam(
 public DoubleParam(String name, String description, Double defaultValue) {
 this(name, description, defaultValue, ParamValidators.alwaysTrue());
 }
+
+@Override
+public Double jsonDecode(Object json) throws IOException {
+if (json instanceof String && json.equals(String.valueOf(Double.NaN))) 
{
+return Double.NaN;
+}
+return (Double) json;

Review Comment:
   How about the following implementation?
   ```java
   if (json instanceof String) {
   return Double.valueOf((String) json);
   }
   return (Double) json;
   ```
   This applies to `Double.NaN` as well as other special values.



##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/imputer/Imputer.java:
##
@@ -0,0 +1,328 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.feature.imputer;
+
+import org.apache.flink.api.common.functions.AggregateFunction;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.ml.api.Estimator;
+import org.apache.flink.ml.common.datastream.DataStreamUtils;
+import org.apache.flink.ml.common.util.QuantileSummary;
+import org.apache.flink.ml.param.Param;
+import org.apache.flink.ml.util.ParamUtils;
+import org.apache.flink.ml.util.ReadWriteUtils;
+import org.apache.flink.streaming.api.datastream.DataStream;
+import org.apache.flink.table.api.Table;
+import org.apache.flink.table.api.bridge.java.StreamTableEnvironment;
+import org.apache.flink.table.api.internal.TableImpl;
+import org.apache.flink.types.Row;
+import org.apache.flink.util.Preconditions;
+
+import java.io.IOException;
+import java.util.Arrays;
+import java.util.HashMap;
+import java.util.Map;
+
+/**
+ * The Imputer estimator completes missing values in a dataset. Missing values 
can be imputed using

Review Comment:
   nit: it might be better to use "bounded stream" instead of "dataset", as 
dataset has a specific meaning in flink.



##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/imputer/ImputerModel.java:
##
@@ -0,0 +1,174 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * 

[jira] [Comment Edited] (FLINK-29541) [JUnit5 Migration] Module: flink-table-planner

2022-10-27 Thread Lijie Wang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29541?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625437#comment-17625437
 ] 

Lijie Wang edited comment on FLINK-29541 at 10/28/22 5:37 AM:
--

Sure, just assigned to you :). It would be better to create a separate issue to 
track these abstract base classes, just like what you did in FLINK-28542


was (Author: wanglijie95):
Sure, just assigned to you :). It would be better to create a separate issue 
for these abstract base classes to track them, just like what you did in 
FLINK-28542

> [JUnit5 Migration] Module: flink-table-planner
> --
>
> Key: FLINK-29541
> URL: https://issues.apache.org/jira/browse/FLINK-29541
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner, Tests
>Reporter: Lijie Wang
>Assignee: Ryan Skraba
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-29541) [JUnit5 Migration] Module: flink-table-planner

2022-10-27 Thread Lijie Wang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29541?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lijie Wang reassigned FLINK-29541:
--

Assignee: Ryan Skraba

> [JUnit5 Migration] Module: flink-table-planner
> --
>
> Key: FLINK-29541
> URL: https://issues.apache.org/jira/browse/FLINK-29541
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner, Tests
>Reporter: Lijie Wang
>Assignee: Ryan Skraba
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29541) [JUnit5 Migration] Module: flink-table-planner

2022-10-27 Thread Lijie Wang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29541?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625437#comment-17625437
 ] 

Lijie Wang commented on FLINK-29541:


Sure, just assigned to you :). It would be better to create a separate issue 
for these abstract base classes to track them, just like what you did in 
FLINK-28542

> [JUnit5 Migration] Module: flink-table-planner
> --
>
> Key: FLINK-29541
> URL: https://issues.apache.org/jira/browse/FLINK-29541
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner, Tests
>Reporter: Lijie Wang
>Assignee: Ryan Skraba
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-29565) In Flink per job mode, the logs printed by taskManager on the web UI will not be highlighted, because the log contents are annotated due to special symbols, which will

2022-10-27 Thread Yun Gao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29565?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yun Gao reassigned FLINK-29565:
---

Assignee: wangshiwei

> In Flink per job mode, the logs printed by taskManager on the web UI will not 
> be highlighted, because the log contents are annotated due to special 
> symbols, which will affect the use experience. For more information, see Fig
> 
>
> Key: FLINK-29565
> URL: https://issues.apache.org/jira/browse/FLINK-29565
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Web Frontend
>Affects Versions: 1.14.3
>Reporter: wangshiwei
>Assignee: wangshiwei
>Priority: Minor
> Attachments: image-2022-10-10-18-40-27-721.png, 
> image-2022-10-10-18-43-31-897.png, image-2022-10-10-18-43-53-713.png, 
> image-2022-10-10-18-45-17-228.png, image-2022-10-10-19-02-29-796.png, 
> image-2022-10-10-19-03-27-670.png
>
>
>  
> !image-2022-10-10-19-03-27-670.png|width=580,height=317!
> !image-2022-10-10-18-43-53-713.png|width=726,height=47!
> This kind of '/*' content will appear in the print log of the 
> logEnvironmentInfo method in the EnvironmentInformation class. The following 
> logs will be commented out without highlighting
> *verification*
> !image-2022-10-10-18-45-17-228.png|width=880,height=161!
> After manually printing '*/' in the business code, the log is normal
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29565) In Flink per job mode, the logs printed by taskManager on the web UI will not be highlighted, because the log contents are annotated due to special symbols, which will

2022-10-27 Thread Yun Gao (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625435#comment-17625435
 ] 

Yun Gao commented on FLINK-29565:
-

Thanks [~1336399775] for helping fix the issue! I have assigned the issue to 
you. 

> In Flink per job mode, the logs printed by taskManager on the web UI will not 
> be highlighted, because the log contents are annotated due to special 
> symbols, which will affect the use experience. For more information, see Fig
> 
>
> Key: FLINK-29565
> URL: https://issues.apache.org/jira/browse/FLINK-29565
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Web Frontend
>Affects Versions: 1.14.3
>Reporter: wangshiwei
>Priority: Minor
> Attachments: image-2022-10-10-18-40-27-721.png, 
> image-2022-10-10-18-43-31-897.png, image-2022-10-10-18-43-53-713.png, 
> image-2022-10-10-18-45-17-228.png, image-2022-10-10-19-02-29-796.png, 
> image-2022-10-10-19-03-27-670.png
>
>
>  
> !image-2022-10-10-19-03-27-670.png|width=580,height=317!
> !image-2022-10-10-18-43-53-713.png|width=726,height=47!
> This kind of '/*' content will appear in the print log of the 
> logEnvironmentInfo method in the EnvironmentInformation class. The following 
> logs will be commented out without highlighting
> *verification*
> !image-2022-10-10-18-45-17-228.png|width=880,height=161!
> After manually printing '*/' in the business code, the log is normal
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] lindong28 commented on a diff in pull request #162: [FLINK-29593] Add QuantileSummary to help calculate approximate quantiles

2022-10-27 Thread GitBox


lindong28 commented on code in PR #162:
URL: https://github.com/apache/flink-ml/pull/162#discussion_r1007633262


##
flink-ml-lib/src/main/java/org/apache/flink/ml/common/util/QuantileSummary.java:
##
@@ -0,0 +1,400 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.common.util;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.util.Preconditions;
+
+import java.io.Serializable;
+import java.util.ArrayList;
+import java.util.Arrays;
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.LinkedList;
+import java.util.List;
+import java.util.Map;
+import java.util.stream.Collectors;
+import java.util.stream.IntStream;
+
+/**
+ * Helper class to compute an approximate quantile summary. This 
implementation is based on the
+ * algorithm proposed in the paper: "Space-efficient Online Computation of 
Quantile Summaries" by
+ * Greenwald, Michael and Khanna, Sanjeev. 
(https://doi.org/10.1145/375663.375670)
+ */
+@Internal
+public class QuantileSummary implements Serializable {
+
+/** The target relative error. */
+private final double relativeError;
+
+/**
+ * The compression threshold. After the internal buffer of statistics 
crosses this size, it
+ * attempts to compress the statistics together.
+ */
+private final int compressThreshold;
+
+/** The count of all the elements inserted to be calculated. */
+private final long count;
+
+/** A buffer of quantile statistics. */
+private final List sampled;
+
+/** The default size of head buffer. */
+private static final int DEFAULT_HEAD_SIZE = 5;

Review Comment:
   `private static final` variables are typically put above `private final` 
variables.



##
flink-ml-lib/src/main/java/org/apache/flink/ml/common/util/QuantileSummary.java:
##
@@ -0,0 +1,400 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.common.util;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.util.Preconditions;
+
+import java.io.Serializable;
+import java.util.ArrayList;
+import java.util.Arrays;
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.LinkedList;
+import java.util.List;
+import java.util.Map;
+import java.util.stream.Collectors;
+import java.util.stream.IntStream;
+
+/**
+ * Helper class to compute an approximate quantile summary. This 
implementation is based on the
+ * algorithm proposed in the paper: "Space-efficient Online Computation of 
Quantile Summaries" by
+ * Greenwald, Michael and Khanna, Sanjeev. 
(https://doi.org/10.1145/375663.375670)
+ */
+@Internal

Review Comment:
   According to the Java doc of `@Internal`, it is an "Annotation to mark 
methods within stable, public APIs as an internal developer API".
   
   Thus it seems unnecessary to add this annotation.
   
   @zhipeng93 Could you confirm that we should remove this annotation?



##
flink-ml-lib/src/main/java/org/apache/flink/ml/common/util/QuantileSummary.java:
##
@@ -0,0 +1,400 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in 

[jira] [Comment Edited] (FLINK-29711) Topic notification not present in metadata after 60000 ms.

2022-10-27 Thread Durgesh Mishra (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17624881#comment-17624881
 ] 

Durgesh Mishra edited comment on FLINK-29711 at 10/28/22 5:25 AM:
--

[~mason6345]  I added logs in class which implements 
KafkaRecordSerializationSchema.


was (Author: JIRAUSER297316):
[~mason6345]  I added logs in class which implements 
KafkaRecordSerializationSchema.

> Topic notification not present in metadata after 6 ms.
> --
>
> Key: FLINK-29711
> URL: https://issues.apache.org/jira/browse/FLINK-29711
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Affects Versions: 1.14.4, 1.14.6
>Reporter: Durgesh Mishra
>Priority: Major
>
> Failed to send data to Kafka null with 
> FlinkKafkaInternalProducer\{transactionalId='null', inTransaction=false, 
> closed=false}
> at 
> org.apache.flink.connector.kafka.sink.KafkaWriter$WriterCallback.throwException(KafkaWriter.java:405)
> at 
> org.apache.flink.connector.kafka.sink.KafkaWriter$WriterCallback.lambda$onCompletion$0(KafkaWriter.java:391)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$1.runThrowing(StreamTaskActionExecutor.java:50)
> at org.apache.flink.streaming.runtime.tasks.mailbox.Mail.run(Mail.java:90)
> at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.processMailsNonBlocking(MailboxProcessor.java:353)
> at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.processMail(MailboxProcessor.java:317)
> at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:201)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:809)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:761)
> at 
> org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:958)
> at org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:937)
> at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:766)
> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:575)
> at java.base/java.lang.Thread.run(Unknown Source)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] liuml07 commented on pull request #20917: [hotfix][doc] Refine code examples in data_stream_api

2022-10-27 Thread GitBox


liuml07 commented on PR #20917:
URL: https://github.com/apache/flink/pull/20917#issuecomment-1294473427

   Thank you @Myasuka for prompt review and commit.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] liuml07 commented on a diff in pull request #20935: [hotfix][doc] Make referencing custom image clearer for Docker

2022-10-27 Thread GitBox


liuml07 commented on code in PR #20935:
URL: https://github.com/apache/flink/pull/20935#discussion_r1007632470


##
docs/content.zh/docs/deployment/resource-providers/standalone/docker.md:
##
@@ -406,15 +406,15 @@ services:
   You can then start creating tables and queries those.
 
 * Note, that all required dependencies (e.g. for connectors) need to be 
available in the cluster as well as the client.
-  For example, if you would like to use the Kafka Connector create a custom 
image with the following Dockerfile
+  For example, if you would like to use the Kafka Connector, you can build a 
custom image.
+  Create a Dockerfile as follows:
 
   ```Dockerfile
   FROM flink:{{< stable >}}{{< version >}}-scala{{< scala_version >}}{{< 
/stable >}}{{< unstable >}}latest{{< /unstable >}}
   RUN wget -P /opt/flink/lib 
https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-kafka_2.12/{{<
 version >}}/flink-sql-connector-kafka_scala{{< scala_version >}}-{{< version 
>}}.jar
   ```
 
-  and reference it (e.g via the `build`) command in the Dockerfile.
-  and reference it (e.g via the `build`) command in the Dockerfile.
+  and reference this Dockerfile with the `build` command in the 
`docker-compose.yml` file.

Review Comment:
   Updated the doc. After building the doc screenshot is like:
   https://user-images.githubusercontent.com/159186/198509170-969dc74b-4df4-4a80-a2ca-d0fa6be7dca1.png;>
   
   



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] liuml07 commented on a diff in pull request #20935: [hotfix][doc] Make referencing custom image clearer for Docker

2022-10-27 Thread GitBox


liuml07 commented on code in PR #20935:
URL: https://github.com/apache/flink/pull/20935#discussion_r1007632470


##
docs/content.zh/docs/deployment/resource-providers/standalone/docker.md:
##
@@ -406,15 +406,15 @@ services:
   You can then start creating tables and queries those.
 
 * Note, that all required dependencies (e.g. for connectors) need to be 
available in the cluster as well as the client.
-  For example, if you would like to use the Kafka Connector create a custom 
image with the following Dockerfile
+  For example, if you would like to use the Kafka Connector, you can build a 
custom image.
+  Create a Dockerfile as follows:
 
   ```Dockerfile
   FROM flink:{{< stable >}}{{< version >}}-scala{{< scala_version >}}{{< 
/stable >}}{{< unstable >}}latest{{< /unstable >}}
   RUN wget -P /opt/flink/lib 
https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-kafka_2.12/{{<
 version >}}/flink-sql-connector-kafka_scala{{< scala_version >}}-{{< version 
>}}.jar
   ```
 
-  and reference it (e.g via the `build`) command in the Dockerfile.
-  and reference it (e.g via the `build`) command in the Dockerfile.
+  and reference this Dockerfile with the `build` command in the 
`docker-compose.yml` file.

Review Comment:
   Updated the doc. After building the doc screenshot is like:
   https://user-images.githubusercontent.com/159186/198507938-d68b6ee6-c642-48cc-8baa-5f125c210109.png;>
   



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] flinkbot commented on pull request #21183: [FLINK-29789][tests] Fix flaky test in CheckpointCoordinatorTest

2022-10-27 Thread GitBox


flinkbot commented on PR #21183:
URL: https://github.com/apache/flink/pull/21183#issuecomment-1294459420

   
   ## CI report:
   
   * 6cb844d32692779b408f5f216899183ff2f849b7 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] sopan98 opened a new pull request, #21183: [FLINK-29789][tests] Fix flaky test in CheckpointCoordinatorTest

2022-10-27 Thread GitBox


sopan98 opened a new pull request, #21183:
URL: https://github.com/apache/flink/pull/21183

   This PR aims to solve the issue presented here: 
https://issues.apache.org/jira/browse/FLINK-29789
   
   ## What is the purpose of the change
   The fix is to change the HashMap of PendingCheckpoint in 
CheckpointCoordinator to LinkedHashMap to make the tests more stable (less 
flaky).
   
   ## Brief change log
   - Changing the map implementation of PendingCheckpoint in 
CheckpointCoordinator
   
   ## Verifying this change
   Existing tests already cover this change, and it can pass them successfully.
   The tests are: 
   1. 
org.apache.flink.runtime.checkpoint.CheckpointCoordinatorTest.testTriggerAndDeclineCheckpointComplex
   
   ## Does this pull request potentially affect one of the following parts:
   * Dependencies (does it add or upgrade a dependency): No
   * The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: No
   * The serializers: No
   * The runtime per-record code paths (performance sensitive): No
   * Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos,, ZooKeeper: No
   * The S3 file system connector: No
   
   ## Documentation
   Does this pull request introduce a new feature? No


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-29789) Fix flaky tests in CheckpointCoordinatorTest

2022-10-27 Thread Sopan Phaltankar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29789?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sopan Phaltankar updated FLINK-29789:
-
Description: 
The test 
org.apache.flink.runtime.checkpoint.CheckpointCoordinatorTest.testTriggerAndDeclineCheckpointComplex
 is flaky and has the following failure:

Failures:
[ERROR] Failures:
[ERROR]   CheckpointCoordinatorTest.testTriggerAndDeclineCheckpointComplex:1054 
expected:<2> but was:<1>

I used the tool [NonDex|https://github.com/TestingResearchIllinois/NonDex] to 
find this flaky test.
Command: mvn -pl flink-runtime edu.illinois:nondex-maven-plugun:1.1.2:nondex 
-Dtest=org.apache.flink.runtime.checkpoint.CheckpointCoordinatorTest#testTriggerAndDeclineCheckpointComplex

I analyzed the assertion failure and found that checkpoint1Id and checkpoint2Id 
are getting assigned by iterating over a HashMap.
As we know, iterator() returns elements in a random order 
[(JavaDoc|https://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html#entrySet--])
 and this might cause test failures for some orders.

Therefore, to remove this non-determinism, we would change HashMap to 
LinkedHashMap.
On further analysis, it was found that the Map is getting initialized on line 
1894 of org.apache.flink.runtime.checkpoint.CheckpointCoordinator class.

After changing from HashMap to LinkedHashMap, the above test is passing without 
any non-determinism.

  was:
The test 
org.apache.flink.runtime.checkpoint.CheckpointCoordinatorTest.testTriggerAndDeclineCheckpointComplex
 is flaky and has the following failure:

Failures:
[ERROR] Failures:
[ERROR]   CheckpointCoordinatorTest.testTriggerAndDeclineCheckpointComplex:1054 
expected:<2> but was:<1>

I used the tool (NonDex|https://github.com/TestingResearchIllinois/NonDex) to 
find this flaky test.
Command: mvn -pl flink-runtime edu.illinois:nondex-maven-plugun:1.1.2:nondex 
-Dtest=org.apache.flink.runtime.checkpoint.CheckpointCoordinatorTest#testTriggerAndDeclineCheckpointComplex

I analyzed the assertion failure and found that checkpoint1Id and checkpoint2Id 
are getting assigned by iterating over a HashMap.
As we know, iterator() returns elements in a random order 
(JavaDoc|https://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html#entrySet--)
 and this might cause test failures for some orders.

Therefore, to remove this non-determinism, we would change HashMap to 
LinkedHashMap.
On further analysis, it was found that the Map is getting initialized on line 
1894 of org.apache.flink.runtime.checkpoint.CheckpointCoordinator class.

After changing from HashMap to LinkedHashMap, the above test is passing without 
any non-determinism.


> Fix flaky tests in CheckpointCoordinatorTest
> 
>
> Key: FLINK-29789
> URL: https://issues.apache.org/jira/browse/FLINK-29789
> Project: Flink
>  Issue Type: Bug
>Reporter: Sopan Phaltankar
>Priority: Minor
>  Labels: pull-request-available
>
> The test 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinatorTest.testTriggerAndDeclineCheckpointComplex
>  is flaky and has the following failure:
> Failures:
> [ERROR] Failures:
> [ERROR]   
> CheckpointCoordinatorTest.testTriggerAndDeclineCheckpointComplex:1054 
> expected:<2> but was:<1>
> I used the tool [NonDex|https://github.com/TestingResearchIllinois/NonDex] to 
> find this flaky test.
> Command: mvn -pl flink-runtime edu.illinois:nondex-maven-plugun:1.1.2:nondex 
> -Dtest=org.apache.flink.runtime.checkpoint.CheckpointCoordinatorTest#testTriggerAndDeclineCheckpointComplex
> I analyzed the assertion failure and found that checkpoint1Id and 
> checkpoint2Id are getting assigned by iterating over a HashMap.
> As we know, iterator() returns elements in a random order 
> [(JavaDoc|https://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html#entrySet--])
>  and this might cause test failures for some orders.
> Therefore, to remove this non-determinism, we would change HashMap to 
> LinkedHashMap.
> On further analysis, it was found that the Map is getting initialized on line 
> 1894 of org.apache.flink.runtime.checkpoint.CheckpointCoordinator class.
> After changing from HashMap to LinkedHashMap, the above test is passing 
> without any non-determinism.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-web] godfreyhe closed pull request #574: Announcement blogpost for the 1.16 release

2022-10-27 Thread GitBox


godfreyhe closed pull request #574: Announcement blogpost for the 1.16 release
URL: https://github.com/apache/flink-web/pull/574


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-web] godfreyhe commented on pull request #574: Announcement blogpost for the 1.16 release

2022-10-27 Thread GitBox


godfreyhe commented on PR #574:
URL: https://github.com/apache/flink-web/pull/574#issuecomment-1294451826

   merge & close


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-29789) Fix flaky tests in CheckpointCoordinatorTest

2022-10-27 Thread Sopan Phaltankar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29789?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sopan Phaltankar updated FLINK-29789:
-
Description: 
The test 
org.apache.flink.runtime.checkpoint.CheckpointCoordinatorTest.testTriggerAndDeclineCheckpointComplex
 is flaky and has the following failure:

Failures:
[ERROR] Failures:
[ERROR]   CheckpointCoordinatorTest.testTriggerAndDeclineCheckpointComplex:1054 
expected:<2> but was:<1>

I used the tool (NonDex|https://github.com/TestingResearchIllinois/NonDex) to 
find this flaky test.
Command: mvn -pl flink-runtime edu.illinois:nondex-maven-plugun:1.1.2:nondex 
-Dtest=org.apache.flink.runtime.checkpoint.CheckpointCoordinatorTest#testTriggerAndDeclineCheckpointComplex

I analyzed the assertion failure and found that checkpoint1Id and checkpoint2Id 
are getting assigned by iterating over a HashMap.
As we know, iterator() returns elements in a random order 
(JavaDoc|https://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html#entrySet--)
 and this might cause test failures for some orders.

Therefore, to remove this non-determinism, we would change HashMap to 
LinkedHashMap.
On further analysis, it was found that the Map is getting initialized on line 
1894 of org.apache.flink.runtime.checkpoint.CheckpointCoordinator class.

After changing from HashMap to LinkedHashMap, the above test is passing without 
any non-determinism.

  was:
The test 
_org.apache.flink.streaming.api.operators.co.CoBroadcastWithNonKeyedOperatorTest.testMultiStateSupport_
 has the following failure:

Failures:
[ERROR]   CoBroadcastWithNonKeyedOperatorTest.testMultiStateSupport:74 
Wrong Side Output: arrays first differed at element [0]; expected:6> but was:5>

I used the tool [NonDex|https://github.com/TestingResearchIllinois/NonDex] to 
find this flaky test. 
Command: mvn edu.illinois:nondex-maven-plugun:1.1.2:nondex -Dtest='Fully 
Qualified Test Name'

I analyzed the assertion failure and found that the root cause is because the 
test method calls ctx.getBroadcastState(STATE_DESCRIPTOR).immutableEntries() 
which calls the entrySet() method of the underlying HashMap. entrySet() returns 
the entries in a non-deterministic way, causing the test to be flaky. 

The fix would be to change _HashMap_ to _LinkedHashMap_ where the Map is 
getting initialized.
On further analysis, it was found that the Map is getting initialized on line 
53 of org.apache.flink.runtime.state.HeapBroadcastState class.

After changing from HashMap to LinkedHashMap, the above test is passing.

Edit: Upon making this change and running the CI, it was found that the tests 
org.apache.flink.api.datastream.DataStreamBatchExecutionITCase.batchKeyedBroadcastExecution
 and 
org.apache.flink.api.datastream.DataStreamBatchExecutionITCase.batchBroadcastExecution
 were failing. Upon further investigation, I found that these tests were also 
flaky and depended on the earlier made change.


> Fix flaky tests in CheckpointCoordinatorTest
> 
>
> Key: FLINK-29789
> URL: https://issues.apache.org/jira/browse/FLINK-29789
> Project: Flink
>  Issue Type: Bug
>Reporter: Sopan Phaltankar
>Priority: Minor
>  Labels: pull-request-available
>
> The test 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinatorTest.testTriggerAndDeclineCheckpointComplex
>  is flaky and has the following failure:
> Failures:
> [ERROR] Failures:
> [ERROR]   
> CheckpointCoordinatorTest.testTriggerAndDeclineCheckpointComplex:1054 
> expected:<2> but was:<1>
> I used the tool (NonDex|https://github.com/TestingResearchIllinois/NonDex) to 
> find this flaky test.
> Command: mvn -pl flink-runtime edu.illinois:nondex-maven-plugun:1.1.2:nondex 
> -Dtest=org.apache.flink.runtime.checkpoint.CheckpointCoordinatorTest#testTriggerAndDeclineCheckpointComplex
> I analyzed the assertion failure and found that checkpoint1Id and 
> checkpoint2Id are getting assigned by iterating over a HashMap.
> As we know, iterator() returns elements in a random order 
> (JavaDoc|https://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html#entrySet--)
>  and this might cause test failures for some orders.
> Therefore, to remove this non-determinism, we would change HashMap to 
> LinkedHashMap.
> On further analysis, it was found that the Map is getting initialized on line 
> 1894 of org.apache.flink.runtime.checkpoint.CheckpointCoordinator class.
> After changing from HashMap to LinkedHashMap, the above test is passing 
> without any non-determinism.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29789) Fix flaky tests in CheckpointCoordinatorTest

2022-10-27 Thread Sopan Phaltankar (Jira)
Sopan Phaltankar created FLINK-29789:


 Summary: Fix flaky tests in CheckpointCoordinatorTest
 Key: FLINK-29789
 URL: https://issues.apache.org/jira/browse/FLINK-29789
 Project: Flink
  Issue Type: Bug
Reporter: Sopan Phaltankar


The test 
_org.apache.flink.streaming.api.operators.co.CoBroadcastWithNonKeyedOperatorTest.testMultiStateSupport_
 has the following failure:

Failures:
[ERROR]   CoBroadcastWithNonKeyedOperatorTest.testMultiStateSupport:74 
Wrong Side Output: arrays first differed at element [0]; expected:6> but was:5>

I used the tool [NonDex|https://github.com/TestingResearchIllinois/NonDex] to 
find this flaky test. 
Command: mvn edu.illinois:nondex-maven-plugun:1.1.2:nondex -Dtest='Fully 
Qualified Test Name'

I analyzed the assertion failure and found that the root cause is because the 
test method calls ctx.getBroadcastState(STATE_DESCRIPTOR).immutableEntries() 
which calls the entrySet() method of the underlying HashMap. entrySet() returns 
the entries in a non-deterministic way, causing the test to be flaky. 

The fix would be to change _HashMap_ to _LinkedHashMap_ where the Map is 
getting initialized.
On further analysis, it was found that the Map is getting initialized on line 
53 of org.apache.flink.runtime.state.HeapBroadcastState class.

After changing from HashMap to LinkedHashMap, the above test is passing.

Edit: Upon making this change and running the CI, it was found that the tests 
org.apache.flink.api.datastream.DataStreamBatchExecutionITCase.batchKeyedBroadcastExecution
 and 
org.apache.flink.api.datastream.DataStreamBatchExecutionITCase.batchBroadcastExecution
 were failing. Upon further investigation, I found that these tests were also 
flaky and depended on the earlier made change.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-29611) Fix flaky tests in CoBroadcastWithNonKeyedOperatorTest

2022-10-27 Thread Sopan Phaltankar (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29611?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17619810#comment-17619810
 ] 

Sopan Phaltankar edited comment on FLINK-29611 at 10/28/22 4:35 AM:


[~martijnvisser] 
This test depends on the order of iteration in `HashMap.entrySet()` and can 
fail for some orders. I am running this on my local machine. I have used a 
maven plugin NonDex, which can be used to identify such tests. One can 
reproduce it with the command `mvn 
edu.illinois:nondex-maven-plugun:1.1.2:nondex 
-Dtest=org.apache.flink.streaming.api.operators.co.CoBroadcastWithNonKeyedOperatorTest#testMultiStateSupport`.
 Even if the test was not failing during the daily jobs, it'd be good to not 
depend on the `HashMap.entrySet()` that gives back results in an undefined 
manner 
[JavaDoc|https://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html#entrySet--]
 for this. Therefore, to remove non-determinism completely, we can do this 
change.


was (Author: JIRAUSER296865):
[~martijnvisser] 
This test depends on the order of iteration in `HashMap.entrySet()` and can 
fail for some orders. I am running this on my local machine. I have used a 
maven plugin NonDex, which can be used to identify such tests. One can 
reproduce it with the command `mvn 
edu.illinois:nondex-maven-plugun:1.1.2:nondex 
-Dtest=org.apache.flink.streaming.api.operators.co.CoBroadcastWithNonKeyedOperatorTest#testMultiStateSupport`.
 Even if the test was not failing during the daily jobs, it'd be good to not 
depend on the `HashMap.entrySet()` that gives back results in an undefined 
manner 
[JavaDoc|https://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html#entrySet--]for
 this. Therefore, to remove non-determinism completely, we can do this change.

> Fix flaky tests in CoBroadcastWithNonKeyedOperatorTest
> --
>
> Key: FLINK-29611
> URL: https://issues.apache.org/jira/browse/FLINK-29611
> Project: Flink
>  Issue Type: Bug
>Reporter: Sopan Phaltankar
>Priority: Minor
>  Labels: pull-request-available
>
> The test 
> _org.apache.flink.streaming.api.operators.co.CoBroadcastWithNonKeyedOperatorTest.testMultiStateSupport_
>  has the following failure:
> Failures:
> [ERROR]   CoBroadcastWithNonKeyedOperatorTest.testMultiStateSupport:74 
> Wrong Side Output: arrays first differed at element [0]; expected: 15 : 9:key.6->6> but was:5>
> I used the tool [NonDex|https://github.com/TestingResearchIllinois/NonDex] to 
> find this flaky test. 
> Command: mvn edu.illinois:nondex-maven-plugun:1.1.2:nondex -Dtest='Fully 
> Qualified Test Name'
> I analyzed the assertion failure and found that the root cause is because the 
> test method calls ctx.getBroadcastState(STATE_DESCRIPTOR).immutableEntries() 
> which calls the entrySet() method of the underlying HashMap. entrySet() 
> returns the entries in a non-deterministic way, causing the test to be flaky. 
> The fix would be to change _HashMap_ to _LinkedHashMap_ where the Map is 
> getting initialized.
> On further analysis, it was found that the Map is getting initialized on line 
> 53 of org.apache.flink.runtime.state.HeapBroadcastState class.
> After changing from HashMap to LinkedHashMap, the above test is passing.
> Edit: Upon making this change and running the CI, it was found that the tests 
> org.apache.flink.api.datastream.DataStreamBatchExecutionITCase.batchKeyedBroadcastExecution
>  and 
> org.apache.flink.api.datastream.DataStreamBatchExecutionITCase.batchBroadcastExecution
>  were failing. Upon further investigation, I found that these tests were also 
> flaky and depended on the earlier made change.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-29611) Fix flaky tests in CoBroadcastWithNonKeyedOperatorTest

2022-10-27 Thread Sopan Phaltankar (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29611?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17619810#comment-17619810
 ] 

Sopan Phaltankar edited comment on FLINK-29611 at 10/28/22 4:34 AM:


[~martijnvisser] 
This test depends on the order of iteration in `HashMap.entrySet()` and can 
fail for some orders. I am running this on my local machine. I have used a 
maven plugin NonDex, which can be used to identify such tests. One can 
reproduce it with the command `mvn 
edu.illinois:nondex-maven-plugun:1.1.2:nondex 
-Dtest=org.apache.flink.streaming.api.operators.co.CoBroadcastWithNonKeyedOperatorTest#testMultiStateSupport`.
 Even if the test was not failing during the daily jobs, it'd be good to not 
depend on the `HashMap.entrySet()` that gives back results in an undefined 
manner 
[JavaDoc|https://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html#entrySet--]for
 this. Therefore, to remove non-determinism completely, we can do this change.


was (Author: JIRAUSER296865):
[~martijnvisser] 
This test depends on the order of iteration in `HashMap.entrySet()` and can 
fail for some orders. I am running this on my local machine. I have used a 
maven plugin NonDex, which can be used to identify such tests. One can 
reproduce it with the command `mvn 
edu.illinois:nondex-maven-plugun:1.1.2:nondex 
-Dtest=org.apache.flink.streaming.api.operators.co.CoBroadcastWithNonKeyedOperatorTest#testMultiStateSupport`.
 Even if the test was not failing during the daily jobs, it'd be good to not 
depend on the `HashMap.entrySet()` that gives back results in an undefined 
manner 
[JavaDoc|[https://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html#entrySet--]]
 for this. Therefore, to remove non-determinism completely, we can do this 
change.

> Fix flaky tests in CoBroadcastWithNonKeyedOperatorTest
> --
>
> Key: FLINK-29611
> URL: https://issues.apache.org/jira/browse/FLINK-29611
> Project: Flink
>  Issue Type: Bug
>Reporter: Sopan Phaltankar
>Priority: Minor
>  Labels: pull-request-available
>
> The test 
> _org.apache.flink.streaming.api.operators.co.CoBroadcastWithNonKeyedOperatorTest.testMultiStateSupport_
>  has the following failure:
> Failures:
> [ERROR]   CoBroadcastWithNonKeyedOperatorTest.testMultiStateSupport:74 
> Wrong Side Output: arrays first differed at element [0]; expected: 15 : 9:key.6->6> but was:5>
> I used the tool [NonDex|https://github.com/TestingResearchIllinois/NonDex] to 
> find this flaky test. 
> Command: mvn edu.illinois:nondex-maven-plugun:1.1.2:nondex -Dtest='Fully 
> Qualified Test Name'
> I analyzed the assertion failure and found that the root cause is because the 
> test method calls ctx.getBroadcastState(STATE_DESCRIPTOR).immutableEntries() 
> which calls the entrySet() method of the underlying HashMap. entrySet() 
> returns the entries in a non-deterministic way, causing the test to be flaky. 
> The fix would be to change _HashMap_ to _LinkedHashMap_ where the Map is 
> getting initialized.
> On further analysis, it was found that the Map is getting initialized on line 
> 53 of org.apache.flink.runtime.state.HeapBroadcastState class.
> After changing from HashMap to LinkedHashMap, the above test is passing.
> Edit: Upon making this change and running the CI, it was found that the tests 
> org.apache.flink.api.datastream.DataStreamBatchExecutionITCase.batchKeyedBroadcastExecution
>  and 
> org.apache.flink.api.datastream.DataStreamBatchExecutionITCase.batchBroadcastExecution
>  were failing. Upon further investigation, I found that these tests were also 
> flaky and depended on the earlier made change.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] reswqa commented on a diff in pull request #21137: [FLINK-29234][runtime] JobMasterServiceLeadershipRunner handle leader event in a separate executor to avoid dead lock

2022-10-27 Thread GitBox


reswqa commented on code in PR #21137:
URL: https://github.com/apache/flink/pull/21137#discussion_r100709


##
flink-runtime/src/test/java/org/apache/flink/runtime/jobmaster/JobMasterServiceLeadershipRunnerTest.java:
##
@@ -683,21 +665,66 @@ public void testJobAlreadyDone() throws Exception {
 jobManagerRunner.getResultFuture();
 
 JobManagerRunnerResult result = resultFuture.get();
-assertEquals(
-JobStatus.FAILED,
-
result.getExecutionGraphInfo().getArchivedExecutionGraph().getState());
+
assertThat(result.getExecutionGraphInfo().getArchivedExecutionGraph().getState())
+.isEqualTo(JobStatus.FAILED);
 }
 }
 
+@Test
+void 
testJobMasterServiceLeadershipRunnerCloseWhenElectionServiceGrantLeaderShip()
+throws Exception {
+final TestingLeaderElectionDriver.TestingLeaderElectionDriverFactory
+testingLeaderElectionDriverFactory =
+new 
TestingLeaderElectionDriver.TestingLeaderElectionDriverFactory();
+final LeaderElectionService defaultLeaderElectionService =
+new 
DefaultLeaderElectionService(testingLeaderElectionDriverFactory);
+
+final JobMasterServiceLeadershipRunner jobManagerRunner =
+newJobMasterServiceLeadershipRunnerBuilder()
+.setJobMasterServiceProcessFactory(
+
TestingJobMasterServiceProcessFactory.newBuilder().build())
+.setLeaderElectionService(defaultLeaderElectionService)
+.build();
+
+jobManagerRunner.start();
+final TestingLeaderElectionDriver currentLeaderDriver =
+Preconditions.checkNotNull(
+
testingLeaderElectionDriverFactory.getCurrentLeaderDriver());
+
+final CheckedThread contenderCloseThread =
+new CheckedThread() {
+@Override
+public void go() {
+try {
+jobManagerRunner.close();
+} catch (Exception e) {
+throw new RuntimeException(e);
+}
+}
+};
+contenderCloseThread.start();
+
+// grant leadership.
+currentLeaderDriver.isLeader();

Review Comment:
   @XComp Thank you for your advice. 
   At the beginning, I didn't think of a good way to control the completion of 
completableFuture, and the probability of deadlock reproduce in my local 
environment is very high. But your suggestion did find a good control injection 
point, so i decided to test it in this way. 
   But if we want to reproduce the problem 100%, it seems that we need another 
control point. Maybe `TestingLeaderElectionDriver#isLeader` is a suitable 
choice. What's your opinion?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-29788) StatefulJobWBroadcastStateMigrationITCase failed in native savepoints

2022-10-27 Thread Hangxiang Yu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29788?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625417#comment-17625417
 ] 

Hangxiang Yu commented on FLINK-29788:
--

Sure. It's known limitation that changelog hasn't supported native savepoint in 
1.16.
So I think we could just disable changelog when triggering native savepoint in 
this case.
It will not block the release of 1.16.
BTW, IIUC, Should the migration it case be executed before we decide to release 
1.16 ? [~hxbks2ks] 

   

> StatefulJobWBroadcastStateMigrationITCase failed in native savepoints
> -
>
> Key: FLINK-29788
> URL: https://issues.apache.org/jira/browse/FLINK-29788
> Project: Flink
>  Issue Type: Bug
>  Components: Release System, Runtime / State Backends
>Affects Versions: 1.16.0
>Reporter: Huang Xingbo
>Priority: Critical
> Attachments: image-2022-10-28-11-18-45-471.png
>
>
>  !image-2022-10-28-11-18-45-471.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29565) In Flink per job mode, the logs printed by taskManager on the web UI will not be highlighted, because the log contents are annotated due to special symbols, which will

2022-10-27 Thread wangshiwei (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625409#comment-17625409
 ] 

wangshiwei commented on FLINK-29565:


yes 

> In Flink per job mode, the logs printed by taskManager on the web UI will not 
> be highlighted, because the log contents are annotated due to special 
> symbols, which will affect the use experience. For more information, see Fig
> 
>
> Key: FLINK-29565
> URL: https://issues.apache.org/jira/browse/FLINK-29565
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Web Frontend
>Affects Versions: 1.14.3
>Reporter: wangshiwei
>Priority: Minor
> Attachments: image-2022-10-10-18-40-27-721.png, 
> image-2022-10-10-18-43-31-897.png, image-2022-10-10-18-43-53-713.png, 
> image-2022-10-10-18-45-17-228.png, image-2022-10-10-19-02-29-796.png, 
> image-2022-10-10-19-03-27-670.png
>
>
>  
> !image-2022-10-10-19-03-27-670.png|width=580,height=317!
> !image-2022-10-10-18-43-53-713.png|width=726,height=47!
> This kind of '/*' content will appear in the print log of the 
> logEnvironmentInfo method in the EnvironmentInformation class. The following 
> logs will be commented out without highlighting
> *verification*
> !image-2022-10-10-18-45-17-228.png|width=880,height=161!
> After manually printing '*/' in the business code, the log is normal
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] liyubin117 commented on pull request #21182: [FLINK-29679][table] DESCRIBE statement shows column comment

2022-10-27 Thread GitBox


liyubin117 commented on PR #21182:
URL: https://github.com/apache/flink/pull/21182#issuecomment-1294406889

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-ml] zhipeng93 commented on a diff in pull request #162: [FLINK-29593] Add QuantileSummary to help calculate approximate quant…

2022-10-27 Thread GitBox


zhipeng93 commented on code in PR #162:
URL: https://github.com/apache/flink-ml/pull/162#discussion_r1007572419


##
flink-ml-lib/src/main/java/org/apache/flink/ml/common/util/QuantileSummary.java:
##
@@ -0,0 +1,400 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.common.util;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.util.Preconditions;
+
+import java.io.Serializable;
+import java.util.ArrayList;
+import java.util.Arrays;
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.LinkedList;
+import java.util.List;
+import java.util.Map;
+import java.util.stream.Collectors;
+import java.util.stream.IntStream;
+
+/**
+ * Helper class to compute an approximate quantile summary. This 
implementation is based on the
+ * algorithm proposed in the paper: "Space-efficient Online Computation of 
Quantile Summaries" by
+ * Greenwald, Michael and Khanna, Sanjeev. 
(https://doi.org/10.1145/375663.375670)
+ */
+@Internal
+public class QuantileSummary implements Serializable {
+
+/** The target relative error. */
+private final double relativeError;
+
+/**
+ * The compression threshold. After the internal buffer of statistics 
crosses this size, it
+ * attempts to compress the statistics together.
+ */
+private final int compressThreshold;
+
+/** The count of all the elements inserted to be calculated. */
+private final long count;
+
+/** A buffer of quantile statistics. */
+private final List sampled;
+
+/** The default size of head buffer. */
+private static final int DEFAULT_HEAD_SIZE = 5;
+
+/** The default compression threshold. */
+private static final int DEFAULT_COMPRESS_THRESHOLD = 1;
+
+/** A buffer of the latest samples seen so far. */
+private List headBuffer = new ArrayList<>(DEFAULT_HEAD_SIZE);

Review Comment:
   nit: Do you think making it a `double[]` could be more memory efficient? I 
am also fine the keep the current implementation.



##
flink-ml-lib/src/main/java/org/apache/flink/ml/common/util/QuantileSummary.java:
##
@@ -0,0 +1,400 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.common.util;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.util.Preconditions;
+
+import java.io.Serializable;
+import java.util.ArrayList;
+import java.util.Arrays;
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.LinkedList;
+import java.util.List;
+import java.util.Map;
+import java.util.stream.Collectors;
+import java.util.stream.IntStream;
+
+/**
+ * Helper class to compute an approximate quantile summary. This 
implementation is based on the
+ * algorithm proposed in the paper: "Space-efficient Online Computation of 
Quantile Summaries" by
+ * Greenwald, Michael and Khanna, Sanjeev. 
(https://doi.org/10.1145/375663.375670)
+ */
+@Internal
+public class QuantileSummary implements Serializable {
+
+/** The target relative error. */
+private final double relativeError;
+
+/**
+ * The compression threshold. After the internal buffer of statistics 
crosses this size, it
+ * attempts to compress the statistics together.
+ */
+private final int compressThreshold;
+
+/** The count of all the elements inserted to be calculated. */
+private final long count;
+
+/** A buffer of quantile 

[jira] [Updated] (FLINK-29788) StatefulJobWBroadcastStateMigrationITCase failed in native savepoints

2022-10-27 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29788?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29788:
-
Priority: Critical  (was: Major)

> StatefulJobWBroadcastStateMigrationITCase failed in native savepoints
> -
>
> Key: FLINK-29788
> URL: https://issues.apache.org/jira/browse/FLINK-29788
> Project: Flink
>  Issue Type: Bug
>  Components: Release System, Runtime / State Backends
>Affects Versions: 1.16.0
>Reporter: Huang Xingbo
>Priority: Critical
> Attachments: image-2022-10-28-11-18-45-471.png
>
>
>  !image-2022-10-28-11-18-45-471.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29788) StatefulJobWBroadcastStateMigrationITCase failed in native savepoints

2022-10-27 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29788?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625404#comment-17625404
 ] 

Huang Xingbo commented on FLINK-29788:
--

[~masteryhx] Could you help take a look? Thx.

> StatefulJobWBroadcastStateMigrationITCase failed in native savepoints
> -
>
> Key: FLINK-29788
> URL: https://issues.apache.org/jira/browse/FLINK-29788
> Project: Flink
>  Issue Type: Bug
>  Components: Release System, Runtime / State Backends
>Affects Versions: 1.16.0
>Reporter: Huang Xingbo
>Priority: Major
> Attachments: image-2022-10-28-11-18-45-471.png
>
>
>  !image-2022-10-28-11-18-45-471.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] liyubin117 commented on pull request #21182: [FLINK-29679][table] DESCRIBE statement shows column comment

2022-10-27 Thread GitBox


liyubin117 commented on PR #21182:
URL: https://github.com/apache/flink/pull/21182#issuecomment-1294397909

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-29788) StatefulJobWBroadcastStateMigrationITCase failed in native savepoints

2022-10-27 Thread Huang Xingbo (Jira)
Huang Xingbo created FLINK-29788:


 Summary: StatefulJobWBroadcastStateMigrationITCase failed in 
native savepoints
 Key: FLINK-29788
 URL: https://issues.apache.org/jira/browse/FLINK-29788
 Project: Flink
  Issue Type: Bug
  Components: Release System, Runtime / State Backends
Affects Versions: 1.16.0
Reporter: Huang Xingbo
 Attachments: image-2022-10-28-11-18-45-471.png

 !image-2022-10-28-11-18-45-471.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] xintongsong commented on pull request #21122: [FLINK-28889] Hybrid shuffle supports multiple consumer and broadcast optimization

2022-10-27 Thread GitBox


xintongsong commented on PR #21122:
URL: https://github.com/apache/flink/pull/21122#issuecomment-1294395156

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-28863) Snapshot result of RocksDB native savepoint should have empty shared-state

2022-10-27 Thread Jinzhong Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-28863?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625402#comment-17625402
 ] 

Jinzhong Li commented on FLINK-28863:
-

In my opinion, for Rocksdb native savepoint SnapshotResult, we should put 
[sstFiles|https://github.com/apache/flink/blob/bb9f2525e6e16d00ef2f0739d9cb96c2e47e35e7/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/snapshot/RocksIncrementalSnapshotStrategy.java#L302]
 into privateState Map of IncrementalRemoteKeyedStateHandle to fix this issue.

This change has no effect on restore, which [downloads both the priavateStates 
and 
shareStates|[https://github.com/apache/flink/blob/35c5f674041bcefea93e1de459cea0d1789f98e0/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBStateDownloader.java#L53].]

 

[~yunta] WDYT?  If my understanding is correct, I can fix the issue.

> Snapshot result of RocksDB native savepoint should have empty shared-state
> --
>
> Key: FLINK-28863
> URL: https://issues.apache.org/jira/browse/FLINK-28863
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing, Runtime / State Backends
>Reporter: Yun Tang
>Priority: Major
> Fix For: 1.17.0, 1.15.3, 1.16.1
>
>
> The current snapshot result of RocksDB native savepoint has non-empty shared 
> state, which is obviously not correct as all snapshot artifacts already stay 
> in the exclusive checkpoint scope folder.
> This does not bring real harmful result due to we would not register the 
> snapshot results of RocksDB native savepoint.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29787) fix ci METHOD_NEW_DEFAULT issue

2022-10-27 Thread Yubin Li (Jira)
Yubin Li created FLINK-29787:


 Summary: fix ci METHOD_NEW_DEFAULT issue
 Key: FLINK-29787
 URL: https://issues.apache.org/jira/browse/FLINK-29787
 Project: Flink
  Issue Type: Bug
  Components: Build System / CI
Reporter: Yubin Li


`org.apache.flink.api.connector.source.SourceReader` declared a new default 
function `pauseOrResumeSplits()`, japicmp plugin failed during ci running, we 
should configure the plugin to make it compatible.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] Myasuka merged pull request #20917: [hotfix][doc] Refine code examples in data_stream_api

2022-10-27 Thread GitBox


Myasuka merged PR #20917:
URL: https://github.com/apache/flink/pull/20917


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-28973) Extending /jars/:jarid/plan API to support setting Flink configs

2022-10-27 Thread Zhanghao Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-28973?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625399#comment-17625399
 ] 

Zhanghao Chen commented on FLINK-28973:
---

Hi [~gaoyunhaii]. The concern is totally valid, but as have been discussed in 
[FLIP-256: Support Job Dynamic Parameter With Flink Rest Api - Apache Flink - 
Apache Software 
Foundation|https://cwiki.apache.org/confluence/display/FLINK/FLIP-256%3A+Support+Job+Dynamic+Parameter+With+Flink+Rest+Api],
 a pragmatic approach would just be allow to take both types of config options 
as input, and simply ignore the cluster-level ones, which is also consistent 
with how we handle it when submitting jobs.

> Extending /jars/:jarid/plan API to support setting Flink configs
> 
>
> Key: FLINK-28973
> URL: https://issues.apache.org/jira/browse/FLINK-28973
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / REST
>Reporter: Zhanghao Chen
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29565) In Flink per job mode, the logs printed by taskManager on the web UI will not be highlighted, because the log contents are annotated due to special symbols, which will

2022-10-27 Thread Yun Gao (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625396#comment-17625396
 ] 

Yun Gao commented on FLINK-29565:
-

Hi [~1336399775] then would you like to open a PR to fix this issue? 

> In Flink per job mode, the logs printed by taskManager on the web UI will not 
> be highlighted, because the log contents are annotated due to special 
> symbols, which will affect the use experience. For more information, see Fig
> 
>
> Key: FLINK-29565
> URL: https://issues.apache.org/jira/browse/FLINK-29565
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Web Frontend
>Affects Versions: 1.14.3
>Reporter: wangshiwei
>Priority: Minor
> Attachments: image-2022-10-10-18-40-27-721.png, 
> image-2022-10-10-18-43-31-897.png, image-2022-10-10-18-43-53-713.png, 
> image-2022-10-10-18-45-17-228.png, image-2022-10-10-19-02-29-796.png, 
> image-2022-10-10-19-03-27-670.png
>
>
>  
> !image-2022-10-10-19-03-27-670.png|width=580,height=317!
> !image-2022-10-10-18-43-53-713.png|width=726,height=47!
> This kind of '/*' content will appear in the print log of the 
> logEnvironmentInfo method in the EnvironmentInformation class. The following 
> logs will be commented out without highlighting
> *verification*
> !image-2022-10-10-18-45-17-228.png|width=880,height=161!
> After manually printing '*/' in the business code, the log is normal
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] HuangXingBo commented on pull request #20859: [release] release notes for the 1.16 release

2022-10-27 Thread GitBox


HuangXingBo commented on PR #20859:
URL: https://github.com/apache/flink/pull/20859#issuecomment-1294382816

   @zentol Ok. I have addressed the comment at the latest commit. :)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] flinkbot commented on pull request #21182: [FLINK-29679][table] DESCRIBE statement shows column comment

2022-10-27 Thread GitBox


flinkbot commented on PR #21182:
URL: https://github.com/apache/flink/pull/21182#issuecomment-1294381431

   
   ## CI report:
   
   * 0a7afa1d3cb5370c44b1bea729afa6577ae7d3c9 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-29565) In Flink per job mode, the logs printed by taskManager on the web UI will not be highlighted, because the log contents are annotated due to special symbols, which will

2022-10-27 Thread wangshiwei (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625395#comment-17625395
 ] 

wangshiwei commented on FLINK-29565:


yes, I have

> In Flink per job mode, the logs printed by taskManager on the web UI will not 
> be highlighted, because the log contents are annotated due to special 
> symbols, which will affect the use experience. For more information, see Fig
> 
>
> Key: FLINK-29565
> URL: https://issues.apache.org/jira/browse/FLINK-29565
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Web Frontend
>Affects Versions: 1.14.3
>Reporter: wangshiwei
>Priority: Minor
> Attachments: image-2022-10-10-18-40-27-721.png, 
> image-2022-10-10-18-43-31-897.png, image-2022-10-10-18-43-53-713.png, 
> image-2022-10-10-18-45-17-228.png, image-2022-10-10-19-02-29-796.png, 
> image-2022-10-10-19-03-27-670.png
>
>
>  
> !image-2022-10-10-19-03-27-670.png|width=580,height=317!
> !image-2022-10-10-18-43-53-713.png|width=726,height=47!
> This kind of '/*' content will appear in the print log of the 
> logEnvironmentInfo method in the EnvironmentInformation class. The following 
> logs will be commented out without highlighting
> *verification*
> !image-2022-10-10-18-45-17-228.png|width=880,height=161!
> After manually printing '*/' in the business code, the log is normal
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] reswqa commented on a diff in pull request #21137: [FLINK-29234][runtime] JobMasterServiceLeadershipRunner handle leader event in a separate executor to avoid dead lock

2022-10-27 Thread GitBox


reswqa commented on code in PR #21137:
URL: https://github.com/apache/flink/pull/21137#discussion_r1007555607


##
flink-runtime/src/test/java/org/apache/flink/runtime/jobmaster/JobMasterServiceLeadershipRunnerTest.java:
##
@@ -683,21 +665,66 @@ public void testJobAlreadyDone() throws Exception {
 jobManagerRunner.getResultFuture();
 
 JobManagerRunnerResult result = resultFuture.get();
-assertEquals(
-JobStatus.FAILED,
-
result.getExecutionGraphInfo().getArchivedExecutionGraph().getState());
+
assertThat(result.getExecutionGraphInfo().getArchivedExecutionGraph().getState())
+.isEqualTo(JobStatus.FAILED);
 }
 }
 
+@Test
+void 
testJobMasterServiceLeadershipRunnerCloseWhenElectionServiceGrantLeaderShip()
+throws Exception {
+final TestingLeaderElectionDriver.TestingLeaderElectionDriverFactory
+testingLeaderElectionDriverFactory =
+new 
TestingLeaderElectionDriver.TestingLeaderElectionDriverFactory();
+final LeaderElectionService defaultLeaderElectionService =
+new 
DefaultLeaderElectionService(testingLeaderElectionDriverFactory);
+
+final JobMasterServiceLeadershipRunner jobManagerRunner =
+newJobMasterServiceLeadershipRunnerBuilder()
+.setJobMasterServiceProcessFactory(
+
TestingJobMasterServiceProcessFactory.newBuilder().build())
+.setLeaderElectionService(defaultLeaderElectionService)
+.build();
+
+jobManagerRunner.start();
+final TestingLeaderElectionDriver currentLeaderDriver =
+Preconditions.checkNotNull(
+
testingLeaderElectionDriverFactory.getCurrentLeaderDriver());
+
+final CheckedThread contenderCloseThread =
+new CheckedThread() {
+@Override
+public void go() {
+try {
+jobManagerRunner.close();
+} catch (Exception e) {
+throw new RuntimeException(e);
+}
+}
+};
+contenderCloseThread.start();
+
+// grant leadership.
+currentLeaderDriver.isLeader();

Review Comment:
   @XComp Thank you for your advice.
   At the beginning, I didn't think of a good way to control the completion of 
completableFuture, and the probability of deadlock reproduce in my local 
environment is very high. But your suggestion did find a good control injection 
point, so i decided to test it in this way.
   But if we want to reproduce the problem 100%, it seems that we need another 
control point. Maybe `TestingLeaderElectionDriver#isLeader` is a suitable 
choice. What's your opinion?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] reswqa commented on a diff in pull request #21137: [FLINK-29234][runtime] JobMasterServiceLeadershipRunner handle leader event in a separate executor to avoid dead lock

2022-10-27 Thread GitBox


reswqa commented on code in PR #21137:
URL: https://github.com/apache/flink/pull/21137#discussion_r1007555607


##
flink-runtime/src/test/java/org/apache/flink/runtime/jobmaster/JobMasterServiceLeadershipRunnerTest.java:
##
@@ -683,21 +665,66 @@ public void testJobAlreadyDone() throws Exception {
 jobManagerRunner.getResultFuture();
 
 JobManagerRunnerResult result = resultFuture.get();
-assertEquals(
-JobStatus.FAILED,
-
result.getExecutionGraphInfo().getArchivedExecutionGraph().getState());
+
assertThat(result.getExecutionGraphInfo().getArchivedExecutionGraph().getState())
+.isEqualTo(JobStatus.FAILED);
 }
 }
 
+@Test
+void 
testJobMasterServiceLeadershipRunnerCloseWhenElectionServiceGrantLeaderShip()
+throws Exception {
+final TestingLeaderElectionDriver.TestingLeaderElectionDriverFactory
+testingLeaderElectionDriverFactory =
+new 
TestingLeaderElectionDriver.TestingLeaderElectionDriverFactory();
+final LeaderElectionService defaultLeaderElectionService =
+new 
DefaultLeaderElectionService(testingLeaderElectionDriverFactory);
+
+final JobMasterServiceLeadershipRunner jobManagerRunner =
+newJobMasterServiceLeadershipRunnerBuilder()
+.setJobMasterServiceProcessFactory(
+
TestingJobMasterServiceProcessFactory.newBuilder().build())
+.setLeaderElectionService(defaultLeaderElectionService)
+.build();
+
+jobManagerRunner.start();
+final TestingLeaderElectionDriver currentLeaderDriver =
+Preconditions.checkNotNull(
+
testingLeaderElectionDriverFactory.getCurrentLeaderDriver());
+
+final CheckedThread contenderCloseThread =
+new CheckedThread() {
+@Override
+public void go() {
+try {
+jobManagerRunner.close();
+} catch (Exception e) {
+throw new RuntimeException(e);
+}
+}
+};
+contenderCloseThread.start();
+
+// grant leadership.
+currentLeaderDriver.isLeader();

Review Comment:
   @XComp Thank you for your advice.
   At the beginning, I didn't think of a good way to control the completion of 
completableFuture, and the probability of deadlock reproduce in my local 
environment is very high. But your suggestion did find a good control injection 
point, so i decided to test it in this way.
   But if we want to reproduce the problem 100%, it seems that we need another 
control point. Maybe TestingLeaderElectionDriver#isLeader is a suitable choice. 
What's your opinion?



##
flink-runtime/src/test/java/org/apache/flink/runtime/jobmaster/JobMasterServiceLeadershipRunnerTest.java:
##
@@ -683,21 +665,66 @@ public void testJobAlreadyDone() throws Exception {
 jobManagerRunner.getResultFuture();
 
 JobManagerRunnerResult result = resultFuture.get();
-assertEquals(
-JobStatus.FAILED,
-
result.getExecutionGraphInfo().getArchivedExecutionGraph().getState());
+
assertThat(result.getExecutionGraphInfo().getArchivedExecutionGraph().getState())
+.isEqualTo(JobStatus.FAILED);
 }
 }
 
+@Test
+void 
testJobMasterServiceLeadershipRunnerCloseWhenElectionServiceGrantLeaderShip()
+throws Exception {
+final TestingLeaderElectionDriver.TestingLeaderElectionDriverFactory
+testingLeaderElectionDriverFactory =
+new 
TestingLeaderElectionDriver.TestingLeaderElectionDriverFactory();
+final LeaderElectionService defaultLeaderElectionService =
+new 
DefaultLeaderElectionService(testingLeaderElectionDriverFactory);
+
+final JobMasterServiceLeadershipRunner jobManagerRunner =
+newJobMasterServiceLeadershipRunnerBuilder()
+.setJobMasterServiceProcessFactory(
+
TestingJobMasterServiceProcessFactory.newBuilder().build())
+.setLeaderElectionService(defaultLeaderElectionService)
+.build();
+
+jobManagerRunner.start();
+final TestingLeaderElectionDriver currentLeaderDriver =
+Preconditions.checkNotNull(
+
testingLeaderElectionDriverFactory.getCurrentLeaderDriver());
+
+final CheckedThread contenderCloseThread =
+new CheckedThread() {
+@Override
+public void go() {
+try {
+  

[GitHub] [flink] liyubin117 closed pull request #21116: [FLINK-29679][table] DESCRIBE statement shows column comment

2022-10-27 Thread GitBox


liyubin117 closed pull request #21116: [FLINK-29679][table] DESCRIBE statement 
shows column comment
URL: https://github.com/apache/flink/pull/21116


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] liyubin117 opened a new pull request, #21182: [FLINK-29679][table] DESCRIBE statement shows column comment

2022-10-27 Thread GitBox


liyubin117 opened a new pull request, #21182:
URL: https://github.com/apache/flink/pull/21182

   ## What is the purpose of the change
   
   DESCRIBE statement print column comment to make results user-friendly, such 
as help analysts write sql adaptive to corresponding business logics.
   
   ## Brief change log
   
   * use new schema framework to get comment.
   
   ## Verifying this change
   
   This change added tests and can be verified as follows:
   
   flink-table/flink-sql-client/src/test/resources/sql/table.q
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? yes
 - If yes, how is the feature documented? yes


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-29765) SQL query not executing properly

2022-10-27 Thread Shengkai Fang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29765?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625391#comment-17625391
 ] 

Shengkai Fang commented on FLINK-29765:
---

It's better you can share more infos, e.g. source table schema, json plan?

> SQL query not executing properly
> 
>
> Key: FLINK-29765
> URL: https://issues.apache.org/jira/browse/FLINK-29765
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.15.0
>Reporter: Aqib Mehmood
>Priority: Major
>
> I am using this query to compare difference between last and second last 
> price of our order sku
> WITH CTE AS (
> SELECT a.sku, a.name, a.updatedAt, b.price FROM (
> SELECT sku, name, max(updatedAt) AS updatedAt from (
> SELECT sku, name, updatedAt FROM wms.PurchaseOrderProduct
> WHERE CONCAT(sku, DATE_FORMAT(updatedAt, '%Y-%m-%d %H:%m:%s')) 
> not in (
> SELECT CONCAT(sku, DATE_FORMAT(updatedAt, '%Y-%m-%d 
> %H:%m:%s')) FROM (
> SELECT sku, max(updatedAt) as updatedAt from 
> wms.PurchaseOrderProduct
> GROUP BY sku
> ) AS x
> )
> ) AS z
> GROUP BY sku, name
> ) AS a
> LEFT JOIN wms.PurchaseOrderProduct b
> ON a.sku=b.sku AND a.name=b.name and a.updatedAt=b.updatedAt
> )
> SELECT a.sku, a.name, a.updatedAt AS latestupdatedAt, a.price AS latestPrice, 
> b.updatedAt AS lastUpdatedAt, b.price AS lastPrice
> FROM (
> SELECT a.sku, a.name, a.updatedAt, b.price from (
> SELECT sku, name, max(updatedAt) as updatedAt from 
> wms.PurchaseOrderProduct
> GROUP BY sku, name
> ) AS a
> LEFT JOIN wms.PurchaseOrderProduct b
> ON a.sku=b.sku AND a.name=b.name and a.updatedAt=b.updatedAt
> ) AS a
> LEFT JOIN CTE AS b
> ON a.sku=b.sku AND a.name=b.name;
> This issue is that Im getting *NULLs* for columns *lastUpdatedAt* and 
> {*}lastPrice{*}. But when I run the same query on our prod database, I'm 
> getting desired results.I suspect that flink is not processing the entire 
> query before giving the results.
> I get desired results for a couple of rows in while *lastUpdatedAt* and 
> *lastPrice* are not *NULL* in the beginning of the table{*}.{*} But then 
> after that the entire two columns return *NULLs*
> I would like to know why flink is not executing the above query properly?TIA



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29748) Expose the optimize phase in the connector context

2022-10-27 Thread Shengkai Fang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29748?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625388#comment-17625388
 ] 

Shengkai Fang commented on FLINK-29748:
---

I think you can validate  when StreamExecSource invokes getScanRuntimeProvider. 
When execnode -> transformat, it means the optimization finishes.

> Expose the optimize phase in the connector context
> --
>
> Key: FLINK-29748
> URL: https://issues.apache.org/jira/browse/FLINK-29748
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner, Table SQL / Runtime
>Reporter: Aitozi
>Priority: Minor
>
> Currently, in the connector it can not know whether the whole optimize is 
> finished.
> When the optimize finished, the all information is static, eg: the reading 
> partitions. If I want to validate the final optimized result (like whether 
> the reading partition is too much or empty), it needs the context of what is 
> the current phase. I think the {{ScanContext}} is ok to expose this 
> information. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] yunfengzhou-hub commented on a diff in pull request #160: [FLINK-29434] Add AlgoOperator for Splitter

2022-10-27 Thread GitBox


yunfengzhou-hub commented on code in PR #160:
URL: https://github.com/apache/flink-ml/pull/160#discussion_r1007543761


##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/randomsplitter/RandomSplitterParams.java:
##
@@ -0,0 +1,66 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.feature.randomsplitter;
+
+import org.apache.flink.ml.param.DoubleArrayParam;
+import org.apache.flink.ml.param.Param;
+import org.apache.flink.ml.param.ParamValidator;
+import org.apache.flink.ml.param.WithParams;
+
+import java.util.Arrays;
+import java.util.HashSet;
+import java.util.Set;
+
+/**
+ * Params of {@link RandomSplitter}.
+ *
+ * @param  The class type of this instance.
+ */
+public interface RandomSplitterParams extends WithParams {
+Param WEIGHTS =
+new DoubleArrayParam(
+"weights", "The weights of data splitting.", null, 
weightsValidator());

Review Comment:
   Weight is different from fraction. Weight means the proportion of the number 
of elements in each output split table,  instead of a possibility threshold 
like the fraction in the previous implementation.
   
   `setWeight(1,1)` is equal to `setThreshold(0.5)`, `setWeight(1,2,2)` is 
equal to `setThreshold(0.2,0.6)`.



##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/randomsplitter/RandomSplitter.java:
##
@@ -0,0 +1,127 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.feature.randomsplitter;
+
+import org.apache.flink.api.java.typeutils.RowTypeInfo;
+import org.apache.flink.ml.api.AlgoOperator;
+import org.apache.flink.ml.common.datastream.TableUtils;
+import org.apache.flink.ml.param.Param;
+import org.apache.flink.ml.util.ParamUtils;
+import org.apache.flink.ml.util.ReadWriteUtils;
+import org.apache.flink.streaming.api.datastream.DataStream;
+import org.apache.flink.streaming.api.datastream.SingleOutputStreamOperator;
+import org.apache.flink.streaming.api.operators.AbstractStreamOperator;
+import org.apache.flink.streaming.api.operators.OneInputStreamOperator;
+import org.apache.flink.streaming.runtime.streamrecord.StreamRecord;
+import org.apache.flink.table.api.Table;
+import org.apache.flink.table.api.bridge.java.StreamTableEnvironment;
+import org.apache.flink.table.api.internal.TableImpl;
+import org.apache.flink.types.Row;
+import org.apache.flink.util.OutputTag;
+import org.apache.flink.util.Preconditions;
+
+import java.io.IOException;
+import java.util.Arrays;
+import java.util.HashMap;
+import java.util.Map;
+import java.util.Random;
+
+/** An AlgoOperator which splits a datastream into N datastreams according to 
the given weights. */
+public class RandomSplitter
+implements AlgoOperator, 
RandomSplitterParams {
+private final Map, Object> paramMap = new HashMap<>();
+
+public RandomSplitter() {
+ParamUtils.initializeMapWithDefaultValues(paramMap, this);
+}
+
+@Override
+public Table[] transform(Table... inputs) {
+Preconditions.checkArgument(inputs.length == 1);
+StreamTableEnvironment tEnv =
+(StreamTableEnvironment) ((TableImpl) 
inputs[0]).getTableEnvironment();
+RowTypeInfo outputTypeInfo = 
TableUtils.getRowTypeInfo(inputs[0].getResolvedSchema());
+
+final Double[] weights = getWeights();
+OutputTag[] outputTags = new OutputTag[weights.length];
+for (int i = 0; i < 

[jira] [Updated] (FLINK-29786) VarianceThresholdSelector Uses FeaturesCol Incorrectly

2022-10-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29786?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-29786:
---
Labels: pull-request-available  (was: )

> VarianceThresholdSelector Uses FeaturesCol Incorrectly
> --
>
> Key: FLINK-29786
> URL: https://issues.apache.org/jira/browse/FLINK-29786
> Project: Flink
>  Issue Type: Bug
>  Components: Library / Machine Learning
>Reporter: Jiang Xin
>Priority: Major
>  Labels: pull-request-available
> Fix For: ml-2.2.0
>
>
> Flink ML has both `FeaturesCol` and `InputCol` to be used to set input. The 
> `FeaturesCol` should only be used with `LabelCol`, while `InputCol` can be 
> used more generally.
> The `VarianceThresholdSelector` should use `InputCol` instead of 
> `FeaturesCol`.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] jiangxin369 opened a new pull request, #167: [FLINK-29786] VarianceThresholdSelector Uses FeaturesCol Incorrectly

2022-10-27 Thread GitBox


jiangxin369 opened a new pull request, #167:
URL: https://github.com/apache/flink-ml/pull/167

   
   
   ## What is the purpose of the change
   
   Replace `FeaturesCol` with `InputCol` in VarianceThresholdSelector.
   
   ## Brief change log
   
 - Add Javadoc for `HasFeaturesCol` to distinguish `HasInputCol`
 - Replace `FeaturesCol` with `InputCol` in VarianceThresholdSelector.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] xintongsong commented on pull request #21122: [FLINK-28889] Hybrid shuffle supports multiple consumer and broadcast optimization

2022-10-27 Thread GitBox


xintongsong commented on PR #21122:
URL: https://github.com/apache/flink/pull/21122#issuecomment-1294359915

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] wangyang0918 merged pull request #21179: [BP-1.16][FLINK-29539][Deployment/Kubernetes] dnsPolicy in FlinkPod is not overridable

2022-10-27 Thread GitBox


wangyang0918 merged PR #21179:
URL: https://github.com/apache/flink/pull/21179


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] wangyang0918 merged pull request #20982: [FLINK-29539] [Deployment/Kubernetes] dnsPolicy in FlinkPod is not overridable

2022-10-27 Thread GitBox


wangyang0918 merged PR #20982:
URL: https://github.com/apache/flink/pull/20982


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-29572) Flink Task Manager skip loopback interface for resource manager registration

2022-10-27 Thread Xintong Song (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29572?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625383#comment-17625383
 ] 

Xintong Song edited comment on FLINK-29572 at 10/28/22 2:15 AM:


You don't have to configure different ports for each task manager. You just 
need to remove `taskmanager.rpc.port` from your configuration, and Flink by 
default should use random ports.

IMHO, allowing components from different hosts to bi-directionally communicate 
with each other via the loopback address does not sounds like a common use 
case. It seems to be a violation of the loopback address semantics. This is 
probably more a problem of the proxy software, rather than Flink.

I'd love to help you find a workaround to get things work if possible. However, 
as one of the maintainers of the Flink project, I'd be negative to introduce 
such a knob for such an unusual use case. 


was (Author: xintongsong):
You don't have to configure different ports for each task manager. You just 
need to remove `taskmanager.rpc.port` from your configuration, and Flink by 
default should use random ports.

IMHO, allowing components from different hosts to bi-directionally communicate 
with each other via the loopback address does not sounds like a common use 
case. It seems to be a violation of the loopback address semantics. This is 
probably more a problem of the proxy software, rather than Flink.

I'd love to help you find a workaround to get things work. However, as one of 
the maintainers of the Flink project, I'd be negative to introduce such a knob 
for such an unusual use case. 

> Flink Task Manager skip loopback interface for resource manager registration
> 
>
> Key: FLINK-29572
> URL: https://issues.apache.org/jira/browse/FLINK-29572
> Project: Flink
>  Issue Type: Bug
>  Components: API / Core
>Affects Versions: 1.15.2
> Environment: Flink 1.15.2
> Kubernetes with Istio Proxy
>Reporter: Kevin Li
>Priority: Major
>
> Currently Flink Task Manager use different local interface to bind to connect 
> to Resource Manager. First one is Loopback interface. Normally if Job Manager 
> is running on remote host/container, using loopback interface to connect will 
> fail and it will pick up correct IP address.
> However, if Task Manager is running with some proxy, loopback interface can 
> connect to remote host as well. This will result 127.0.0.1 reported to 
> Resource Manager during registration, even Job Manager/Resource Manager runs 
> on remote host, and problem will happen. For us, only one Task Manager can 
> register in this case.
> I suggest adding configuration to skip Loopback interface check if we know 
> Job/Resource Manager is running on remote host/container.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29572) Flink Task Manager skip loopback interface for resource manager registration

2022-10-27 Thread Xintong Song (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29572?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625383#comment-17625383
 ] 

Xintong Song commented on FLINK-29572:
--

You don't have to configure different ports for each task manager. You just 
need to remove `taskmanager.rpc.port` from your configuration, and Flink by 
default should use random ports.

IMHO, allowing components from different hosts to bi-directionally communicate 
with each other via the loopback address does not sounds like a common use 
case. It seems to be a violation of the loopback address semantics. This is 
probably more a problem of the proxy software, rather than Flink.

I'd love to help you find a workaround to get things work. However, as one of 
the maintainers of the Flink project, I'd be negative to introduce such a knob 
for such an unusual use case. 

> Flink Task Manager skip loopback interface for resource manager registration
> 
>
> Key: FLINK-29572
> URL: https://issues.apache.org/jira/browse/FLINK-29572
> Project: Flink
>  Issue Type: Bug
>  Components: API / Core
>Affects Versions: 1.15.2
> Environment: Flink 1.15.2
> Kubernetes with Istio Proxy
>Reporter: Kevin Li
>Priority: Major
>
> Currently Flink Task Manager use different local interface to bind to connect 
> to Resource Manager. First one is Loopback interface. Normally if Job Manager 
> is running on remote host/container, using loopback interface to connect will 
> fail and it will pick up correct IP address.
> However, if Task Manager is running with some proxy, loopback interface can 
> connect to remote host as well. This will result 127.0.0.1 reported to 
> Resource Manager during registration, even Job Manager/Resource Manager runs 
> on remote host, and problem will happen. For us, only one Task Manager can 
> register in this case.
> I suggest adding configuration to skip Loopback interface check if we know 
> Job/Resource Manager is running on remote host/container.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] wangyang0918 commented on pull request #21178: [BP-1.15][FLINK-29539][Deployment/Kubernetes] dnsPolicy in FlinkPod is not overridable

2022-10-27 Thread GitBox


wangyang0918 commented on PR #21178:
URL: https://github.com/apache/flink/pull/21178#issuecomment-1294356340

   Fix the compile error and force pushed.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-29786) VarianceThresholdSelector Uses FeaturesCol Incorrectly

2022-10-27 Thread Jiang Xin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29786?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jiang Xin updated FLINK-29786:
--
Description: 
Flink ML has both `FeaturesCol` and `InputCol` to be used to set input. The 
`FeaturesCol` should only be used with `LabelCol`, while `InputCol` can be used 
more generally.

The `VarianceThresholdSelector` should use `InputCol` instead of `FeaturesCol`.

  was:Flink ML has both `FeaturesCol` and `InputCol` to be used to set input. 
The `FeaturesCol` should only be used with `LabelCol`, while `InputCol` can be 
used more generally. The `VarianceThresholdSelector` should use `InputCol` 
instead of `FeaturesCol`.


> VarianceThresholdSelector Uses FeaturesCol Incorrectly
> --
>
> Key: FLINK-29786
> URL: https://issues.apache.org/jira/browse/FLINK-29786
> Project: Flink
>  Issue Type: Bug
>  Components: Library / Machine Learning
>Reporter: Jiang Xin
>Priority: Major
> Fix For: ml-2.2.0
>
>
> Flink ML has both `FeaturesCol` and `InputCol` to be used to set input. The 
> `FeaturesCol` should only be used with `LabelCol`, while `InputCol` can be 
> used more generally.
> The `VarianceThresholdSelector` should use `InputCol` instead of 
> `FeaturesCol`.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29786) VarianceThresholdSelector Uses FeaturesCol Incorrectly

2022-10-27 Thread Jiang Xin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29786?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jiang Xin updated FLINK-29786:
--
Summary: VarianceThresholdSelector Uses FeaturesCol Incorrectly  (was: 
VarianceThresholdSelector Uses FeaturesCol as Input Param)

> VarianceThresholdSelector Uses FeaturesCol Incorrectly
> --
>
> Key: FLINK-29786
> URL: https://issues.apache.org/jira/browse/FLINK-29786
> Project: Flink
>  Issue Type: Bug
>  Components: Library / Machine Learning
>Reporter: Jiang Xin
>Priority: Major
> Fix For: ml-2.2.0
>
>
> Flink ML has both `FeaturesCol` and `InputCol` to be used to set input. The 
> `FeaturesCol` should only be used with `LabelCol`, while `InputCol` can be 
> used more generally. The `VarianceThresholdSelector` should use `InputCol` 
> instead of `FeaturesCol`.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29749) flink info command support dynamic properties

2022-10-27 Thread Yang Wang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29749?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625374#comment-17625374
 ] 

Yang Wang commented on FLINK-29749:
---

Thanks [~jackylau] for your contribution.

> flink info command support dynamic properties
> -
>
> Key: FLINK-29749
> URL: https://issues.apache.org/jira/browse/FLINK-29749
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.17.0
>Reporter: jackylau
>Assignee: jackylau
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0, 1.15.3, 1.16.1
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-29749) flink info command support dynamic properties

2022-10-27 Thread Yang Wang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29749?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Wang closed FLINK-29749.
-
Resolution: Fixed

Fixed via:

master: f8c6a668cd2b887f33a0cf4608de2d6b95c71f03

release-1.16: 38e90428bf7e603fdd353243f1edeba3553af2a3

release-1.15: 1d29f540a0692540a01b951033a8dc04fdb74d4f

> flink info command support dynamic properties
> -
>
> Key: FLINK-29749
> URL: https://issues.apache.org/jira/browse/FLINK-29749
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.17.0
>Reporter: jackylau
>Assignee: jackylau
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0, 1.15.3, 1.16.1
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] wangyang0918 merged pull request #21166: [FLINK-29749][client] Make 'flink info' command could support dynamic…

2022-10-27 Thread GitBox


wangyang0918 merged PR #21166:
URL: https://github.com/apache/flink/pull/21166


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-29786) VarianceThresholdSelector Uses FeaturesCol as Input Param

2022-10-27 Thread Jiang Xin (Jira)
Jiang Xin created FLINK-29786:
-

 Summary: VarianceThresholdSelector Uses FeaturesCol as Input Param
 Key: FLINK-29786
 URL: https://issues.apache.org/jira/browse/FLINK-29786
 Project: Flink
  Issue Type: Bug
  Components: Library / Machine Learning
Reporter: Jiang Xin
 Fix For: ml-2.2.0


Flink ML has both `FeaturesCol` and `InputCol` to be used to set input. The 
`FeaturesCol` should only be used with `LabelCol`, while `InputCol` can be used 
more generally. The `VarianceThresholdSelector` should use `InputCol` instead 
of `FeaturesCol`.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] yunfengzhou-hub commented on a diff in pull request #156: [FLINK-29323] Refine Transformer for VectorAssembler

2022-10-27 Thread GitBox


yunfengzhou-hub commented on code in PR #156:
URL: https://github.com/apache/flink-ml/pull/156#discussion_r1007512230


##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/vectorassembler/VectorAssembler.java:
##
@@ -47,10 +47,16 @@
 
 /**
  * A Transformer which combines a given list of input columns into a vector 
column. Types of input
- * columns must be either vector or numerical value.
+ * columns must be either vector or numerical types. The operator deals with 
null values or records
+ * with wrong sizes according to the strategy specified by the {@link 
HasHandleInvalid} parameter as
+ * follows:
  *
- * The `keep` option of {@link HasHandleInvalid} means that we output bad 
rows with output column
- * set to null.
+ * The `keep` option means that if the input column data is NaN, then it 
keeps this value and if
+ * data is null vector, then uses a NaN vector to replace it.
+ *
+ * The `skip` option means that it filters out rows with invalid elements.
+ *
+ * The `error` option means that it throws an error exception when meeting 
some invalid data.

Review Comment:
   The following tags could better format the JavaDoc.
   ```html
   
AAA
BBB
CCC
   
   ```



##
flink-ml-lib/src/test/java/org/apache/flink/ml/feature/VectorAssemblerTest.java:
##
@@ -45,13 +45,16 @@
 

Review Comment:
   Let's further add test cases to verify different handleInvalid strategies 
when the input vectors are valid (no null or NaN values), but the size of the 
vectors does not match with that set in `inputSizes`.



##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/vectorassembler/VectorAssembler.java:
##
@@ -74,64 +81,96 @@ public Table[] transform(Table... inputs) {
 DataStream output =
 tEnv.toDataStream(inputs[0])
 .flatMap(
-new AssemblerFunc(getInputCols(), 
getHandleInvalid()),
+new AssemblerFunction(
+getInputCols(), getHandleInvalid(), 
getInputSizes()),
 outputTypeInfo);
 Table outputTable = tEnv.fromDataStream(output);
 return new Table[] {outputTable};
 }
 
-private static class AssemblerFunc implements FlatMapFunction {
+private static class AssemblerFunction implements FlatMapFunction {
 private final String[] inputCols;
 private final String handleInvalid;
+private final Integer[] inputSizes;
+private int vectorSize = 0;
+private final boolean keepInvalid;
 
-public AssemblerFunc(String[] inputCols, String handleInvalid) {
+public AssemblerFunction(String[] inputCols, String handleInvalid, 
Integer[] sizeArray) {
 this.inputCols = inputCols;
 this.handleInvalid = handleInvalid;
+this.inputSizes = sizeArray;
+for (Integer inputSize : inputSizes) {
+vectorSize += inputSize;
+}
+keepInvalid = handleInvalid.equals(HasHandleInvalid.KEEP_INVALID);
 }
 
 @Override
 public void flatMap(Row value, Collector out) {
 int nnz = 0;
-int vectorSize = 0;
 try {
-for (String inputCol : inputCols) {
-Object object = value.getField(inputCol);
-Preconditions.checkNotNull(object, "Input column value 
should not be null.");
+for (int i = 0; i < inputCols.length; ++i) {
+Object object = value.getField(inputCols[i]);
+if (object == null) {
+if (keepInvalid) {
+if (inputSizes[i] > 1) {
+DenseVector tmpVec = new 
DenseVector(inputSizes[i]);
+for (int j = 0; j < inputSizes[i]; ++j) {
+tmpVec.values[j] = Double.NaN;
+}
+object = tmpVec;
+} else {
+object = Double.NaN;
+}
+value.setField(inputCols[i], object);
+} else {
+throw new RuntimeException(
+"Input column value is null. Please check 
the input data or using handleInvalid = 'keep'.");
+}
+}
 if (object instanceof Number) {
+if (Double.isNaN(((Number) object).doubleValue()) && 
!keepInvalid) {
+throw new RuntimeException(
+"Encountered NaN while assembling a row 
with handleInvalid = 'error'. Consider "
++ "removing NaNs from dataset or 
using handleInvalid = 'keep' 

[jira] [Created] (FLINK-29785) Upgrade Flink Elasticsearch-7 connector elasticsearch.version to 7.17.0

2022-10-27 Thread Elkhan Dadashov (Jira)
Elkhan Dadashov created FLINK-29785:
---

 Summary: Upgrade Flink Elasticsearch-7 connector 
elasticsearch.version to 7.17.0
 Key: FLINK-29785
 URL: https://issues.apache.org/jira/browse/FLINK-29785
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / ElasticSearch
Reporter: Elkhan Dadashov


`elasticsearch-7` connector uses ElasticSearch `7.10.2` version.

[https://github.com/apache/flink-connector-elasticsearch/blob/main/flink-connector-elasticsearch7/pom.xml#L39]

 

When ElasticSearch server side is `8.X.X` then, `elasticsearch-7` connector 
does not work. For Example for ElasticSearch `8.2.2` version on the server 
side, the minimum required version on the `elasticsearch-7` Flink connector 
side is `"minimum_wire_compatibility_version" : "7.17.0"`.

As of today `elasticsearch-8` does not exist yet. [ Add Elasticsearch 8.0 
support Jira ticket |https://issues.apache.org/jira/browse/FLINK-26088 ] .

With this intermediare step - upgrading `elasticsearch-7` connector 
`elasticsearch.version` to 7.17.0 can help Flink users still ingest into 
ElasticSearch which has `8.X.X` version deployed on the server side. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] ericxiao251 commented on pull request #21077: [FLINK-29498] Add Scala Async Retry Strategies and ResultPredicates Helper Classes

2022-10-27 Thread GitBox


ericxiao251 commented on PR #21077:
URL: https://github.com/apache/flink/pull/21077#issuecomment-1294158906

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-29783) Flaky test: KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime

2022-10-27 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29783?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625337#comment-17625337
 ] 

Martijn Visser edited comment on FLINK-29783 at 10/27/22 9:14 PM:
--

Duplicate of FLINK-24119


was (Author: martijnvisser):
Duplicate of FLINk-24119

> Flaky test: 
> KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime
> -
>
> Key: FLINK-29783
> URL: https://issues.apache.org/jira/browse/FLINK-29783
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.17.0
>Reporter: Gabor Somogyi
>Priority: Major
>  Labels: pull-request-available
>
> {code:java}
> Oct 27 15:07:54 java.lang.AssertionError: Create test topic : 
> partition_failure_recovery_EventTime failed, 
> org.apache.kafka.common.errors.TopicExistsException: Topic 
> 'partition_failure_recovery_EventTime' already exists.
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.KafkaTestEnvironmentImpl.createTestTopic(KafkaTestEnvironmentImpl.java:207)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.KafkaTestEnvironment.createTestTopic(KafkaTestEnvironment.java:97)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.KafkaTestBase.createTestTopic(KafkaTestBase.java:217)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.shuffle.KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecovery(KafkaShuffleExactlyOnceITCase.java:158)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.shuffle.KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime(KafkaShuffleExactlyOnceITCase.java:101)
> Oct 27 15:07:54   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> Oct 27 15:07:54   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> Oct 27 15:07:54   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> Oct 27 15:07:54   at java.lang.reflect.Method.invoke(Method.java:498)
> Oct 27 15:07:54   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> Oct 27 15:07:54   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
> Oct 27 15:07:54   at 
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> Oct 27 15:07:54   at java.lang.Thread.run(Thread.java:748)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-29783) Flaky test: KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime

2022-10-27 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser closed FLINK-29783.
--
Resolution: Duplicate

Duplicate of FLINk-24119

> Flaky test: 
> KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime
> -
>
> Key: FLINK-29783
> URL: https://issues.apache.org/jira/browse/FLINK-29783
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.17.0
>Reporter: Gabor Somogyi
>Priority: Major
>  Labels: pull-request-available
>
> {code:java}
> Oct 27 15:07:54 java.lang.AssertionError: Create test topic : 
> partition_failure_recovery_EventTime failed, 
> org.apache.kafka.common.errors.TopicExistsException: Topic 
> 'partition_failure_recovery_EventTime' already exists.
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.KafkaTestEnvironmentImpl.createTestTopic(KafkaTestEnvironmentImpl.java:207)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.KafkaTestEnvironment.createTestTopic(KafkaTestEnvironment.java:97)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.KafkaTestBase.createTestTopic(KafkaTestBase.java:217)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.shuffle.KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecovery(KafkaShuffleExactlyOnceITCase.java:158)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.shuffle.KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime(KafkaShuffleExactlyOnceITCase.java:101)
> Oct 27 15:07:54   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> Oct 27 15:07:54   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> Oct 27 15:07:54   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> Oct 27 15:07:54   at java.lang.reflect.Method.invoke(Method.java:498)
> Oct 27 15:07:54   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> Oct 27 15:07:54   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
> Oct 27 15:07:54   at 
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> Oct 27 15:07:54   at java.lang.Thread.run(Thread.java:748)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-29716) Separate slf4j jar in the lib folder from the distribution

2022-10-27 Thread Alexis Sarda-Espinosa (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29716?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexis Sarda-Espinosa closed FLINK-29716.
-
Resolution: Won't Do

> Separate slf4j jar in the lib folder from the distribution
> --
>
> Key: FLINK-29716
> URL: https://issues.apache.org/jira/browse/FLINK-29716
> Project: Flink
>  Issue Type: Improvement
>Affects Versions: 1.15.2
>Reporter: Alexis Sarda-Espinosa
>Priority: Major
>
> Flink's binary distribution includes several jars under the {{lib}} folder, 
> which has individual jars for all log4j artifacts. This makes it relatively 
> easy to swap out those logging jars when necessary, for example when critical 
> vulnerabilities are found (as was recently the case).
> With SLF4J 2.+, some breaking changes mean that many implementations are not 
> directly backwards compatible, see for example the [notes for 
> log4j2|https://logging.apache.org/log4j/2.x/log4j-slf4j-impl/index.html]. 
> This means that, in the future, if swapping logging jars were necessary, the 
> SLF4J jar might have to be changed as well.
> Right now the SLF4J jar is not included separately in the distribution, I 
> believe it's packed inside the {{flink-dist}} jar, although I'm not sure. It 
> would be better to separate that as it is done for the default log4j2 jars.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29716) Separate slf4j jar in the lib folder from the distribution

2022-10-27 Thread Alexis Sarda-Espinosa (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29716?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625336#comment-17625336
 ] 

Alexis Sarda-Espinosa commented on FLINK-29716:
---

I see, all right.

> Separate slf4j jar in the lib folder from the distribution
> --
>
> Key: FLINK-29716
> URL: https://issues.apache.org/jira/browse/FLINK-29716
> Project: Flink
>  Issue Type: Improvement
>Affects Versions: 1.15.2
>Reporter: Alexis Sarda-Espinosa
>Priority: Major
>
> Flink's binary distribution includes several jars under the {{lib}} folder, 
> which has individual jars for all log4j artifacts. This makes it relatively 
> easy to swap out those logging jars when necessary, for example when critical 
> vulnerabilities are found (as was recently the case).
> With SLF4J 2.+, some breaking changes mean that many implementations are not 
> directly backwards compatible, see for example the [notes for 
> log4j2|https://logging.apache.org/log4j/2.x/log4j-slf4j-impl/index.html]. 
> This means that, in the future, if swapping logging jars were necessary, the 
> SLF4J jar might have to be changed as well.
> Right now the SLF4J jar is not included separately in the distribution, I 
> believe it's packed inside the {{flink-dist}} jar, although I'm not sure. It 
> would be better to separate that as it is done for the default log4j2 jars.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] ericxiao251 commented on pull request #21077: [FLINK-29498] Add Scala Async Retry Strategies and ResultPredicates Helper Classes

2022-10-27 Thread GitBox


ericxiao251 commented on PR #21077:
URL: https://github.com/apache/flink/pull/21077#issuecomment-1294048147

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-29572) Flink Task Manager skip loopback interface for resource manager registration

2022-10-27 Thread Kevin Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29572?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625277#comment-17625277
 ] 

Kevin Li commented on FLINK-29572:
--

It will work if we configure different ports for each different task manager. 
But that will be cumbersome. If you have 10 task manager, you need to create 10 
different deployments for each of them. Also autoscale could be issue too. 
Rather than you have one deployment with 10 replicas and they can scale up and 
down.

I downgrade my Flink to 1.14.6 and it works fine. Looks like it was introduced 
by  FLINK-24474.

> Flink Task Manager skip loopback interface for resource manager registration
> 
>
> Key: FLINK-29572
> URL: https://issues.apache.org/jira/browse/FLINK-29572
> Project: Flink
>  Issue Type: Bug
>  Components: API / Core
>Affects Versions: 1.15.2
> Environment: Flink 1.15.2
> Kubernetes with Istio Proxy
>Reporter: Kevin Li
>Priority: Major
>
> Currently Flink Task Manager use different local interface to bind to connect 
> to Resource Manager. First one is Loopback interface. Normally if Job Manager 
> is running on remote host/container, using loopback interface to connect will 
> fail and it will pick up correct IP address.
> However, if Task Manager is running with some proxy, loopback interface can 
> connect to remote host as well. This will result 127.0.0.1 reported to 
> Resource Manager during registration, even Job Manager/Resource Manager runs 
> on remote host, and problem will happen. For us, only one Task Manager can 
> register in this case.
> I suggest adding configuration to skip Loopback interface check if we know 
> Job/Resource Manager is running on remote host/container.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] gaborgsomogyi commented on pull request #21180: [FLINK-29783][tests] Add some random to KafkaShuffleExactlyOnceITCase topic names

2022-10-27 Thread GitBox


gaborgsomogyi commented on PR #21180:
URL: https://github.com/apache/flink/pull/21180#issuecomment-1293940368

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-29784) Build fails with There is at least one incompatibility: org.apache.flink.api.connector.source.SourceReader

2022-10-27 Thread Sergey Nuyanzin (Jira)
Sergey Nuyanzin created FLINK-29784:
---

 Summary: Build fails with There is at least one incompatibility: 
org.apache.flink.api.connector.source.SourceReader
 Key: FLINK-29784
 URL: https://issues.apache.org/jira/browse/FLINK-29784
 Project: Flink
  Issue Type: Bug
  Components: API / Core
Affects Versions: 1.17.0
Reporter: Sergey Nuyanzin


currently {{./mvnw clean install -DskipTests}} fails with 
{noformat}

[ERROR] Failed to execute goal 
io.github.zentol.japicmp:japicmp-maven-plugin:0.16.0_m325:cmp (default) on 
project flink-core: There is at least one incompatibility: 
org.apache.flink.api.connector.source.SourceReader.pauseOrResumeSplits(java.util.Collection,java.util.Collection):METHOD_NEW_DEFAULT
 -> [Help 1]
[ERROR] 

{noformat}
It starts failing after this commit 
https://github.com/apache/flink/commit/82567cc9e9a23a2b6ca41f433c4b9310c0075767
removal of that commit fixes the behavior 
// cc [~hxbks2ks] may be you know more details about that



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] gaborgsomogyi commented on pull request #21160: [FLINK-29761][runtime][security] Simplify HadoopModule

2022-10-27 Thread GitBox


gaborgsomogyi commented on PR #21160:
URL: https://github.com/apache/flink/pull/21160#issuecomment-1293904151

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] gaborgsomogyi commented on pull request #21180: [FLINK-29783][tests] Add some random to KafkaShuffleExactlyOnceITCase topic names

2022-10-27 Thread GitBox


gaborgsomogyi commented on PR #21180:
URL: https://github.com/apache/flink/pull/21180#issuecomment-1293903766

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] gaborgsomogyi commented on pull request #21147: [FLINK-28330][runtime][security] Remove old delegation token framework code

2022-10-27 Thread GitBox


gaborgsomogyi commented on PR #21147:
URL: https://github.com/apache/flink/pull/21147#issuecomment-1293897838

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] flinkbot commented on pull request #21181: [hotfix][doc] `FlinkKafkaConsumer` will be removed with Flink 1.17 instead of 1.15

2022-10-27 Thread GitBox


flinkbot commented on PR #21181:
URL: https://github.com/apache/flink/pull/21181#issuecomment-1293882353

   
   ## CI report:
   
   * 85079793ed48230d97a1921862c7d28304f34fc7 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] JingGe opened a new pull request, #21181: [hotfix][doc] `FlinkKafkaConsumer` will be removed with Flink 1.17 instead of 1.15

2022-10-27 Thread GitBox


JingGe opened a new pull request, #21181:
URL: https://github.com/apache/flink/pull/21181

   ## What is the purpose of the change
   
   Back port #21172  to master: Change the docs: FlinkKafkaConsumer is 
deprecated and will be removed with Flink 1.17 instead of 1.15 since Flink 1.16 
has code freeze and will be released soon.
   
   ## Brief change log
   
- kafka.md
- zh kafka.md
   
   
   ## Verifying this change
   
   This change is a trivial doc update without any test coverage.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] sap1ens commented on a diff in pull request #20475: [FLINK-28842][Connector/Kafka] Add client.id.prefix for the KafkaSink

2022-10-27 Thread GitBox


sap1ens commented on code in PR #20475:
URL: https://github.com/apache/flink/pull/20475#discussion_r1007187247


##
flink-connectors/flink-connector-kafka/src/main/java/org/apache/flink/connector/kafka/sink/KafkaWriter.java:
##
@@ -398,6 +407,11 @@ private void registerMetricSync() {
 });
 }
 
+private String createProducerClientId(Properties props, int subtaskId) {
+String prefix = 
props.getProperty(KafkaSinkOptions.CLIENT_ID_PREFIX.key());
+return prefix + "-" + subtaskId;

Review Comment:
   Right, updated!



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] SaravananArumugam-Opus commented on a diff in pull request #20922: [DRAFT] Apply changes for calcite 1.3x

2022-10-27 Thread GitBox


SaravananArumugam-Opus commented on code in PR #20922:
URL: https://github.com/apache/flink/pull/20922#discussion_r1007141005


##
flink-table/pom.xml:
##
@@ -76,10 +76,10 @@ under the License.

 

-   1.26.0
+   1.32.0

3.0.11

Review Comment:
   Got it… Let me give a try … 



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] flinkbot commented on pull request #21180: [FLINK-29783][Tests] Add some random to KafkaShuffleExactlyOnceITCasetopic names

2022-10-27 Thread GitBox


flinkbot commented on PR #21180:
URL: https://github.com/apache/flink/pull/21180#issuecomment-1293803451

   
   ## CI report:
   
   * ccc5e22c6617909e90705c11e4e28dfc12d9310b UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] gaborgsomogyi commented on pull request #21180: [FLINK-29783][Tests] Add some random to KafkaShuffleExactlyOnceITCasetopic names

2022-10-27 Thread GitBox


gaborgsomogyi commented on PR #21180:
URL: https://github.com/apache/flink/pull/21180#issuecomment-1293804227

   Let's see whether it helps.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-29783) Flaky test: KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime

2022-10-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-29783:
---
Labels: pull-request-available  (was: )

> Flaky test: 
> KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime
> -
>
> Key: FLINK-29783
> URL: https://issues.apache.org/jira/browse/FLINK-29783
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.17.0
>Reporter: Gabor Somogyi
>Priority: Major
>  Labels: pull-request-available
>
> {code:java}
> Oct 27 15:07:54 java.lang.AssertionError: Create test topic : 
> partition_failure_recovery_EventTime failed, 
> org.apache.kafka.common.errors.TopicExistsException: Topic 
> 'partition_failure_recovery_EventTime' already exists.
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.KafkaTestEnvironmentImpl.createTestTopic(KafkaTestEnvironmentImpl.java:207)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.KafkaTestEnvironment.createTestTopic(KafkaTestEnvironment.java:97)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.KafkaTestBase.createTestTopic(KafkaTestBase.java:217)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.shuffle.KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecovery(KafkaShuffleExactlyOnceITCase.java:158)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.shuffle.KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime(KafkaShuffleExactlyOnceITCase.java:101)
> Oct 27 15:07:54   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> Oct 27 15:07:54   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> Oct 27 15:07:54   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> Oct 27 15:07:54   at java.lang.reflect.Method.invoke(Method.java:498)
> Oct 27 15:07:54   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> Oct 27 15:07:54   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
> Oct 27 15:07:54   at 
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> Oct 27 15:07:54   at java.lang.Thread.run(Thread.java:748)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] gaborgsomogyi opened a new pull request, #21180: [FLINK-29783][Tests] Add some random to KafkaShuffleExactlyOnceITCasetopic names

2022-10-27 Thread GitBox


gaborgsomogyi opened a new pull request, #21180:
URL: https://github.com/apache/flink/pull/21180

   ## What is the purpose of the change
   
   
`KafkaShuffleExactlyOnceITCasetopic.testAssignedToPartitionFailureRecoveryEventTime`
 fails with the following exception:
   ```
   {code:java}
   Oct 27 15:07:54 java.lang.AssertionError: Create test topic : 
partition_failure_recovery_EventTime failed, 
org.apache.kafka.common.errors.TopicExistsException: Topic 
'partition_failure_recovery_EventTime' already exists.
   Oct 27 15:07:54  at 
org.apache.flink.streaming.connectors.kafka.KafkaTestEnvironmentImpl.createTestTopic(KafkaTestEnvironmentImpl.java:207)
   Oct 27 15:07:54  at 
org.apache.flink.streaming.connectors.kafka.KafkaTestEnvironment.createTestTopic(KafkaTestEnvironment.java:97)
   Oct 27 15:07:54  at 
org.apache.flink.streaming.connectors.kafka.KafkaTestBase.createTestTopic(KafkaTestBase.java:217)
   Oct 27 15:07:54  at 
org.apache.flink.streaming.connectors.kafka.shuffle.KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecovery(KafkaShuffleExactlyOnceITCase.java:158)
   Oct 27 15:07:54  at 
org.apache.flink.streaming.connectors.kafka.shuffle.KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime(KafkaShuffleExactlyOnceITCase.java:101)
   Oct 27 15:07:54  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
Method)
   Oct 27 15:07:54  at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   Oct 27 15:07:54  at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   Oct 27 15:07:54  at java.lang.reflect.Method.invoke(Method.java:498)
   Oct 27 15:07:54  at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
   Oct 27 15:07:54  at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
   Oct 27 15:07:54  at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
   Oct 27 15:07:54  at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
   Oct 27 15:07:54  at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
   Oct 27 15:07:54  at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
   Oct 27 15:07:54  at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
   Oct 27 15:07:54  at 
java.util.concurrent.FutureTask.run(FutureTask.java:266)
   Oct 27 15:07:54  at java.lang.Thread.run(Thread.java:748)
   {code}
   ```
   
   I don't see any reason why multiple times the same topic would be created so 
what I assume is that somehow the test is executed multiple times within the 
JVM and Kafka instance is class bounded.
   
   In the current PR I've made a quick fix to make this disappear and allow 
peoples to work on features.
   
   ## Brief change log
   
   Added some random to KafkaShuffleExactlyOnceITCasetopic names.
   
   ## Verifying this change
   
   Existing unit tests.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? not applicable
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] snuyanzin commented on a diff in pull request #20922: [DRAFT] Apply changes for calcite 1.3x

2022-10-27 Thread GitBox


snuyanzin commented on code in PR #20922:
URL: https://github.com/apache/flink/pull/20922#discussion_r1007127903


##
flink-table/pom.xml:
##
@@ -76,10 +76,10 @@ under the License.

 

-   1.26.0
+   1.32.0

3.0.11

Review Comment:
   yes, i know
   however from another side current there is a huge amount of tests which 
should be passed.
   With current changes they are passing.
   With janino 3.1.x no.
   Once you have them passing - feel free to update the version



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-29783) Flaky test: KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime

2022-10-27 Thread Gabor Somogyi (Jira)
Gabor Somogyi created FLINK-29783:
-

 Summary: Flaky test: 
KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime
 Key: FLINK-29783
 URL: https://issues.apache.org/jira/browse/FLINK-29783
 Project: Flink
  Issue Type: Bug
  Components: Tests
Affects Versions: 1.17.0
Reporter: Gabor Somogyi


{code:java}
Oct 27 15:07:54 java.lang.AssertionError: Create test topic : 
partition_failure_recovery_EventTime failed, 
org.apache.kafka.common.errors.TopicExistsException: Topic 
'partition_failure_recovery_EventTime' already exists.
Oct 27 15:07:54 at 
org.apache.flink.streaming.connectors.kafka.KafkaTestEnvironmentImpl.createTestTopic(KafkaTestEnvironmentImpl.java:207)
Oct 27 15:07:54 at 
org.apache.flink.streaming.connectors.kafka.KafkaTestEnvironment.createTestTopic(KafkaTestEnvironment.java:97)
Oct 27 15:07:54 at 
org.apache.flink.streaming.connectors.kafka.KafkaTestBase.createTestTopic(KafkaTestBase.java:217)
Oct 27 15:07:54 at 
org.apache.flink.streaming.connectors.kafka.shuffle.KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecovery(KafkaShuffleExactlyOnceITCase.java:158)
Oct 27 15:07:54 at 
org.apache.flink.streaming.connectors.kafka.shuffle.KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime(KafkaShuffleExactlyOnceITCase.java:101)
Oct 27 15:07:54 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
Method)
Oct 27 15:07:54 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
Oct 27 15:07:54 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
Oct 27 15:07:54 at java.lang.reflect.Method.invoke(Method.java:498)
Oct 27 15:07:54 at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
Oct 27 15:07:54 at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
Oct 27 15:07:54 at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
Oct 27 15:07:54 at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
Oct 27 15:07:54 at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
Oct 27 15:07:54 at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
Oct 27 15:07:54 at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
Oct 27 15:07:54 at 
java.util.concurrent.FutureTask.run(FutureTask.java:266)
Oct 27 15:07:54 at java.lang.Thread.run(Thread.java:748)
{code}




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29783) Flaky test: KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime

2022-10-27 Thread Gabor Somogyi (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29783?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625201#comment-17625201
 ] 

Gabor Somogyi commented on FLINK-29783:
---

It's super annoying to restart CI and hope that it works. Trying to add a quick 
fix.

> Flaky test: 
> KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime
> -
>
> Key: FLINK-29783
> URL: https://issues.apache.org/jira/browse/FLINK-29783
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.17.0
>Reporter: Gabor Somogyi
>Priority: Major
>
> {code:java}
> Oct 27 15:07:54 java.lang.AssertionError: Create test topic : 
> partition_failure_recovery_EventTime failed, 
> org.apache.kafka.common.errors.TopicExistsException: Topic 
> 'partition_failure_recovery_EventTime' already exists.
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.KafkaTestEnvironmentImpl.createTestTopic(KafkaTestEnvironmentImpl.java:207)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.KafkaTestEnvironment.createTestTopic(KafkaTestEnvironment.java:97)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.KafkaTestBase.createTestTopic(KafkaTestBase.java:217)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.shuffle.KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecovery(KafkaShuffleExactlyOnceITCase.java:158)
> Oct 27 15:07:54   at 
> org.apache.flink.streaming.connectors.kafka.shuffle.KafkaShuffleExactlyOnceITCase.testAssignedToPartitionFailureRecoveryEventTime(KafkaShuffleExactlyOnceITCase.java:101)
> Oct 27 15:07:54   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> Oct 27 15:07:54   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> Oct 27 15:07:54   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> Oct 27 15:07:54   at java.lang.reflect.Method.invoke(Method.java:498)
> Oct 27 15:07:54   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> Oct 27 15:07:54   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
> Oct 27 15:07:54   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
> Oct 27 15:07:54   at 
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> Oct 27 15:07:54   at java.lang.Thread.run(Thread.java:748)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] SaravananArumugam-Opus commented on a diff in pull request #20922: [DRAFT] Apply changes for calcite 1.3x

2022-10-27 Thread GitBox


SaravananArumugam-Opus commented on code in PR #20922:
URL: https://github.com/apache/flink/pull/20922#discussion_r1007121946


##
flink-table/pom.xml:
##
@@ -76,10 +76,10 @@ under the License.

 

-   1.26.0
+   1.32.0

3.0.11

Review Comment:
   Calcite - 1.32.0 has dependency on Janino 3.1.8.  (\)



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] flinkbot commented on pull request #21179: [BP-1.16][FLINK-29539][Deployment/Kubernetes] dnsPolicy in FlinkPod is not overridable

2022-10-27 Thread GitBox


flinkbot commented on PR #21179:
URL: https://github.com/apache/flink/pull/21179#issuecomment-1293778963

   
   ## CI report:
   
   * 1f30c93213d37ea19210229027af68bcefe51a2f UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] snuyanzin commented on a diff in pull request #20922: [DRAFT] Apply changes for calcite 1.3x

2022-10-27 Thread GitBox


snuyanzin commented on code in PR #20922:
URL: https://github.com/apache/flink/pull/20922#discussion_r1007111394


##
flink-table/pom.xml:
##
@@ -76,10 +76,10 @@ under the License.

 

-   1.26.0
+   1.32.0

3.0.11

Review Comment:
   Before doing such update it's required to make it working...
   right now it doesn't work...



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] SaravananArumugam-Opus commented on a diff in pull request #20922: [DRAFT] Apply changes for calcite 1.3x

2022-10-27 Thread GitBox


SaravananArumugam-Opus commented on code in PR #20922:
URL: https://github.com/apache/flink/pull/20922#discussion_r1007109247


##
flink-table/pom.xml:
##
@@ -76,10 +76,10 @@ under the License.

 

-   1.26.0
+   1.32.0

3.0.11

Review Comment:
   Update Janino version to 3.1.8 
(https://github.com/apache/calcite/blob/calcite-1.32.0/gradle.properties#L118)



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] flinkbot commented on pull request #21178: [BP-1.15][FLINK-29539][Deployment/Kubernetes] dnsPolicy in FlinkPod is not overridable

2022-10-27 Thread GitBox


flinkbot commented on PR #21178:
URL: https://github.com/apache/flink/pull/21178#issuecomment-1293768219

   
   ## CI report:
   
   * 096aa77e435541c36e0442b2892894e4e6c419f3 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-29654) Vulnerable libraries - Flink 1.15.2

2022-10-27 Thread nagasudhakar (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29654?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625193#comment-17625193
 ] 

nagasudhakar commented on FLINK-29654:
--

Thanks [~martijnvisser]

> Vulnerable libraries - Flink 1.15.2
> ---
>
> Key: FLINK-29654
> URL: https://issues.apache.org/jira/browse/FLINK-29654
> Project: Flink
>  Issue Type: Bug
>  Components: Build System
>Affects Versions: 1.15.2
>Reporter: nagasudhakar
>Priority: Major
>
> Hi, our organisation ran a security scan on Flink-1.15.2 release and found 
> the following vulnerable open source libraries being used -
> JDOM1.1
> kryo2.24.0
> libnetty-3.9-java3.9.0.Final
> Netty Project3.10.6.Final
> Play2.6.11
> Apache Tika1.28.1
> Apache Avro1.7.7
> Apache Kafka2.8.1
> The recommended versions for these libraries are -
> JDOM2.0.2
> kryo-5.5.0
> libnetty-3.9-java3.9.9.Final
> Netty Project 5.0.0.Final
> Play2.8.16
> Apache Tika2.4.1
> Apache Avro1.8.2
> Apache Kafka2.8.2



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] carloscastrojumo commented on pull request #20982: [FLINK-29539] [Deployment/Kubernetes] dnsPolicy in FlinkPod is not overridable

2022-10-27 Thread GitBox


carloscastrojumo commented on PR #20982:
URL: https://github.com/apache/flink/pull/20982#issuecomment-1293765226

   > @carloscastrojumo Could you please create the backport PR for release-1.15 
and release-1.16?
   
   Done @wangyang0918 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] carloscastrojumo opened a new pull request, #21179: [BP-1.16][FLINK-29539][Deployment/Kubernetes] dnsPolicy in FlinkPod is not overridable

2022-10-27 Thread GitBox


carloscastrojumo opened a new pull request, #21179:
URL: https://github.com/apache/flink/pull/21179

   This PR is a backport of PR https://github.com/apache/flink/pull/20982 that 
relates to [FLINK-29539](https://issues.apache.org/jira/browse/FLINK-29539)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] carloscastrojumo opened a new pull request, #21178: [BP-1.15][FLINK-29539][Deployment/Kubernetes] dnsPolicy in FlinkPod is not overridable

2022-10-27 Thread GitBox


carloscastrojumo opened a new pull request, #21178:
URL: https://github.com/apache/flink/pull/21178

   This PR is a backport of PR https://github.com/apache/flink/pull/20982 that 
relates to [FLINK-29539](https://issues.apache.org/jira/browse/FLINK-29539)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



  1   2   3   >