[jira] [Updated] (FLINK-29424) Add Pull Request Template for flink-ml Project

2022-09-26 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29424?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-29424:
---
Labels: pull-request-available  (was: )

> Add Pull Request Template for flink-ml Project
> --
>
> Key: FLINK-29424
> URL: https://issues.apache.org/jira/browse/FLINK-29424
> Project: Flink
>  Issue Type: New Feature
>  Components: Library / Machine Learning
>Reporter: Jiang Xin
>Priority: Major
>  Labels: pull-request-available
>
> Add Pull Request Template for flink-ml Project



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] jiangxin369 opened a new pull request, #159: [FLINK-29424] Add Pull Request Template

2022-09-26 Thread GitBox


jiangxin369 opened a new pull request, #159:
URL: https://github.com/apache/flink-ml/pull/159

   ## What is the purpose of the change
   
   Add Pull Request Template for Flink ML
   
   ## Brief change log
   
 - Add Pull Request Template for Flink ML
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ not documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-ml] jiangxin369 commented on a diff in pull request #158: [FLINK-29409] Add Transformer and Estimator for VarianceThresholdSelector

2022-09-26 Thread GitBox


jiangxin369 commented on code in PR #158:
URL: https://github.com/apache/flink-ml/pull/158#discussion_r980765655


##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/variancethresholdselector/VarianceThresholdSelectorParams.java:
##
@@ -0,0 +1,48 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.feature.variancethresholdselector;
+
+import org.apache.flink.ml.param.DoubleParam;
+import org.apache.flink.ml.param.Param;
+import org.apache.flink.ml.param.ParamValidators;
+
+/**
+ * * Params of VarianceThresholdSelectorModel.

Review Comment:
   Thanks for your code review, I'll remove it.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-ml] yunfengzhou-hub commented on a diff in pull request #158: [FLINK-29409] Add Transformer and Estimator for VarianceThresholdSelector

2022-09-26 Thread GitBox


yunfengzhou-hub commented on code in PR #158:
URL: https://github.com/apache/flink-ml/pull/158#discussion_r980705519


##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/variancethresholdselector/VarianceThresholdSelectorParams.java:
##
@@ -0,0 +1,48 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.feature.variancethresholdselector;
+
+import org.apache.flink.ml.param.DoubleParam;
+import org.apache.flink.ml.param.Param;
+import org.apache.flink.ml.param.ParamValidators;
+
+/**
+ * * Params of VarianceThresholdSelectorModel.

Review Comment:
   nit: "Params of VarianceThresholdSelector". The `*` before "Params" should 
be removed.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-ml] yunfengzhou-hub commented on a diff in pull request #158: [FLINK-29409] Add Transformer and Estimator for VarianceThresholdSelector

2022-09-26 Thread GitBox


yunfengzhou-hub commented on code in PR #158:
URL: https://github.com/apache/flink-ml/pull/158#discussion_r980705519


##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/variancethresholdselector/VarianceThresholdSelectorParams.java:
##
@@ -0,0 +1,48 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.feature.variancethresholdselector;
+
+import org.apache.flink.ml.param.DoubleParam;
+import org.apache.flink.ml.param.Param;
+import org.apache.flink.ml.param.ParamValidators;
+
+/**
+ * * Params of VarianceThresholdSelectorModel.

Review Comment:
   nit: "Params of VarianceThresholdSelector". The `*` before "Params` should 
be removed.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Assigned] (FLINK-28256) Move the write and prepareCommit logic of AbstractTableWrite to FileStoreWrite

2022-09-26 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28256?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee reassigned FLINK-28256:


Assignee: Shammon

> Move the write and prepareCommit logic of AbstractTableWrite to FileStoreWrite
> --
>
> Key: FLINK-28256
> URL: https://issues.apache.org/jira/browse/FLINK-28256
> Project: Flink
>  Issue Type: New Feature
>  Components: Table Store
>Reporter: Jingsong Lee
>Assignee: Shammon
>Priority: Minor
>  Labels: pull-request-available
> Fix For: table-store-0.3.0
>
>
> TableWrite can be a simple key value converter wrapper build on 
> FileStoreWrite.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-28256) Move the write and prepareCommit logic of AbstractTableWrite to FileStoreWrite

2022-09-26 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28256?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-28256:
---
Labels: pull-request-available  (was: )

> Move the write and prepareCommit logic of AbstractTableWrite to FileStoreWrite
> --
>
> Key: FLINK-28256
> URL: https://issues.apache.org/jira/browse/FLINK-28256
> Project: Flink
>  Issue Type: New Feature
>  Components: Table Store
>Reporter: Jingsong Lee
>Priority: Minor
>  Labels: pull-request-available
> Fix For: table-store-0.3.0
>
>
> TableWrite can be a simple key value converter wrapper build on 
> FileStoreWrite.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-table-store] zjureel opened a new pull request, #305: [FLINK-28256] Move the write and prepareCommit logic of AbstractTableWrite to FileStoreWrite

2022-09-26 Thread GitBox


zjureel opened a new pull request, #305:
URL: https://github.com/apache/flink-table-store/pull/305

   Currently `TableWrite` creates and manages the `RecordWriter`, which is 
created by `FileStoreWrite`. As `TableWrite` corresponds to `FileStoreTable` 
and `FileStoreWrite` corresponds to `FileStore`, it's better to move the 
`write` and `prepareCommit` logic of `AbstractTableWrite` to `FileStoreWrite`


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-29426) Flink ML 2.1.0 Installation

2022-09-26 Thread Tony Sun (Jira)
Tony Sun created FLINK-29426:


 Summary: Flink ML 2.1.0 Installation
 Key: FLINK-29426
 URL: https://issues.apache.org/jira/browse/FLINK-29426
 Project: Flink
  Issue Type: Bug
  Components: Quickstarts
Affects Versions: ml-2.1.0
 Environment: Java version: 11.0.16

Flink version: 1.15.2
Reporter: Tony Sun


Trying to go through Flink-ML quickstart but having an issue with building 
Flink ML's source code.

Steps to reproduce:

1. sudo wget 
https://dlcdn.apache.org/flink/flink-ml-2.1.0/flink-ml-2.1.0-src.tgz
2. sudo tar -xzf flink-ml-2.1.0-src.tgz
3. mvn clean package -DskipTests -Dcheckstyle.skip
{quote}Failed to execute goal 
org.apache.maven.plugins:maven-checkstyle-plugin:2.17:check (validate) on 
project flink-ml-parent: Unable to create output stream: 
/usr/src/flink-ml-2.1.0/target/checkstyle-result.xml: 
/usr/src/flink-ml-2.1.0/target/checkstyle-result.xml (No such file or 
directory) -> [Help 1]
{quote}
It doesn't seem like `target` is a folder in `flink-ml-2.1.0`.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-docker] HuangXingBo closed pull request #136: Update Dockerfiles for 1.14.6 release

2022-09-26 Thread GitBox


HuangXingBo closed pull request #136: Update Dockerfiles for 1.14.6 release
URL: https://github.com/apache/flink-docker/pull/136


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-docker] HuangXingBo closed pull request #135: Add GPG key for 1.14.6 release

2022-09-26 Thread GitBox


HuangXingBo closed pull request #135: Add GPG key for 1.14.6 release
URL: https://github.com/apache/flink-docker/pull/135


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-29425) Hybrid full spilling strategy triggering spilling frequently

2022-09-26 Thread Weijie Guo (Jira)
Weijie Guo created FLINK-29425:
--

 Summary: Hybrid full spilling strategy triggering spilling 
frequently
 Key: FLINK-29425
 URL: https://issues.apache.org/jira/browse/FLINK-29425
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Network
Affects Versions: 1.16.0
Reporter: Weijie Guo


In hybrid shuffle mode, we have an internal config option 
'DEFAULT_FULL_STRATEGY_NUM_BUFFERS_TRIGGER_SPILLED' to control spilling 
frequency in the full spilling strategy. Unfortunately, the default value(10) 
is too small. As a result, frequent disk spilling calls are observed in the 
TPC-DS test, which seriously affects performance. When we increase the value, 
the query performance is improved significantly. We should set a more 
reasonable default value, or adopt an adaptive strategy to determine the 
spilling frequency.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-19822) Remove redundant shuffle for streaming

2022-09-26 Thread Alexander Smirnov (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19822?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609812#comment-17609812
 ] 

Alexander Smirnov commented on FLINK-19822:
---

Hi [~godfreyhe]! I'm wondering why this improvement is still open, although it 
looks quite useful.  Are there any pitfalls in implementation from linked 
Github PR?

> Remove redundant shuffle for streaming
> --
>
> Key: FLINK-19822
> URL: https://issues.apache.org/jira/browse/FLINK-19822
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner
>Reporter: godfrey he
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> auto-unassigned, pull-request-available
>
> This is similar 
> [FLINK-12575|https://issues.apache.org/jira/browse/FLINK-12575], we could 
> implement {{satisfyTraits}} method for stream nodes to remove redundant 
> shuffle. This could add more possibilities that more operators can be merged 
> into multiple input operator.
> Different batch, stream operators require the shuffle keys and the state keys 
> must be exactly the same, otherwise the state may be not correct.
> We only support a few operators in this issue, such as Join and regular 
> Aggregate. Other operators will be supported in the future.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] lindong28 commented on a diff in pull request #157: [FLINK-28906] Support windowing in AgglomerativeClustering

2022-09-26 Thread GitBox


lindong28 commented on code in PR #157:
URL: https://github.com/apache/flink-ml/pull/157#discussion_r980594756


##
flink-ml-python/pyflink/ml/lib/clustering/agglomerativeclustering.py:
##
@@ -112,15 +113,20 @@ def compute_full_tree(self):
 class AgglomerativeClustering(JavaClusteringAlgoOperator, 
_AgglomerativeClusteringParams):
 """
 An AlgoOperator that performs a hierarchical clustering using a bottom-up 
approach. Each
-observation starts in its own cluster and the clusters are merged together 
one by one.
-Users can choose different strategies to merge two clusters by setting
-{@link AgglomerativeClusteringParams#LINKAGE} and different distance 
measures by setting
-{@link AgglomerativeClusteringParams#DISTANCE_MEASURE}.
+observation starts in its own cluster and the clusters are merged together 
one by one. Users can
+choose different strategies to merge two clusters by setting
+AgglomerativeClusteringParams#LINKAGE and different distance measures by 
setting
+AgglomerativeClusteringParams#DISTANCE_MEASURE.
 
 The output contains two tables. The first one assigns one cluster Id 
for each data point.
 The second one contains the information of merging two clusters at each 
step. The data format
 of the merging information is (clusterId1, clusterId2, distance, 
sizeOfMergedCluster).
 
+This operator supports the HasWindows parameter, which creates 
mini-batches(windows) from

Review Comment:
   Update the doc to be consistent with the corresponding Javadoc.



##
flink-ml-python/pyflink/ml/lib/clustering/tests/test_agglomerativeclustering.py:
##
@@ -58,6 +63,18 @@ def setUp(self):
 {Vectors.dense(4, 1.5), Vectors.dense(4, 0)}
 ]
 
+self.eucliean_ward_count_five_window_as_two_result = [

Review Comment:
   eucliean -> euclidean
   
   Same for other lines in this file.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-29424) Add Pull Request Template for flink-ml Project

2022-09-26 Thread Jiang Xin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29424?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jiang Xin updated FLINK-29424:
--
Component/s: Library / Machine Learning

> Add Pull Request Template for flink-ml Project
> --
>
> Key: FLINK-29424
> URL: https://issues.apache.org/jira/browse/FLINK-29424
> Project: Flink
>  Issue Type: New Feature
>  Components: Library / Machine Learning
>Reporter: Jiang Xin
>Priority: Major
>
> Add Pull Request Template for flink-ml Project



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29424) Add Pull Request Template for flink-ml Project

2022-09-26 Thread Jiang Xin (Jira)
Jiang Xin created FLINK-29424:
-

 Summary: Add Pull Request Template for flink-ml Project
 Key: FLINK-29424
 URL: https://issues.apache.org/jira/browse/FLINK-29424
 Project: Flink
  Issue Type: New Feature
Reporter: Jiang Xin


Add Pull Request Template for flink-ml Project



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] wangyang0918 commented on a diff in pull request #20867: [FLINK-29372] Add suffix to all options that conflict with YAML

2022-09-26 Thread GitBox


wangyang0918 commented on code in PR #20867:
URL: https://github.com/apache/flink/pull/20867#discussion_r980692851


##
flink-kubernetes/src/main/java/org/apache/flink/kubernetes/configuration/KubernetesConfigOptions.java:
##
@@ -447,7 +450,7 @@ public class KubernetesConfigOptions {
 + Constants.MAIN_CONTAINER_NAME
 + "'. If not explicitly configured, config 
option '"
 + KUBERNETES_POD_TEMPLATE_FILE_KEY
-+ "' will be used.");
++ ".default' will be used.");
 
 public static final ConfigOption TASK_MANAGER_POD_TEMPLATE =

Review Comment:
   We also need to update the fallback keys.
   
   ```
   .withFallbackKeys(KUBERNETES_POD_TEMPLATE_FILE_KEY + ".default")



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-29181) log.system can be congiured by dynamic options

2022-09-26 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29181?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-29181:
-
Issue Type: Improvement  (was: New Feature)

> log.system can be congiured by dynamic options
> --
>
> Key: FLINK-29181
> URL: https://issues.apache.org/jira/browse/FLINK-29181
> Project: Flink
>  Issue Type: Improvement
>  Components: Table Store
>Reporter: Jingsong Lee
>Assignee: Jingsong Lee
>Priority: Major
>  Labels: pull-request-available
> Fix For: table-store-0.3.0, table-store-0.2.1
>
>
> Now log.system default value is null, it can not be configured by dynamic 
> option.
> We can let default value is 'none'.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29154) Support LookupTableSource for table store

2022-09-26 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-29154:
-
Issue Type: Improvement  (was: New Feature)

> Support LookupTableSource for table store
> -
>
> Key: FLINK-29154
> URL: https://issues.apache.org/jira/browse/FLINK-29154
> Project: Flink
>  Issue Type: Improvement
>  Components: Table Store
>Reporter: Jingsong Lee
>Assignee: Jingsong Lee
>Priority: Major
>  Labels: pull-request-available
> Fix For: table-store-0.3.0, table-store-0.2.1
>
>
> At present, the bottom layer of Table Store is LSM structure, and it has full 
> and incremental reading, which can have certain lookup capability. We can 
> unlock Table Store to implement LookupTableSource.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] jiangxin369 commented on a diff in pull request #158: [FLINK-29409] Add Transformer and Estimator for VarianceThresholdSelector

2022-09-26 Thread GitBox


jiangxin369 commented on code in PR #158:
URL: https://github.com/apache/flink-ml/pull/158#discussion_r980677136


##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/variancethresholdselector/VarianceThresholdSelectorModelData.java:
##
@@ -0,0 +1,121 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.feature.variancethresholdselector;
+
+import org.apache.flink.api.common.serialization.Encoder;
+import org.apache.flink.api.common.typeinfo.TypeInformation;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.connector.file.src.reader.SimpleStreamFormat;
+import org.apache.flink.core.fs.FSDataInputStream;
+import org.apache.flink.streaming.api.datastream.DataStream;
+import org.apache.flink.table.api.Table;
+import org.apache.flink.table.api.bridge.java.StreamTableEnvironment;
+import org.apache.flink.table.api.internal.TableImpl;
+
+import com.esotericsoftware.kryo.io.Input;
+import com.esotericsoftware.kryo.io.Output;
+
+import java.io.IOException;
+import java.io.OutputStream;
+import java.util.ArrayList;
+import java.util.List;
+
+/**
+ * Model data of {@link VarianceThresholdSelectorModel}.
+ *
+ * This class also provides methods to convert model data from Table to a 
data stream, and
+ * classes to save/load model data.
+ */
+public class VarianceThresholdSelectorModelData {
+
+public int numOfFeatures;
+public int[] indices;

Review Comment:
   I think there should only be indices in the model data (so do Spark ML and 
sklearn),  because the model needs to be transformed on different datasets and 
it has better performance.
   Indeed, it's better to divide `VarianceThresholdSelectorParams` and 
`VarianceThresholdSelectorModelParams`, I'll do it.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-29154) Support LookupTableSource for table store

2022-09-26 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-29154:
-
Fix Version/s: table-store-0.2.1

> Support LookupTableSource for table store
> -
>
> Key: FLINK-29154
> URL: https://issues.apache.org/jira/browse/FLINK-29154
> Project: Flink
>  Issue Type: New Feature
>  Components: Table Store
>Reporter: Jingsong Lee
>Assignee: Jingsong Lee
>Priority: Major
>  Labels: pull-request-available
> Fix For: table-store-0.3.0, table-store-0.2.1
>
>
> At present, the bottom layer of Table Store is LSM structure, and it has full 
> and incremental reading, which can have certain lookup capability. We can 
> unlock Table Store to implement LookupTableSource.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-29154) Support LookupTableSource for table store

2022-09-26 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29154?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17600665#comment-17600665
 ] 

Jingsong Lee edited comment on FLINK-29154 at 9/27/22 2:09 AM:
---

master: e55dde87296b51132d37fb4265bd3c863f05db40
release-0.2: c24016d512b5c01c04b6976b19ebdbf9a2ba78d4


was (Author: lzljs3620320):
master: e55dde87296b51132d37fb4265bd3c863f05db40

> Support LookupTableSource for table store
> -
>
> Key: FLINK-29154
> URL: https://issues.apache.org/jira/browse/FLINK-29154
> Project: Flink
>  Issue Type: New Feature
>  Components: Table Store
>Reporter: Jingsong Lee
>Assignee: Jingsong Lee
>Priority: Major
>  Labels: pull-request-available
> Fix For: table-store-0.3.0, table-store-0.2.1
>
>
> At present, the bottom layer of Table Store is LSM structure, and it has full 
> and incremental reading, which can have certain lookup capability. We can 
> unlock Table Store to implement LookupTableSource.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29421) Support python 3.10

2022-09-26 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609778#comment-17609778
 ] 

Huang Xingbo commented on FLINK-29421:
--

Currently, pyflink hasn't supported Python3.10. In release-1.16, pyflink will 
support 3.9. For 3.10 support, it will probably be in release 1.17.

> Support python 3.10
> ---
>
> Key: FLINK-29421
> URL: https://issues.apache.org/jira/browse/FLINK-29421
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Reporter: Eric Sirianni
>Priority: Minor
>
> The {{apache-flink}} package fails to install on Python 3.10 due to inability 
> to compile {{numpy}}
> {noformat}
> numpy/core/src/multiarray/scalartypes.c.src:3242:12: error: too 
> few arguments to function ‘_Py_HashDouble’
>  3242 | return 
> _Py_HashDouble(npy_half_to_double(((PyHalfScalarObject *)obj)->obval));
>   |^~
> In file included from 
> /home/sirianni/.asdf/installs/python/3.10.6/include/python3.10/Python.h:77,
>  from 
> numpy/core/src/multiarray/scalartypes.c.src:3:
> 
> /home/sirianni/.asdf/installs/python/3.10.6/include/python3.10/pyhash.h:10:23:
>  note: declared here
>10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double);
> {noformat}
> Numpy issue https://github.com/numpy/numpy/issues/19033
> [Mailing list 
> thread|https://lists.apache.org/thread/f4r9hjt1l33xf5ngnswszhnls4cxkk52]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] jiangxin369 commented on a diff in pull request #158: [FLINK-29409] Add Transformer and Estimator for VarianceThresholdSelector

2022-09-26 Thread GitBox


jiangxin369 commented on code in PR #158:
URL: https://github.com/apache/flink-ml/pull/158#discussion_r980656357


##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/variancethresholdselector/VarianceThresholdSelectorModel.java:
##
@@ -0,0 +1,172 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.feature.variancethresholdselector;
+
+import org.apache.flink.api.common.functions.RichMapFunction;
+import org.apache.flink.api.common.typeinfo.TypeInformation;
+import org.apache.flink.api.java.typeutils.RowTypeInfo;
+import org.apache.flink.ml.api.Model;
+import org.apache.flink.ml.common.broadcast.BroadcastUtils;
+import org.apache.flink.ml.common.datastream.TableUtils;
+import org.apache.flink.ml.linalg.DenseVector;
+import org.apache.flink.ml.linalg.Vector;
+import org.apache.flink.ml.linalg.Vectors;
+import org.apache.flink.ml.param.Param;
+import org.apache.flink.ml.util.ParamUtils;
+import org.apache.flink.ml.util.ReadWriteUtils;
+import org.apache.flink.streaming.api.datastream.DataStream;
+import org.apache.flink.table.api.Table;
+import org.apache.flink.table.api.bridge.java.StreamTableEnvironment;
+import org.apache.flink.table.api.internal.TableImpl;
+import org.apache.flink.types.Row;
+import org.apache.flink.util.Preconditions;
+
+import org.apache.commons.lang3.ArrayUtils;
+
+import java.io.IOException;
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.Map;
+
+/**
+ * A Model which removes low-variance data using the model data computed by 
{@link
+ * VarianceThresholdSelector}.
+ */
+public class VarianceThresholdSelectorModel
+implements Model,
+
VarianceThresholdSelectorParams {
+
+private final Map, Object> paramMap = new HashMap<>();
+private Table modelDataTable;
+
+public VarianceThresholdSelectorModel() {
+ParamUtils.initializeMapWithDefaultValues(paramMap, this);
+}
+
+@Override
+public VarianceThresholdSelectorModel setModelData(Table... inputs) {
+Preconditions.checkArgument(inputs.length == 1);
+modelDataTable = inputs[0];
+return this;
+}
+
+@Override
+public Table[] getModelData() {
+return new Table[] {modelDataTable};
+}
+
+@Override
+public Map, Object> getParamMap() {
+return paramMap;
+}
+
+@Override
+public void save(String path) throws IOException {
+ReadWriteUtils.saveMetadata(this, path);
+ReadWriteUtils.saveModelData(
+
VarianceThresholdSelectorModelData.getModelDataStream(modelDataTable),
+path,
+new VarianceThresholdSelectorModelData.ModelDataEncoder());
+}
+
+public static VarianceThresholdSelectorModel load(StreamTableEnvironment 
tEnv, String path)
+throws IOException {
+VarianceThresholdSelectorModel model = 
ReadWriteUtils.loadStageParam(path);
+Table modelDataTable =
+ReadWriteUtils.loadModelData(
+tEnv, path, new 
VarianceThresholdSelectorModelData.ModelDataDecoder());
+return model.setModelData(modelDataTable);
+}
+
+@Override
+public Table[] transform(Table... inputs) {
+Preconditions.checkArgument(inputs.length == 1);
+
+StreamTableEnvironment tEnv =
+(StreamTableEnvironment) ((TableImpl) 
inputs[0]).getTableEnvironment();
+DataStream data = tEnv.toDataStream(inputs[0]);
+DataStream 
varianceThresholdSelectorModel =
+
VarianceThresholdSelectorModelData.getModelDataStream(modelDataTable);
+
+final String broadcastModelKey = "broadcastModelKey";
+RowTypeInfo inputTypeInfo = 
TableUtils.getRowTypeInfo(inputs[0].getResolvedSchema());
+RowTypeInfo outputTypeInfo =
+new RowTypeInfo(
+ArrayUtils.addAll(
+inputTypeInfo.getFieldTypes(),
+TypeInformation.of(DenseVector.class)),
+ArrayUtils.addAll(inputTypeInfo.getFieldNames(), 
getOutputCol()));
+
+DataStream output =
+BroadcastUtils.withBroadcastStream(
+

[jira] [Closed] (FLINK-29219) CREATE TABLE AS statement blocks SQL client's execution

2022-09-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29219?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-29219.
---
Fix Version/s: 1.17.0
   Resolution: Fixed

Fixed in 
 - master: 44009efc5400c6706563fa53335eccf445cf5d0c
 - release-1.16: 4c126223cbc2dec3f08b50c1b398bdf522747ba9

> CREATE TABLE AS statement blocks SQL client's execution
> ---
>
> Key: FLINK-29219
> URL: https://issues.apache.org/jira/browse/FLINK-29219
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API, Table SQL / Client
>Affects Versions: 1.16.0
>Reporter: Qingsheng Ren
>Assignee: dalongliu
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.16.0, 1.17.0
>
>
> When executing CREATE TABLE AS statement to create a sink table in SQL 
> client, the client could create the table in catalog and submit the job to 
> cluster successfully, but stops emitting new prompts and accepts new inputs, 
> and user has to use SIGTERM (Control + C) to forcefully stop the SQL client. 
> As contrast the behavior of INSERT INTO statement in SQL client is printing 
> "Job is submitted with JobID " and being ready to accept user's input. 
> From the log it looks like the client was waiting for the job to finish.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] wuchong merged pull request #20880: [FLINK-29219][table] Fix CREATE TABLE AS statement blocks SQL client's execution

2022-09-26 Thread GitBox


wuchong merged PR #20880:
URL: https://github.com/apache/flink/pull/20880


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] wuchong merged pull request #20869: [FLINK-29219][table] Fix CREATE TABLE AS statement blocks SQL client's execution

2022-09-26 Thread GitBox


wuchong merged PR #20869:
URL: https://github.com/apache/flink/pull/20869


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-ml] yunfengzhou-hub commented on a diff in pull request #158: [FLINK-29409] Add Transformer and Estimator for VarianceThresholdSelector

2022-09-26 Thread GitBox


yunfengzhou-hub commented on code in PR #158:
URL: https://github.com/apache/flink-ml/pull/158#discussion_r980634965


##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/variancethresholdselector/VarianceThresholdSelectorParams.java:
##
@@ -0,0 +1,47 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.feature.variancethresholdselector;
+
+import org.apache.flink.ml.common.param.HasFeaturesCol;
+import org.apache.flink.ml.common.param.HasOutputCol;
+import org.apache.flink.ml.param.DoubleParam;
+import org.apache.flink.ml.param.Param;
+
+/**
+ * * Params of VarianceThresholdSelectorModel.
+ *
+ * @param  The class type of this instance.
+ */
+public interface VarianceThresholdSelectorParams extends HasFeaturesCol, 
HasOutputCol {
+
+Param VARIANCE_THRESHOLD =
+new DoubleParam(
+"varianceThreshold",
+"Param for variance threshold. Features with "
++ "a variance not greater than this threshold will 
be removed.",
+0.0);

Review Comment:
   nit: we can add a `ParamValidators.gtEq(0)` here.



##
flink-ml-lib/src/main/java/org/apache/flink/ml/feature/variancethresholdselector/VarianceThresholdSelectorModel.java:
##
@@ -0,0 +1,172 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.ml.feature.variancethresholdselector;
+
+import org.apache.flink.api.common.functions.RichMapFunction;
+import org.apache.flink.api.common.typeinfo.TypeInformation;
+import org.apache.flink.api.java.typeutils.RowTypeInfo;
+import org.apache.flink.ml.api.Model;
+import org.apache.flink.ml.common.broadcast.BroadcastUtils;
+import org.apache.flink.ml.common.datastream.TableUtils;
+import org.apache.flink.ml.linalg.DenseVector;
+import org.apache.flink.ml.linalg.Vector;
+import org.apache.flink.ml.linalg.Vectors;
+import org.apache.flink.ml.param.Param;
+import org.apache.flink.ml.util.ParamUtils;
+import org.apache.flink.ml.util.ReadWriteUtils;
+import org.apache.flink.streaming.api.datastream.DataStream;
+import org.apache.flink.table.api.Table;
+import org.apache.flink.table.api.bridge.java.StreamTableEnvironment;
+import org.apache.flink.table.api.internal.TableImpl;
+import org.apache.flink.types.Row;
+import org.apache.flink.util.Preconditions;
+
+import org.apache.commons.lang3.ArrayUtils;
+
+import java.io.IOException;
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.Map;
+
+/**
+ * A Model which removes low-variance data using the model data computed by 
{@link
+ * VarianceThresholdSelector}.
+ */
+public class VarianceThresholdSelectorModel
+implements Model,
+
VarianceThresholdSelectorParams {
+
+private final Map, Object> paramMap = new HashMap<>();
+private Table modelDataTable;
+
+public VarianceThresholdSelectorModel() {
+ParamUtils.initializeMapWithDefaultValues(paramMap, this);
+}
+
+@Override
+public VarianceThresholdSelectorModel setModelData(Table... inputs) {
+Preconditions.checkArgument(inputs.length == 1);
+modelDataTable = inputs[0];
+return this;
+}
+
+@Override
+public Table[] getModelData() {
+return new Table[] {modelDataTable};
+}
+
+@Override
+public Map, Object> getParamMap() {
+return paramMap;
+}
+
+@Override
+public void 

[GitHub] [flink-ml] lindong28 commented on pull request #157: [FLINK-28906] Support windowing in AgglomerativeClustering

2022-09-26 Thread GitBox


lindong28 commented on PR #157:
URL: https://github.com/apache/flink-ml/pull/157#issuecomment-1258769768

   Thanks for the PR! LGTM overall. Can you fix the test failures?
   
   @zhipeng93 do you want to take a look at this PR?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-29418) Update flink-shaded

2022-09-26 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29418?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-29418:
---
Labels: pull-request-available  (was: )

> Update flink-shaded
> ---
>
> Key: FLINK-29418
> URL: https://issues.apache.org/jira/browse/FLINK-29418
> Project: Flink
>  Issue Type: Technical Debt
>  Components: BuildSystem / Shaded
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>  Labels: pull-request-available
> Fix For: shaded-1.16
>
>
> Do a round of updates in flink-shaded.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-shaded] zentol opened a new pull request, #112: [FLINK-29418] Everybody gets an upgrade!

2022-09-26 Thread GitBox


zentol opened a new pull request, #112:
URL: https://github.com/apache/flink-shaded/pull/112

   Except guava.
   
   The upgrades were all tested against 1.17-SNAPSHOT. For 1.16 I'm only 
interested in the Jackson upgrade, which has some slight behavior changes 
w.r.t. how the result of the docs generation, but we can handle that.
   
   Also adds Zookeeper 3.7 / 3.8 modules. I haven't tested those; they either 
work for 1.17 or we'll have to do another release anyway. IOW, there's no harm 
in throwing them in now.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] flinkbot commented on pull request #20903: [FLINK-29402][state backends] Add RocksDB options for DirectIO reads …

2022-09-26 Thread GitBox


flinkbot commented on PR #20903:
URL: https://github.com/apache/flink/pull/20903#issuecomment-1258630220

   
   ## CI report:
   
   * 70ab91e25623f474698e153622e52d459eaf0530 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-29402) Add USE_DIRECT_READ configuration parameter for RocksDB

2022-09-26 Thread Donatien (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Donatien updated FLINK-29402:
-
Fix Version/s: 1.16.0
   (was: 1.15.2)
Affects Version/s: 1.16.0
   (was: 1.15.2)
   Labels: Enhancement pull-request-available 
pull_request_available rocksdb  (was: Enhancement pull-request-available 
rocksdb)

> Add USE_DIRECT_READ configuration parameter for RocksDB
> ---
>
> Key: FLINK-29402
> URL: https://issues.apache.org/jira/browse/FLINK-29402
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.16.0
>Reporter: Donatien
>Priority: Not a Priority
>  Labels: Enhancement, pull-request-available, 
> pull_request_available, rocksdb
> Fix For: 1.16.0
>
> Attachments: directIO-performance-comparison.png
>
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> RocksDB allows the use of DirectIO for read operations to bypass the Linux 
> Page Cache. To understand the impact of Linux Page Cache on performance, one 
> can run a heavy workload on a single-tasked Task Manager with a container 
> memory limit identical to the TM process memory. Running this same workload 
> on a TM with no container memory limit will result in better performances but 
> with the host memory exceeding the TM requirement.
> Linux Page Cache are of course useful but can give false results when 
> benchmarking the Managed Memory used by RocksDB. DirectIO is typically 
> enabled for benchmarks on working set estimation [Zwaenepoel et 
> al.|[https://arxiv.org/abs/1702.04323].]
> I propose to add a configuration key allowing users to enable the use of 
> DirectIO for reads thanks to the RocksDB API. This configuration would be 
> disabled by default.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29402) Add USE_DIRECT_READ configuration parameter for RocksDB

2022-09-26 Thread Donatien (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Donatien updated FLINK-29402:
-
Labels: Enhancement pull-request-available rocksdb  (was: Enhancement 
pull-request-available pull_request_available rocksdb)

> Add USE_DIRECT_READ configuration parameter for RocksDB
> ---
>
> Key: FLINK-29402
> URL: https://issues.apache.org/jira/browse/FLINK-29402
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.16.0
>Reporter: Donatien
>Priority: Not a Priority
>  Labels: Enhancement, pull-request-available, rocksdb
> Fix For: 1.16.0
>
> Attachments: directIO-performance-comparison.png
>
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> RocksDB allows the use of DirectIO for read operations to bypass the Linux 
> Page Cache. To understand the impact of Linux Page Cache on performance, one 
> can run a heavy workload on a single-tasked Task Manager with a container 
> memory limit identical to the TM process memory. Running this same workload 
> on a TM with no container memory limit will result in better performances but 
> with the host memory exceeding the TM requirement.
> Linux Page Cache are of course useful but can give false results when 
> benchmarking the Managed Memory used by RocksDB. DirectIO is typically 
> enabled for benchmarks on working set estimation [Zwaenepoel et 
> al.|[https://arxiv.org/abs/1702.04323].]
> I propose to add a configuration key allowing users to enable the use of 
> DirectIO for reads thanks to the RocksDB API. This configuration would be 
> disabled by default.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29402) Add USE_DIRECT_READ configuration parameter for RocksDB

2022-09-26 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-29402:
---
Labels: Enhancement pull-request-available rocksdb  (was: Enhancement 
rocksdb)

> Add USE_DIRECT_READ configuration parameter for RocksDB
> ---
>
> Key: FLINK-29402
> URL: https://issues.apache.org/jira/browse/FLINK-29402
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.15.2
>Reporter: Donatien
>Priority: Not a Priority
>  Labels: Enhancement, pull-request-available, rocksdb
> Fix For: 1.15.2
>
> Attachments: directIO-performance-comparison.png
>
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> RocksDB allows the use of DirectIO for read operations to bypass the Linux 
> Page Cache. To understand the impact of Linux Page Cache on performance, one 
> can run a heavy workload on a single-tasked Task Manager with a container 
> memory limit identical to the TM process memory. Running this same workload 
> on a TM with no container memory limit will result in better performances but 
> with the host memory exceeding the TM requirement.
> Linux Page Cache are of course useful but can give false results when 
> benchmarking the Managed Memory used by RocksDB. DirectIO is typically 
> enabled for benchmarks on working set estimation [Zwaenepoel et 
> al.|[https://arxiv.org/abs/1702.04323].]
> I propose to add a configuration key allowing users to enable the use of 
> DirectIO for reads thanks to the RocksDB API. This configuration would be 
> disabled by default.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] Donaschmi opened a new pull request, #20903: [FLINK-29402][state backends] Add RocksDB options for DirectIO reads …

2022-09-26 Thread GitBox


Donaschmi opened a new pull request, #20903:
URL: https://github.com/apache/flink/pull/20903

   
   
   ## What is the purpose of the change
   
   This pull request adds two configurable RocksDB options enabling the use of 
DirectIO for reads and background flush and compactions.
   
   
   ## Brief change log
   
 - Add `state.backend.rocksdb.use-direct-reads` to enable DirectIO for read 
operations
 - Add `state.backend.rocksdb.use-direct-writes` to enable DirectIO for 
background flush and compactions
   
   ## Verifying this change
   
   This change is already covered by existing tests, such as 
*testConfigurableOptionsFromConfig*.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no 
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: no
 - The S3 file system connector: no

   ## Documentation
   
 - Does this pull request introduce a new feature? yes
 - If yes, how is the feature documented? docs 
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] stevenzwu commented on a diff in pull request #20852: [FLINK-27101][checkpointing][rest] Add restful API to trigger checkpoints

2022-09-26 Thread GitBox


stevenzwu commented on code in PR #20852:
URL: https://github.com/apache/flink/pull/20852#discussion_r980390428


##
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/JobMaster.java:
##
@@ -863,11 +871,6 @@ public CompletableFuture triggerSavepoint(
 return schedulerNG.triggerSavepoint(targetDirectory, cancelJob, 
formatType);
 }
 
-@Override

Review Comment:
   got it. you added a default impl method in the parent interface.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] stevenzwu commented on a diff in pull request #20852: [FLINK-27101][checkpointing][rest] Add restful API to trigger checkpoints

2022-09-26 Thread GitBox


stevenzwu commented on code in PR #20852:
URL: https://github.com/apache/flink/pull/20852#discussion_r980387378


##
flink-runtime/src/main/java/org/apache/flink/runtime/scheduler/SchedulerBase.java:
##
@@ -878,7 +879,8 @@ public CompletableFuture triggerSavepoint(
 }
 
 @Override
-public CompletableFuture triggerCheckpoint() {
+public CompletableFuture triggerCheckpoint(

Review Comment:
   oh. I didn't realize it is needed in the trigger checkpoint code path.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] cfedersp commented on pull request #19861: [hotfix] [docs] Fix use cases for HashMapStateBackend

2022-09-26 Thread GitBox


cfedersp commented on PR #19861:
URL: https://github.com/apache/flink/pull/19861#issuecomment-1258459137

   Theres also a factual error on this page where it states:
   The EmbeddedRocksDBStateBackend holds in-flight data in a RocksDB database 
that is (per default) stored in the TaskManager local data directories.
   
   With minimal RocksDB config:
   `env.setStateBackend(new EmbeddedRocksDBStateBackend());
   env.getCheckpointConfig().setCheckpointStorage("file:///var/data");`
   I find the data on the JobManager. 
   https://user-images.githubusercontent.com/284126/192354886-091dc2ed-6caa-49f1-84c5-675f56767d97.png;>


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] afedulov commented on pull request #20865: [FLINK-14896][connectors/kinesis] Shade and relocate Jackson dependen…

2022-09-26 Thread GitBox


afedulov commented on PR #20865:
URL: https://github.com/apache/flink/pull/20865#issuecomment-1258433441

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Assigned] (FLINK-29423) JobDetails is incorrect OpenAPI spec

2022-09-26 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29423?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler reassigned FLINK-29423:


Assignee: Chesnay Schepler

> JobDetails is incorrect OpenAPI spec 
> -
>
> Key: FLINK-29423
> URL: https://issues.apache.org/jira/browse/FLINK-29423
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation, Runtime / REST
>Affects Versions: 1.16.0
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>
> The JobDetails use custom serialization, but the introspection ignores that 
> and analyzes the class as-is, resulting in various fields being documented 
> that shouldn't be.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29423) JobDetails is incorrect OpenAPI spec

2022-09-26 Thread Chesnay Schepler (Jira)
Chesnay Schepler created FLINK-29423:


 Summary: JobDetails is incorrect OpenAPI spec 
 Key: FLINK-29423
 URL: https://issues.apache.org/jira/browse/FLINK-29423
 Project: Flink
  Issue Type: Bug
  Components: Documentation, Runtime / REST
Affects Versions: 1.16.0
Reporter: Chesnay Schepler


The JobDetails use custom serialization, but the introspection ignores that and 
analyzes the class as-is, resulting in various fields being documented that 
shouldn't be.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] leletan commented on a diff in pull request #20852: [FLINK-27101][checkpointing][rest] Add restful API to trigger checkpoints

2022-09-26 Thread GitBox


leletan commented on code in PR #20852:
URL: https://github.com/apache/flink/pull/20852#discussion_r980354599


##
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/JobMaster.java:
##
@@ -853,6 +855,12 @@ public CompletableFuture 
requestJob(Time timeout) {
 return CompletableFuture.completedFuture(schedulerNG.requestJob());
 }
 
+@Override
+public CompletableFuture triggerCheckpoint(

Review Comment:
   There will be no other `triggerCheckpoint` method in this class, that method 
is implemented in its parent interface so it will not be defined explicitly 
here.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] leletan commented on a diff in pull request #20852: [FLINK-27101][checkpointing][rest] Add restful API to trigger checkpoints

2022-09-26 Thread GitBox


leletan commented on code in PR #20852:
URL: https://github.com/apache/flink/pull/20852#discussion_r980350376


##
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/JobMaster.java:
##
@@ -863,11 +871,6 @@ public CompletableFuture triggerSavepoint(
 return schedulerNG.triggerSavepoint(targetDirectory, cancelJob, 
formatType);
 }
 
-@Override

Review Comment:
   Double checked the code - the backward compatibility should already been 
kept. At this class's parent interface `JobMasterGateway`, this method is 
supported as 
   
   ```
   /**
* Triggers taking a checkpoint of the executed job.
*
* @param timeout for the rpc call
* @return Future which is completed with the checkpoint path once 
completed
*/
   default CompletableFuture triggerCheckpoint(@RpcTimeout final 
Time timeout) {
   return triggerCheckpoint(null, 
timeout).thenApply(CompletedCheckpoint::getExternalPointer);
   }



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] leletan commented on a diff in pull request #20852: [FLINK-27101][checkpointing][rest] Add restful API to trigger checkpoints

2022-09-26 Thread GitBox


leletan commented on code in PR #20852:
URL: https://github.com/apache/flink/pull/20852#discussion_r980350376


##
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/JobMaster.java:
##
@@ -863,11 +871,6 @@ public CompletableFuture triggerSavepoint(
 return schedulerNG.triggerSavepoint(targetDirectory, cancelJob, 
formatType);
 }
 
-@Override

Review Comment:
   The backward compatibility should already been kept. At this class's parent 
interface `JobMasterGateway`, this method is supported as 
   
   ```
   /**
* Triggers taking a checkpoint of the executed job.
*
* @param timeout for the rpc call
* @return Future which is completed with the checkpoint path once 
completed
*/
   default CompletableFuture triggerCheckpoint(@RpcTimeout final 
Time timeout) {
   return triggerCheckpoint(null, 
timeout).thenApply(CompletedCheckpoint::getExternalPointer);
   }



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] stevenzwu commented on a diff in pull request #20852: [FLINK-27101][checkpointing][rest] Add restful API to trigger checkpoints

2022-09-26 Thread GitBox


stevenzwu commented on code in PR #20852:
URL: https://github.com/apache/flink/pull/20852#discussion_r980320317


##
flink-core/src/main/java/org/apache/flink/core/execution/CheckpointBackupType.java:
##
@@ -0,0 +1,52 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.core.execution;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.annotation.PublicEvolving;
+import org.apache.flink.configuration.DescribedEnum;
+import org.apache.flink.configuration.description.InlineElement;
+
+import static org.apache.flink.configuration.description.TextElement.text;
+
+/** Describes the backup type in which a checkpoint should be taken. */
+@PublicEvolving
+public enum CheckpointBackupType implements DescribedEnum {

Review Comment:
   agree that `SnapshotTypeImpl` is not a good choice. `Internal` would be ok 
for me. 
   
   Just to brainstorm more options. what about `CheckpointSnapshotType` and 
`SavepointSnapshotType`?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] reswqa commented on pull request #20902: [FLINK-29419] fix the problem that testHybridFullExchangesRestart hang.

2022-09-26 Thread GitBox


reswqa commented on PR #20902:
URL: https://github.com/apache/flink/pull/20902#issuecomment-1258354359

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-29422) Production tests return/argument types do not take transitivity into account

2022-09-26 Thread Chesnay Schepler (Jira)
Chesnay Schepler created FLINK-29422:


 Summary: Production tests return/argument types do not take 
transitivity into account
 Key: FLINK-29422
 URL: https://issues.apache.org/jira/browse/FLINK-29422
 Project: Flink
  Issue Type: Technical Debt
  Components: API / DataStream
Affects Versions: 1.16.0
Reporter: Chesnay Schepler


In FLINK-29403 I'm marking {{SimpleCondition}} as {{PublicEvolving}}, but the 
production tests reject it:

{code:java}
Architecture Violation [Priority: MEDIUM] - Rule 'Return and argument types of 
methods annotated with @PublicEvolving must be annotated with 
@Public(Evolving).' was violated (1 times):
Sep 26 15:20:12 
org.apache.flink.cep.pattern.conditions.SimpleCondition.filter(java.lang.Object,
 org.apache.flink.cep.pattern.conditions.IterativeCondition$Context): Argument 
leaf type org.apache.flink.cep.pattern.conditions.IterativeCondition$Context 
does not satisfy: reside outside of package 'org.apache.flink..' or reside in 
any package ['..shaded..'] or annotated with @Public or annotated with 
@PublicEvolving or annotated with @Deprecated
{code}

This doesn't make any sense given that {{IterativeCondition}} itself is already 
{{PublicEvolving}} and contains the exact same method.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29413) Make it possible to associate triggered and completed savepoints

2022-09-26 Thread Gyula Fora (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gyula Fora updated FLINK-29413:
---
Fix Version/s: kubernetes-operator-1.3.0

> Make it possible to associate triggered and completed savepoints
> 
>
> Key: FLINK-29413
> URL: https://issues.apache.org/jira/browse/FLINK-29413
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Reporter: Gyula Fora
>Assignee: Gyula Fora
>Priority: Major
> Fix For: kubernetes-operator-1.3.0
>
>
> Currently it is not clear how one would assoicate completed manual savepoints 
> with savepointTriggerNonce-es when using the operator.
> This makes it difficult to track when a savepoint was completed vs when it 
> was abandoned
> One idea would be to add the savepointTriggerNonce to the completed 
> checkpoint info for Manual savepoints.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-29413) Make it possible to associate triggered and completed savepoints

2022-09-26 Thread Gyula Fora (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gyula Fora reassigned FLINK-29413:
--

Assignee: Gyula Fora

> Make it possible to associate triggered and completed savepoints
> 
>
> Key: FLINK-29413
> URL: https://issues.apache.org/jira/browse/FLINK-29413
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Reporter: Gyula Fora
>Assignee: Gyula Fora
>Priority: Major
>
> Currently it is not clear how one would assoicate completed manual savepoints 
> with savepointTriggerNonce-es when using the operator.
> This makes it difficult to track when a savepoint was completed vs when it 
> was abandoned
> One idea would be to add the savepointTriggerNonce to the completed 
> checkpoint info for Manual savepoints.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29413) Make it possible to associate triggered and completed savepoints

2022-09-26 Thread Gyula Fora (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29413?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609572#comment-17609572
 ] 

Gyula Fora commented on FLINK-29413:


We don't really see any other way to retrieve the format info, so it is 
probably valuable to track that also

> Make it possible to associate triggered and completed savepoints
> 
>
> Key: FLINK-29413
> URL: https://issues.apache.org/jira/browse/FLINK-29413
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Reporter: Gyula Fora
>Priority: Major
>
> Currently it is not clear how one would assoicate completed manual savepoints 
> with savepointTriggerNonce-es when using the operator.
> This makes it difficult to track when a savepoint was completed vs when it 
> was abandoned
> One idea would be to add the savepointTriggerNonce to the completed 
> checkpoint info for Manual savepoints.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29413) Make it possible to associate triggered and completed savepoints

2022-09-26 Thread Thomas Weise (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29413?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609561#comment-17609561
 ] 

Thomas Weise commented on FLINK-29413:
--

+1 for tracking the triggering nonce

Would there be any other way to retrieve the savepoint type given that Flink 
does not retain a history beyond current job execution? 

 

> Make it possible to associate triggered and completed savepoints
> 
>
> Key: FLINK-29413
> URL: https://issues.apache.org/jira/browse/FLINK-29413
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Reporter: Gyula Fora
>Priority: Major
>
> Currently it is not clear how one would assoicate completed manual savepoints 
> with savepointTriggerNonce-es when using the operator.
> This makes it difficult to track when a savepoint was completed vs when it 
> was abandoned
> One idea would be to add the savepointTriggerNonce to the completed 
> checkpoint info for Manual savepoints.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] zentol commented on a diff in pull request #20893: [FLINK-29403][cep] Streamline SimpleCondition usage

2022-09-26 Thread GitBox


zentol commented on code in PR #20893:
URL: https://github.com/apache/flink/pull/20893#discussion_r980213947


##
flink-libraries/flink-cep/src/main/java/org/apache/flink/cep/pattern/conditions/SimpleCondition.java:
##
@@ -40,4 +40,13 @@
 public boolean filter(T value, Context ctx) throws Exception {
 return filter(value);
 }
+
+public static  SimpleCondition of(FilterFunction filters) {

Review Comment:
   > replace the `Condition` with the `FilterFunction`
   
   I'm not sure I understand. Which `Condition` to do you want to _replace_?
   
   > other condition implementation should follow this usage
   
   I'm not sure this is needed; AFAICT all other IterativeCondition sub-classes 
are truly internal.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-kubernetes-operator] gaborgsomogyi commented on pull request #386: [FLINK-29376] Set SHUTDOWN_ON_APPLICATION_FINISH to false only on 1.15 and above

2022-09-26 Thread GitBox


gaborgsomogyi commented on PR #386:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/386#issuecomment-1258232704

   cc @gyfora 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-29376) Deployment already exists error if Flink version is not set correctly

2022-09-26 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29376?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-29376:
---
Labels: pull-request-available  (was: )

> Deployment already exists error if Flink version is not set correctly
> -
>
> Key: FLINK-29376
> URL: https://issues.apache.org/jira/browse/FLINK-29376
> Project: Flink
>  Issue Type: Bug
>  Components: Kubernetes Operator
>Affects Versions: kubernetes-operator-1.1.0, kubernetes-operator-1.2.0
>Reporter: Gyula Fora
>Assignee: Gabor Somogyi
>Priority: Major
>  Labels: pull-request-available
> Fix For: kubernetes-operator-1.2.0
>
>
> If the user incorrectly sets a Flink version lower than 1.15 when using 1.15 
> or above there are some strange behaviour currently around cluster shutdown.
> Since we always set SHUTDOWN_ON_APPLICATION_FINISH (in the 
> FlinkConfigBuilder) regardless of version but set the JmDeployStatus to 
> MISSING/READY based on version. It can happen that we set the JmStatus to 
> MISSING when the jm deployment is still running.
> As the observer skips observing UPGRADING CRs it never updates the MISSING 
> status and the deployment logic fails on duplicate deployment.
> An easy fix could be to only set SHUTDOWN_ON_APPLICATION_FINISH for version 
> >= 1.15



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-kubernetes-operator] gaborgsomogyi opened a new pull request, #386: [FLINK-29376] Set SHUTDOWN_ON_APPLICATION_FINISH to false only on 1.15 and above

2022-09-26 Thread GitBox


gaborgsomogyi opened a new pull request, #386:
URL: https://github.com/apache/flink-kubernetes-operator/pull/386

   ## What is the purpose of the change
   
   If the user incorrectly sets a Flink version lower than 1.15 when using 1.15 
or above:
   ```
   image: flink:1.15
 flinkVersion: v1_14
   ```
   then a spec change ends-up in exception and restart loop. In this PR we set 
`SHUTDOWN_ON_APPLICATION_FINISH` to `false` when Flink version is less than 
1.15.
   
   ## Brief change log
   
   * Set `SHUTDOWN_ON_APPLICATION_FINISH` to false only on 1.15 and above
   * Added unit test for all available Flink versions
   
   ## Verifying this change
   
   * New unit test
   * Manually
 * Deploy app where
 `image: flink:1.15 flinkVersion: v1_14`
 * Wait for RUNNING state
 * Change Flink config
 * Double check that Flink is re-deployed and in RUNNING state again
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changes to the `CustomResourceDescriptors`: 
no
 - Core observer or reconciler logic that is regularly executed: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? not applicable
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-kubernetes-operator] gaborgsomogyi commented on a diff in pull request #382: [FLINK-29392] Trigger error when session job is lost without HA

2022-09-26 Thread GitBox


gaborgsomogyi commented on code in PR #382:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/382#discussion_r980187297


##
flink-kubernetes-operator/src/main/java/org/apache/flink/kubernetes/operator/observer/JobStatusObserver.java:
##
@@ -72,22 +81,91 @@ public boolean observe(
 }
 
 if (!clusterJobStatuses.isEmpty()) {
+// There are jobs on the cluster, we filter the ones for this 
resource
 Optional targetJobStatusMessage =
 filterTargetJob(jobStatus, clusterJobStatuses);
+
 if (targetJobStatusMessage.isEmpty()) {
-
jobStatus.setState(org.apache.flink.api.common.JobStatus.RECONCILING.name());
+LOG.warn("No matching jobs found on the cluster");
+ifRunningMoveToReconciling(jobStatus, previousJobStatus);
+// We could list the jobs but cannot find the one for this 
resource
+if (resource instanceof FlinkDeployment) {
+// This should never happen for application clusters, 
there is something wrong
+setUnknownJobError((FlinkDeployment) resource);
+} else {
+ifHaDisabledMarkSessionJobMissing((FlinkSessionJob) 
resource, deployedConfig);
+}
 return false;
 } else {
 updateJobStatus(resource, targetJobStatusMessage.get(), 
deployedConfig);
 }
 ReconciliationUtils.checkAndUpdateStableSpec(resource.getStatus());
 return true;
 } else {
+LOG.debug("No jobs found on the cluster");
+// No jobs found on the cluster, it is possible that the 
jobmanager is still starting up
 ifRunningMoveToReconciling(jobStatus, previousJobStatus);
+
+if (resource instanceof FlinkSessionJob) {

Review Comment:
   Seems like this is not common logic, maybe this can be extracted from here.



##
flink-kubernetes-operator/src/main/java/org/apache/flink/kubernetes/operator/observer/JobStatusObserver.java:
##
@@ -72,22 +81,91 @@ public boolean observe(
 }
 
 if (!clusterJobStatuses.isEmpty()) {
+// There are jobs on the cluster, we filter the ones for this 
resource
 Optional targetJobStatusMessage =
 filterTargetJob(jobStatus, clusterJobStatuses);
+
 if (targetJobStatusMessage.isEmpty()) {
-
jobStatus.setState(org.apache.flink.api.common.JobStatus.RECONCILING.name());
+LOG.warn("No matching jobs found on the cluster");
+ifRunningMoveToReconciling(jobStatus, previousJobStatus);
+// We could list the jobs but cannot find the one for this 
resource
+if (resource instanceof FlinkDeployment) {
+// This should never happen for application clusters, 
there is something wrong
+setUnknownJobError((FlinkDeployment) resource);
+} else {
+ifHaDisabledMarkSessionJobMissing((FlinkSessionJob) 
resource, deployedConfig);
+}
 return false;
 } else {
 updateJobStatus(resource, targetJobStatusMessage.get(), 
deployedConfig);
 }
 ReconciliationUtils.checkAndUpdateStableSpec(resource.getStatus());
 return true;
 } else {
+LOG.debug("No jobs found on the cluster");
+// No jobs found on the cluster, it is possible that the 
jobmanager is still starting up
 ifRunningMoveToReconciling(jobStatus, previousJobStatus);
+
+if (resource instanceof FlinkSessionJob) {
+ifHaDisabledMarkSessionJobMissing((FlinkSessionJob) resource, 
deployedConfig);
+}
 return false;
 }
 }
 
+/**
+ * When HA is disabled the session job will not recover on JM restarts. If 
the JM goes down /
+ * restarted the session job should be marked missing.
+ *
+ * @param sessionJob Flink session job.
+ * @param conf Flink config.
+ */
+private void ifHaDisabledMarkSessionJobMissing(FlinkSessionJob sessionJob, 
Configuration conf) {

Review Comment:
   This is more like a taste question so we may or may not do it. I prefer 
`get/is` and `set` functions and build upper level functionality like
   ```
   if (something) {
 react
   }
   ```
   More re-usable if somebody wants to implement 
`ifTheMoonIsGreenMarkSessionJobMissing` functionality.
   But again, this is taste question so we can leave it like it if you don't 
agree.
   



##
flink-kubernetes-operator/src/main/java/org/apache/flink/kubernetes/operator/observer/JobStatusObserver.java:
##
@@ -72,22 +81,91 @@ public boolean observe(
 }
 
 if (!clusterJobStatuses.isEmpty()) {
+// There are jobs on the cluster, we filter the 

[jira] [Closed] (FLINK-29109) Checkpoint path conflict with stateless upgrade mode

2022-09-26 Thread Gyula Fora (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gyula Fora closed FLINK-29109.
--
Fix Version/s: kubernetes-operator-1.2.0
   Resolution: Fixed

merged to main 6f0914bbd296c9daf2664afe0a77d1df4f2e157e

> Checkpoint path conflict with stateless upgrade mode
> 
>
> Key: FLINK-29109
> URL: https://issues.apache.org/jira/browse/FLINK-29109
> Project: Flink
>  Issue Type: Bug
>  Components: Kubernetes Operator
>Affects Versions: kubernetes-operator-1.1.0
>Reporter: Thomas Weise
>Assignee: Thomas Weise
>Priority: Major
>  Labels: pull-request-available
> Fix For: kubernetes-operator-1.2.0
>
>
> A stateful job with stateless upgrade mode (yes, there are such use cases) 
> fails with checkpoint path conflict due to constant jobId and FLINK-19358 
> (applies to Flink < 1.16x). Since with stateless upgrade mode the checkpoint 
> id resets on restart the job is going to write to previously used locations 
> and fail. The workaround is to rotate the jobId on every redeploy when the 
> upgrade mode is stateless. While this can be worked around externally it is 
> best done in the operator itself because reconciliation resolves when a 
> restart is actually required while rotating jobId externally may trigger 
> unnecessary restarts.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29413) Make it possible to associate triggered and completed savepoints

2022-09-26 Thread Gyula Fora (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29413?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609537#comment-17609537
 ] 

Gyula Fora commented on FLINK-29413:


I guess we could add that too if there is no other way to get that info, [~thw] 
what do you think?

> Make it possible to associate triggered and completed savepoints
> 
>
> Key: FLINK-29413
> URL: https://issues.apache.org/jira/browse/FLINK-29413
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Reporter: Gyula Fora
>Priority: Major
>
> Currently it is not clear how one would assoicate completed manual savepoints 
> with savepointTriggerNonce-es when using the operator.
> This makes it difficult to track when a savepoint was completed vs when it 
> was abandoned
> One idea would be to add the savepointTriggerNonce to the completed 
> checkpoint info for Manual savepoints.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29413) Make it possible to associate triggered and completed savepoints

2022-09-26 Thread Matyas Orhidi (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29413?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609527#comment-17609527
 ] 

Matyas Orhidi commented on FLINK-29413:
---

Would it makes sense to add the format type too, for all cases until 
https://issues.apache.org/jira/browse/FLINK-29322 is fixed?

> Make it possible to associate triggered and completed savepoints
> 
>
> Key: FLINK-29413
> URL: https://issues.apache.org/jira/browse/FLINK-29413
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Reporter: Gyula Fora
>Priority: Major
>
> Currently it is not clear how one would assoicate completed manual savepoints 
> with savepointTriggerNonce-es when using the operator.
> This makes it difficult to track when a savepoint was completed vs when it 
> was abandoned
> One idea would be to add the savepointTriggerNonce to the completed 
> checkpoint info for Manual savepoints.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-kubernetes-operator] mbalassi commented on a diff in pull request #383: [FLINK-29384] Bump snakeyaml version to 1.32

2022-09-26 Thread GitBox


mbalassi commented on code in PR #383:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/383#discussion_r980116122


##
flink-kubernetes-operator/pom.xml:
##
@@ -47,6 +47,20 @@ under the License.
 io.fabric8
 kubernetes-client
 ${fabric8.version}
+
+
+
+org.yaml
+snakeyaml
+
+
+
+
+
+

Review Comment:
   I tried the dependency management route in `flink-kubernetes-parent`, it did 
not seems to work. I see your point, and can give it a try if you prefer. Given 
that it is hopefully just a temporary fix I would suggest to leave it as is.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-29421) Support python 3.10

2022-09-26 Thread Eric Sirianni (Jira)
Eric Sirianni created FLINK-29421:
-

 Summary: Support python 3.10
 Key: FLINK-29421
 URL: https://issues.apache.org/jira/browse/FLINK-29421
 Project: Flink
  Issue Type: Bug
  Components: API / Python
Reporter: Eric Sirianni


The {{apache-flink}} package fails to install on Python 3.10 due to inability 
to compile {{numpy}}

{noformat}
numpy/core/src/multiarray/scalartypes.c.src:3242:12: error: too few 
arguments to function ‘_Py_HashDouble’
 3242 | return 
_Py_HashDouble(npy_half_to_double(((PyHalfScalarObject *)obj)->obval));
  |^~
In file included from 
/home/sirianni/.asdf/installs/python/3.10.6/include/python3.10/Python.h:77,
 from numpy/core/src/multiarray/scalartypes.c.src:3:

/home/sirianni/.asdf/installs/python/3.10.6/include/python3.10/pyhash.h:10:23: 
note: declared here
   10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double);

{noformat}

Numpy issue https://github.com/numpy/numpy/issues/19033

[Mailing list 
thread|https://lists.apache.org/thread/f4r9hjt1l33xf5ngnswszhnls4cxkk52]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-kubernetes-operator] mbalassi commented on a diff in pull request #383: [FLINK-29384] Bump snakeyaml version to 1.32

2022-09-26 Thread GitBox


mbalassi commented on code in PR #383:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/383#discussion_r980109419


##
flink-kubernetes-shaded/pom.xml:
##
@@ -88,6 +103,7 @@ under the License.
 
META-INF/DEPENDENCIES
 META-INF/LICENSE
 META-INF/MANIFEST.MF
+
org/apache/flink/kubernetes/shaded/org/yaml/snakeyaml/**

Review Comment:
   @wangyang0918 
   The NOTICE has this and only this relevant entry:
   ```
 - SnakeYAML (https://bitbucket.org/snakeyaml/snakeyaml) 
org.yaml:snakeyaml:bundle:1.32
   License: Apache License, Version 2.0  
(http://www.apache.org/licenses/LICENSE-2.0.txt)
   ```
   
   @gyfora Because of the relocation, I added comments to make it more clear.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-29420) Add support for Zookeeper 3.7 / 3.8

2022-09-26 Thread Chesnay Schepler (Jira)
Chesnay Schepler created FLINK-29420:


 Summary: Add support for Zookeeper 3.7 / 3.8
 Key: FLINK-29420
 URL: https://issues.apache.org/jira/browse/FLINK-29420
 Project: Flink
  Issue Type: Improvement
  Components: BuildSystem / Shaded, Runtime / Coordination
Reporter: Chesnay Schepler
Assignee: Chesnay Schepler
 Fix For: 1.17.0






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-kubernetes-operator] tweise merged pull request #385: [FLINK-29109] Avoid checkpoint path conflicts (Flink < 1.16)

2022-09-26 Thread GitBox


tweise merged PR #385:
URL: https://github.com/apache/flink-kubernetes-operator/pull/385


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Closed] (FLINK-29414) Setup license checks

2022-09-26 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29414?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler closed FLINK-29414.

Resolution: Fixed

master: 575912de77153544a5e4211f504cc67e5ee776a8

> Setup license checks
> 
>
> Key: FLINK-29414
> URL: https://issues.apache.org/jira/browse/FLINK-29414
> Project: Flink
>  Issue Type: Technical Debt
>  Components: BuildSystem / Shaded
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>  Labels: pull-request-available
> Fix For: shaded-1.16
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-shaded] zentol merged pull request #111: [FLINK-29414] Setup license check

2022-09-26 Thread GitBox


zentol merged PR #111:
URL: https://github.com/apache/flink-shaded/pull/111


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-29416) One or more fetchers have encountered exception

2022-09-26 Thread Weijie Guo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29416?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609511#comment-17609511
 ] 

Weijie Guo commented on FLINK-29416:


It seems that direct memory is not enough, you can try to increase the size of 
taskmanager.memory.task.off-heap.size
 

> One or more fetchers have encountered exception
> ---
>
> Key: FLINK-29416
> URL: https://issues.apache.org/jira/browse/FLINK-29416
> Project: Flink
>  Issue Type: Bug
>Reporter: liuchenhong
>Priority: Major
>
> One machine in the cluster goes offline, most tasks failover, and then it 
> automatically restarts successfully. But the task of writing data to kudu 
> failed to restart. The error is as follows:
>  
> {code:java}
> //代码占位符
> 2022-09-26 16:06:30,634 WARN  org.apache.flink.runtime.taskmanager.Task       
>              [] - KunlunRecLogOutputWindow (11/90)#6 
> (9e23189bdc1ba38b19f922342819dab2) switched from RUNNING to FAILED with 
> failure cause: 
> org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: 
> Error at remote task manager 'bjstream44.jd.163.org/10.196.24.76:39604'.
>     at 
> org.apache.flink.runtime.io.network.netty.CreditBasedPartitionRequestClientHandler.decodeMsg(CreditBasedPartitionRequestClientHandler.java:351)
>     at 
> org.apache.flink.runtime.io.network.netty.CreditBasedPartitionRequestClientHandler.channelRead(CreditBasedPartitionRequestClientHandler.java:240)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
>     at 
> org.apache.flink.runtime.io.network.netty.NettyMessageClientDecoderDelegate.channelRead(NettyMessageClientDecoderDelegate.java:112)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.epoll.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:795)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:480)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:378)
>     at 
> org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
>     at 
> org.apache.flink.shaded.netty4.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
>     at java.lang.Thread.run(Thread.java:745)
> Caused by: 
> org.apache.flink.runtime.io.network.partition.ProducerFailedException: 
> java.lang.RuntimeException: One or more fetchers have encountered exception
>     at 
> org.apache.flink.runtime.io.network.netty.PartitionRequestQueue.writeAndFlushNextMessageIfPossible(PartitionRequestQueue.java:285)
>     at 
> org.apache.flink.runtime.io.network.netty.PartitionRequestQueue.enqueueAvailableReader(PartitionRequestQueue.java:123)
>     at 
> org.apache.flink.runtime.io.network.netty.PartitionRequestQueue.userEventTriggered(PartitionRequestQueue.java:234)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeUserEventTriggered(AbstractChannelHandlerContext.java:346)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeUserEventTriggered(AbstractChannelHandlerContext.java:332)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireUserEventTriggered(AbstractChannelHandlerContext.java:324)
>     at 
> 

[GitHub] [flink-shaded] zentol commented on a diff in pull request #111: [FLINK-29414] Setup license check

2022-09-26 Thread GitBox


zentol commented on code in PR #111:
URL: https://github.com/apache/flink-shaded/pull/111#discussion_r980060247


##
.github/workflows/ci.yml:
##
@@ -5,12 +5,37 @@ on: [push, pull_request]
 jobs:
   build:
 runs-on: ubuntu-latest
+env:
+  MVN_COMMON_OPTIONS: -U -B --no-transfer-progress

Review Comment:
   Yes. AFAIK those are workarounds for older maven versions.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-shaded] MartijnVisser commented on a diff in pull request #111: [FLINK-29414] Setup license check

2022-09-26 Thread GitBox


MartijnVisser commented on code in PR #111:
URL: https://github.com/apache/flink-shaded/pull/111#discussion_r980050857


##
.github/workflows/ci.yml:
##
@@ -5,12 +5,37 @@ on: [push, pull_request]
 jobs:
   build:
 runs-on: ubuntu-latest
+env:
+  MVN_COMMON_OPTIONS: -U -B --no-transfer-progress

Review Comment:
   Did you deliberately omit `MVN_CONNECTION_OPTIONS` ?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] flinkbot commented on pull request #20902: [FLINK-29419] fix the problem that testHybridFullExchangesRestart hang.

2022-09-26 Thread GitBox


flinkbot commented on PR #20902:
URL: https://github.com/apache/flink/pull/20902#issuecomment-1258054514

   
   ## CI report:
   
   * f57a254ef0d910a65628bb709d8030cf7a63f39b UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-29419) HybridShuffle.testHybridFullExchangesRestart hangs

2022-09-26 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29419?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-29419:
---
Labels: pull-request-available test-stability  (was: test-stability)

> HybridShuffle.testHybridFullExchangesRestart hangs
> --
>
> Key: FLINK-29419
> URL: https://issues.apache.org/jira/browse/FLINK-29419
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.17.0
>Reporter: Huang Xingbo
>Priority: Critical
>  Labels: pull-request-available, test-stability
>
> {code:java}
> 2022-09-26T10:56:44.0766792Z Sep 26 10:56:44 "ForkJoinPool-1-worker-25" #27 
> daemon prio=5 os_prio=0 tid=0x7f41a4efa000 nid=0x6d76 waiting on 
> condition [0x7f40ac135000]
> 2022-09-26T10:56:44.0767432Z Sep 26 10:56:44java.lang.Thread.State: 
> WAITING (parking)
> 2022-09-26T10:56:44.0767892Z Sep 26 10:56:44  at sun.misc.Unsafe.park(Native 
> Method)
> 2022-09-26T10:56:44.0768644Z Sep 26 10:56:44  - parking to wait for  
> <0xa0704e18> (a java.util.concurrent.CompletableFuture$Signaller)
> 2022-09-26T10:56:44.0769287Z Sep 26 10:56:44  at 
> java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> 2022-09-26T10:56:44.0769949Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
> 2022-09-26T10:56:44.0770623Z Sep 26 10:56:44  at 
> java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3313)
> 2022-09-26T10:56:44.0771349Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
> 2022-09-26T10:56:44.0772092Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
> 2022-09-26T10:56:44.0772777Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.JobGraphRunningUtil.execute(JobGraphRunningUtil.java:57)
> 2022-09-26T10:56:44.0773534Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.BatchShuffleITCaseBase.executeJob(BatchShuffleITCaseBase.java:115)
> 2022-09-26T10:56:44.0774333Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.HybridShuffleITCase.testHybridFullExchangesRestart(HybridShuffleITCase.java:59)
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=41343=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] reswqa opened a new pull request, #20902: [FLINK-29419] fix the problem that testHybridFullExchangesRestart hang.

2022-09-26 Thread GitBox


reswqa opened a new pull request, #20902:
URL: https://github.com/apache/flink/pull/20902

   
   
   ## What is the purpose of the change
   
   *(For example: This pull request makes task deployment go through the blob 
server, rather than through RPC. That way we avoid re-transferring them on each 
deployment (during recovery).)*
   
   
   ## Brief change log
   
   *(for example:)*
 - *The TaskInfo is stored in the blob store on job creation time as a 
persistent artifact*
 - *Deployments RPC transmits only the blob storage reference*
 - *TaskManagers retrieve the TaskInfo from the blob cache*
   
   
   ## Verifying this change
   
   Please make sure both new and modified tests in this PR follows the 
conventions defined in our code quality guide: 
https://flink.apache.org/contributing/code-style-and-quality-common.html#testing
   
   *(Please pick either of the following options)*
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   *(or)*
   
   This change is already covered by existing tests, such as *(please describe 
tests)*.
   
   *(or)*
   
   This change added tests and can be verified as follows:
   
   *(example:)*
 - *Added integration tests for end-to-end deployment with large payloads 
(100MB)*
 - *Extended integration test for recovery after master (JobManager) 
failure*
 - *Added test that validates that TaskInfo is transferred only once across 
recoveries*
 - *Manually verified the change by running a 4 node cluster with 2 
JobManagers and 4 TaskManagers, a stateful streaming program, and killing one 
JobManager and two TaskManagers during the execution, verifying that recovery 
happens correctly.*
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / no)
 - The serializers: (yes / no / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / no / 
don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / no / don't know)
 - The S3 file system connector: (yes / no / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / no)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ not documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Closed] (FLINK-29415) InitializationFailure when recovering from a checkpoint in Application Mode leads to the cleanup of all HA data

2022-09-26 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29415?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl closed FLINK-29415.
-
Resolution: Invalid

There is a solution for this specific case already with 
{{execution.submit-failed-job-on-application-error=true}} being introduced with 
FLINK-25715 in Flink 1.15 as [~gyfora] pointed out in the ML thread. I'm gonna 
close this issue again.

> InitializationFailure when recovering from a checkpoint in Application Mode 
> leads to the cleanup of all HA data
> ---
>
> Key: FLINK-29415
> URL: https://issues.apache.org/jira/browse/FLINK-29415
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.16.0, 1.17.0, 1.15.2, 1.14.6
>Reporter: Matthias Pohl
>Priority: Major
>
> This issue was raised in the user ML thread [JobManager restarts on job 
> failure|https://lists.apache.org/thread/qkmcty3h4gkkx5g09m19gwqrf8z8d383]. 
> Recovering from a external checkpoint is handled differently than recovering 
> from an internal state (see 
> [Dispatcher#handleJobManagerRunner|https://github.com/apache/flink/blob/41ac1ba13679121f1ddf14b26a36f4f4a3cc73e4/flink-runtime/src/main/java/org/apache/flink/runtime/dispatcher/Dispatcher.java#L651]).
>  For the latter case, we explicitly do a local cleanup (i.e. no HA data is 
> cleaned up). For the case, described in the ML thread, a global cleanup is 
> performed. That's not a problem in session mode where a new job ID is used. 
> The new job ID will result in using a new namespace for the HA data. Data 
> from previous runs are not touched during a cleanup. In Application mode, we 
> use the default job ID `0` which would be reused. In case of a failure, all 
> the HA data will be "namespaced" using the default job id. As a consequence, 
> the related data is cleaned up.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-29412) Connection leak in orc reader

2022-09-26 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29412?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee closed FLINK-29412.

Resolution: Fixed

master: 29fc9adf023a04b02126f50d915feb55fdbc2327
release-0.2: 311ab67dd2d7dcb4a7a18bff1de6a2da5861964e

> Connection leak in orc reader
> -
>
> Key: FLINK-29412
> URL: https://issues.apache.org/jira/browse/FLINK-29412
> Project: Flink
>  Issue Type: Bug
>  Components: Table Store
>Reporter: Jingsong Lee
>Assignee: Jingsong Lee
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: table-store-0.3.0, table-store-0.2.1
>
>
> 1. OrcFileStatsExtractor forget closing reader.
> 2. HadoopReadOnlyFileSystem forget closing fsDataInputStream.
> We need a pocket test to assert all connections are closed. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29419) HybridShuffle.testHybridFullExchangesRestart hangs

2022-09-26 Thread Weijie Guo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29419?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609486#comment-17609486
 ] 

Weijie Guo commented on FLINK-29419:


[~hxbks2ks] Thanks for the feedback. I will check the cause of this problem.

> HybridShuffle.testHybridFullExchangesRestart hangs
> --
>
> Key: FLINK-29419
> URL: https://issues.apache.org/jira/browse/FLINK-29419
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.17.0
>Reporter: Huang Xingbo
>Priority: Critical
>  Labels: test-stability
>
> {code:java}
> 2022-09-26T10:56:44.0766792Z Sep 26 10:56:44 "ForkJoinPool-1-worker-25" #27 
> daemon prio=5 os_prio=0 tid=0x7f41a4efa000 nid=0x6d76 waiting on 
> condition [0x7f40ac135000]
> 2022-09-26T10:56:44.0767432Z Sep 26 10:56:44java.lang.Thread.State: 
> WAITING (parking)
> 2022-09-26T10:56:44.0767892Z Sep 26 10:56:44  at sun.misc.Unsafe.park(Native 
> Method)
> 2022-09-26T10:56:44.0768644Z Sep 26 10:56:44  - parking to wait for  
> <0xa0704e18> (a java.util.concurrent.CompletableFuture$Signaller)
> 2022-09-26T10:56:44.0769287Z Sep 26 10:56:44  at 
> java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> 2022-09-26T10:56:44.0769949Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
> 2022-09-26T10:56:44.0770623Z Sep 26 10:56:44  at 
> java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3313)
> 2022-09-26T10:56:44.0771349Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
> 2022-09-26T10:56:44.0772092Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
> 2022-09-26T10:56:44.0772777Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.JobGraphRunningUtil.execute(JobGraphRunningUtil.java:57)
> 2022-09-26T10:56:44.0773534Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.BatchShuffleITCaseBase.executeJob(BatchShuffleITCaseBase.java:115)
> 2022-09-26T10:56:44.0774333Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.HybridShuffleITCase.testHybridFullExchangesRestart(HybridShuffleITCase.java:59)
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=41343=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-table-store] JingsongLi merged pull request #304: [FLINK-29412] Connection leak in orc reader

2022-09-26 Thread GitBox


JingsongLi merged PR #304:
URL: https://github.com/apache/flink-table-store/pull/304


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-29419) HybridShuffle.testHybridFullExchangesRestart hangs

2022-09-26 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29419?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609482#comment-17609482
 ] 

Huang Xingbo commented on FLINK-29419:
--

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=41345=logs=5c8e7682-d68f-54d1-16a2-a09310218a49=86f654fa-ab48-5c1a-25f4-7e7f6afb9bba

> HybridShuffle.testHybridFullExchangesRestart hangs
> --
>
> Key: FLINK-29419
> URL: https://issues.apache.org/jira/browse/FLINK-29419
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.17.0
>Reporter: Huang Xingbo
>Priority: Critical
>  Labels: test-stability
>
> {code:java}
> 2022-09-26T10:56:44.0766792Z Sep 26 10:56:44 "ForkJoinPool-1-worker-25" #27 
> daemon prio=5 os_prio=0 tid=0x7f41a4efa000 nid=0x6d76 waiting on 
> condition [0x7f40ac135000]
> 2022-09-26T10:56:44.0767432Z Sep 26 10:56:44java.lang.Thread.State: 
> WAITING (parking)
> 2022-09-26T10:56:44.0767892Z Sep 26 10:56:44  at sun.misc.Unsafe.park(Native 
> Method)
> 2022-09-26T10:56:44.0768644Z Sep 26 10:56:44  - parking to wait for  
> <0xa0704e18> (a java.util.concurrent.CompletableFuture$Signaller)
> 2022-09-26T10:56:44.0769287Z Sep 26 10:56:44  at 
> java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> 2022-09-26T10:56:44.0769949Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
> 2022-09-26T10:56:44.0770623Z Sep 26 10:56:44  at 
> java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3313)
> 2022-09-26T10:56:44.0771349Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
> 2022-09-26T10:56:44.0772092Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
> 2022-09-26T10:56:44.0772777Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.JobGraphRunningUtil.execute(JobGraphRunningUtil.java:57)
> 2022-09-26T10:56:44.0773534Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.BatchShuffleITCaseBase.executeJob(BatchShuffleITCaseBase.java:115)
> 2022-09-26T10:56:44.0774333Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.HybridShuffleITCase.testHybridFullExchangesRestart(HybridShuffleITCase.java:59)
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=41343=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29419) HybridShuffle.testHybridFullExchangesRestart hangs

2022-09-26 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29419?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609481#comment-17609481
 ] 

Huang Xingbo commented on FLINK-29419:
--

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=41345=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7

> HybridShuffle.testHybridFullExchangesRestart hangs
> --
>
> Key: FLINK-29419
> URL: https://issues.apache.org/jira/browse/FLINK-29419
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.17.0
>Reporter: Huang Xingbo
>Priority: Critical
>  Labels: test-stability
>
> {code:java}
> 2022-09-26T10:56:44.0766792Z Sep 26 10:56:44 "ForkJoinPool-1-worker-25" #27 
> daemon prio=5 os_prio=0 tid=0x7f41a4efa000 nid=0x6d76 waiting on 
> condition [0x7f40ac135000]
> 2022-09-26T10:56:44.0767432Z Sep 26 10:56:44java.lang.Thread.State: 
> WAITING (parking)
> 2022-09-26T10:56:44.0767892Z Sep 26 10:56:44  at sun.misc.Unsafe.park(Native 
> Method)
> 2022-09-26T10:56:44.0768644Z Sep 26 10:56:44  - parking to wait for  
> <0xa0704e18> (a java.util.concurrent.CompletableFuture$Signaller)
> 2022-09-26T10:56:44.0769287Z Sep 26 10:56:44  at 
> java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> 2022-09-26T10:56:44.0769949Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
> 2022-09-26T10:56:44.0770623Z Sep 26 10:56:44  at 
> java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3313)
> 2022-09-26T10:56:44.0771349Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
> 2022-09-26T10:56:44.0772092Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
> 2022-09-26T10:56:44.0772777Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.JobGraphRunningUtil.execute(JobGraphRunningUtil.java:57)
> 2022-09-26T10:56:44.0773534Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.BatchShuffleITCaseBase.executeJob(BatchShuffleITCaseBase.java:115)
> 2022-09-26T10:56:44.0774333Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.HybridShuffleITCase.testHybridFullExchangesRestart(HybridShuffleITCase.java:59)
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=41343=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29419) HybridShuffle.testHybridFullExchangesRestart hangs

2022-09-26 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29419?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609480#comment-17609480
 ] 

Huang Xingbo commented on FLINK-29419:
--

cc [~Weijie Guo]

> HybridShuffle.testHybridFullExchangesRestart hangs
> --
>
> Key: FLINK-29419
> URL: https://issues.apache.org/jira/browse/FLINK-29419
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.17.0
>Reporter: Huang Xingbo
>Priority: Critical
>  Labels: test-stability
>
> {code:java}
> 2022-09-26T10:56:44.0766792Z Sep 26 10:56:44 "ForkJoinPool-1-worker-25" #27 
> daemon prio=5 os_prio=0 tid=0x7f41a4efa000 nid=0x6d76 waiting on 
> condition [0x7f40ac135000]
> 2022-09-26T10:56:44.0767432Z Sep 26 10:56:44java.lang.Thread.State: 
> WAITING (parking)
> 2022-09-26T10:56:44.0767892Z Sep 26 10:56:44  at sun.misc.Unsafe.park(Native 
> Method)
> 2022-09-26T10:56:44.0768644Z Sep 26 10:56:44  - parking to wait for  
> <0xa0704e18> (a java.util.concurrent.CompletableFuture$Signaller)
> 2022-09-26T10:56:44.0769287Z Sep 26 10:56:44  at 
> java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> 2022-09-26T10:56:44.0769949Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
> 2022-09-26T10:56:44.0770623Z Sep 26 10:56:44  at 
> java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3313)
> 2022-09-26T10:56:44.0771349Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
> 2022-09-26T10:56:44.0772092Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
> 2022-09-26T10:56:44.0772777Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.JobGraphRunningUtil.execute(JobGraphRunningUtil.java:57)
> 2022-09-26T10:56:44.0773534Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.BatchShuffleITCaseBase.executeJob(BatchShuffleITCaseBase.java:115)
> 2022-09-26T10:56:44.0774333Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.HybridShuffleITCase.testHybridFullExchangesRestart(HybridShuffleITCase.java:59)
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=41343=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-27346) [umbrella] Introduce Hive reader for table store

2022-09-26 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee closed FLINK-27346.

Resolution: Fixed

> [umbrella] Introduce Hive reader for table store
> 
>
> Key: FLINK-27346
> URL: https://issues.apache.org/jira/browse/FLINK-27346
> Project: Flink
>  Issue Type: Improvement
>  Components: Table Store
>Reporter: Jingsong Lee
>Assignee: Caizhi Weng
>Priority: Minor
> Fix For: table-store-0.3.0
>
>
> Add the ability of Hive engine to read table of Flink Table Store.
> Currently the typical query engine is still the Hive engine. For storage, the 
> ecology of queries is very important.
> We currently write an InputFormat for Hive, which requires us to:
> - Organize code dependencies to avoid creating dependency conflicts with Hive
> - some necessary refactoring to make the programming interface more generic
> - Write the Hive InputFormat and implement some pushdown



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29419) HybridShuffle.testHybridFullExchangesRestart hangs

2022-09-26 Thread Huang Xingbo (Jira)
Huang Xingbo created FLINK-29419:


 Summary: HybridShuffle.testHybridFullExchangesRestart hangs
 Key: FLINK-29419
 URL: https://issues.apache.org/jira/browse/FLINK-29419
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Network
Affects Versions: 1.17.0
Reporter: Huang Xingbo


{code:java}
2022-09-26T10:56:44.0766792Z Sep 26 10:56:44 "ForkJoinPool-1-worker-25" #27 
daemon prio=5 os_prio=0 tid=0x7f41a4efa000 nid=0x6d76 waiting on condition 
[0x7f40ac135000]
2022-09-26T10:56:44.0767432Z Sep 26 10:56:44java.lang.Thread.State: WAITING 
(parking)
2022-09-26T10:56:44.0767892Z Sep 26 10:56:44at sun.misc.Unsafe.park(Native 
Method)
2022-09-26T10:56:44.0768644Z Sep 26 10:56:44- parking to wait for  
<0xa0704e18> (a java.util.concurrent.CompletableFuture$Signaller)
2022-09-26T10:56:44.0769287Z Sep 26 10:56:44at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
2022-09-26T10:56:44.0769949Z Sep 26 10:56:44at 
java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
2022-09-26T10:56:44.0770623Z Sep 26 10:56:44at 
java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3313)
2022-09-26T10:56:44.0771349Z Sep 26 10:56:44at 
java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
2022-09-26T10:56:44.0772092Z Sep 26 10:56:44at 
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
2022-09-26T10:56:44.0772777Z Sep 26 10:56:44at 
org.apache.flink.test.runtime.JobGraphRunningUtil.execute(JobGraphRunningUtil.java:57)
2022-09-26T10:56:44.0773534Z Sep 26 10:56:44at 
org.apache.flink.test.runtime.BatchShuffleITCaseBase.executeJob(BatchShuffleITCaseBase.java:115)
2022-09-26T10:56:44.0774333Z Sep 26 10:56:44at 
org.apache.flink.test.runtime.HybridShuffleITCase.testHybridFullExchangesRestart(HybridShuffleITCase.java:59)
{code}
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=41343=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-27843) Schema evolution for data file meta

2022-09-26 Thread liwei li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27843?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609470#comment-17609470
 ] 

liwei li commented on FLINK-27843:
--

May i have a try?

> Schema evolution for data file meta
> ---
>
> Key: FLINK-27843
> URL: https://issues.apache.org/jira/browse/FLINK-27843
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table Store
>Reporter: Jingsong Lee
>Priority: Minor
> Fix For: table-store-0.3.0
>
>
> There are quite a few metadata operations on DataFileMeta, such as getting 
> the statistics of each column and the partition of the file.
> We need to evolution to the latest schema based on schemaId when we get this 
> information



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29243) Introduce a COW mode for table store

2022-09-26 Thread liwei li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29243?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609465#comment-17609465
 ] 

liwei li commented on FLINK-29243:
--

i want to have a try, could you gave me the ticket?

> Introduce a COW mode for table store
> 
>
> Key: FLINK-29243
> URL: https://issues.apache.org/jira/browse/FLINK-29243
> Project: Flink
>  Issue Type: New Feature
>  Components: Table Store
>Reporter: Jingsong Lee
>Priority: Major
> Fix For: table-store-0.3.0
>
>
> Currently, we can combine the COW (Copy on write) mode by three options, but 
> the combination may not be optimal and tedious.
> We can introduce a option that specifically turns on the mode of COW.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] flinkbot commented on pull request #20901: [FLINK-26943][table-planner-blink] Support date_add function in sql

2022-09-26 Thread GitBox


flinkbot commented on PR #20901:
URL: https://github.com/apache/flink/pull/20901#issuecomment-1257928071

   
   ## CI report:
   
   * e9ed9084724236b500271fdcf5d1de5aef056496 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-29418) Update flink-shaded

2022-09-26 Thread Chesnay Schepler (Jira)
Chesnay Schepler created FLINK-29418:


 Summary: Update flink-shaded
 Key: FLINK-29418
 URL: https://issues.apache.org/jira/browse/FLINK-29418
 Project: Flink
  Issue Type: Technical Debt
  Components: BuildSystem / Shaded
Reporter: Chesnay Schepler
Assignee: Chesnay Schepler
 Fix For: shaded-1.16


Do a round of updates in flink-shaded.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] y0908105023 commented on pull request #20846: [FLINK-26943] [add date_add function support]

2022-09-26 Thread GitBox


y0908105023 commented on PR #20846:
URL: https://github.com/apache/flink/pull/20846#issuecomment-1257921682

   @snuyanzin i request a new pr: https://github.com/apache/flink/pull/20901. 
thaks for your time.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] y0908105023 commented on pull request #20901: [FLINK-26943][table-planner-blink] Support date_add function in sql

2022-09-26 Thread GitBox


y0908105023 commented on PR #20901:
URL: https://github.com/apache/flink/pull/20901#issuecomment-1257920212

   history pr: https://github.com/apache/flink/pull/20846


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-shaded] zentol commented on a diff in pull request #111: [FLINK-29414] Setup license check

2022-09-26 Thread GitBox


zentol commented on code in PR #111:
URL: https://github.com/apache/flink-shaded/pull/111#discussion_r979915416


##
pom.xml:
##
@@ -51,9 +51,23 @@ under the License.
 
scm:git:https://github.com/apache/flink-shaded.git
 
 
+
+
+
+apache.snapshots.https
+${distMgmtSnapshotsName}
+${distMgmtSnapshotsUrl}
+
+
+
 
 org.apache.flink.shaded
 4.1.70.Final
+

[GitHub] [flink] y0908105023 opened a new pull request, #20901: [FLINK-26943][table-planner-blink] Support date_add function in sql

2022-09-26 Thread GitBox


y0908105023 opened a new pull request, #20901:
URL: https://github.com/apache/flink/pull/20901

   What is the purpose of the change
   Add date_add sql function support
   
   Brief change log
   date_add
   
   add sql function date_add to BuiltInFunction
   
   Verifying this change
   Please make sure both new and modified tests in this PR follows the 
conventions defined in our code quality guide: 
https://flink.apache.org/contributing/code-style-and-quality-common.html#testing
   
   This change is already covered by existing tests, such as (please describe 
tests).
   
   ScalarFunctionsTest#testDateTypeAdd
   
   ScalarFunctionsTest#testDateTypeMinus
   
   Does this pull request potentially affect one of the following parts:
   Dependencies (does it add or upgrade a dependency): (yes / no)
   The public API, i.e., is any changed class annotated with @Public(Evolving): 
(yes / no)
   The serializers: (yes / no / don't know)
   The runtime per-record code paths (performance sensitive): (yes / no / don't 
know)
   Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / no / don't know)
   The S3 file system connector: (yes / no / don't know)
   It does not affect any parts
   
   Documentation
   Does this pull request introduce a new feature? (yes / no)
   If yes, how is the feature documented? (not applicable / docs / JavaDocs / 
not documented)
   No


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-shaded] zentol commented on a diff in pull request #111: [FLINK-29414] Setup license check

2022-09-26 Thread GitBox


zentol commented on code in PR #111:
URL: https://github.com/apache/flink-shaded/pull/111#discussion_r979911314


##
pom.xml:
##
@@ -51,9 +51,23 @@ under the License.
 
scm:git:https://github.com/apache/flink-shaded.git
 
 
+
+
+
+apache.snapshots.https
+${distMgmtSnapshotsName}
+${distMgmtSnapshotsUrl}
+
+
+
 
 org.apache.flink.shaded
 4.1.70.Final
+
+

Review Comment:
   This thing is super jank but I don't have an idea on how to solve it in 
_general_ without potentially breaking _something_ (because we have to either 
_add_ or _remove_ suffixes).



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-shaded] zentol commented on a diff in pull request #111: [FLINK-29414] Setup license check

2022-09-26 Thread GitBox


zentol commented on code in PR #111:
URL: https://github.com/apache/flink-shaded/pull/111#discussion_r979911314


##
pom.xml:
##
@@ -51,9 +51,23 @@ under the License.
 
scm:git:https://github.com/apache/flink-shaded.git
 
 
+
+
+
+apache.snapshots.https
+${distMgmtSnapshotsName}
+${distMgmtSnapshotsUrl}
+
+
+
 
 org.apache.flink.shaded
 4.1.70.Final
+
+

Review Comment:
   This thing is super jank but I don't have an idea on how to solve it in 
_general_ without potentially breaking _something_.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-29414) Setup license checks

2022-09-26 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29414?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-29414:
---
Labels: pull-request-available  (was: )

> Setup license checks
> 
>
> Key: FLINK-29414
> URL: https://issues.apache.org/jira/browse/FLINK-29414
> Project: Flink
>  Issue Type: Technical Debt
>  Components: BuildSystem / Shaded
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>  Labels: pull-request-available
> Fix For: shaded-1.16
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] y0908105023 closed pull request #20846: [FLINK-26943] [add date_add function support]

2022-09-26 Thread GitBox


y0908105023 closed pull request #20846: [FLINK-26943] [add date_add function 
support]
URL: https://github.com/apache/flink/pull/20846


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-shaded] zentol opened a new pull request, #111: [FLINK-29414] Setup license check

2022-09-26 Thread GitBox


zentol opened a new pull request, #111:
URL: https://github.com/apache/flink-shaded/pull/111

   Most of the changes are taken 1:1 from the ES repo.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] y0908105023 commented on a diff in pull request #20846: [FLINK-26943] [add date_add function support]

2022-09-26 Thread GitBox


y0908105023 commented on code in PR #20846:
URL: https://github.com/apache/flink/pull/20846#discussion_r979900275


##
flink-table/flink-table-common/src/main/java/org/apache/flink/table/utils/DateTimeUtils.java:
##
@@ -1400,6 +1401,25 @@ public static String formatUnixTimestamp(long unixtime, 
TimeZone tz) {
 return formatUnixTimestamp(unixtime, TIMESTAMP_FORMAT_STRING, tz);
 }
 
+/** Add offset to string (like '1970-01-01') and return new date For 
String support. */
+public static Integer dateAdd(BinaryStringData binaryStringData, int 
offset) {
+
+if (binaryStringData == null) {
+return null;
+}
+try {
+int days = parseDate(binaryStringData.toString());
+return days + offset;
+} catch (NullPointerException e) {

Review Comment:
   i modify this, add execption log. just reference formatUnixTimestamp,if 
string args is not  legal format,log and return null.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] y0908105023 commented on a diff in pull request #20846: [FLINK-26943] [add date_add function support]

2022-09-26 Thread GitBox


y0908105023 commented on code in PR #20846:
URL: https://github.com/apache/flink/pull/20846#discussion_r979893866


##
flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/codegen/calls/StringCallGen.scala:
##
@@ -190,6 +191,16 @@ object StringCallGen {
 isCharacterString(operands(1).resultType) =>
 methodGen(BuiltInMethods.STRING_TO_DATE_WITH_FORMAT)
 
+  case DATE_ADD
+  if operands.size == 2 &&
+!isCharacterString(operands.head.resultType) =>
+generateToDateAdd(ctx, operands, returnType)
+
+  case DATE_ADD
+  if operands.size == 2 &&
+isCharacterString(operands.head.resultType) =>
+generateToDateAdd(ctx, operands, returnType)
+

Review Comment:
   To support string type and date type. eg: date_add(string type, offset) or 
date_add(date type, offset)



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-29131) Kubernetes operator webhook can use hostPort

2022-09-26 Thread Gyula Fora (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29131?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609443#comment-17609443
 ] 

Gyula Fora commented on FLINK-29131:


I agree changing the helm chart to have a single Deployment to 2 deployments 
(operator + webhook separately) is a quite big change and will affect every 
user who currently uses the webhook.

Before we go any further I think it might make sense to write a proposal on the 
dev mailing list and discuss this openly + have a FLIP afterwards if there is a 
general agreement.

 

> Kubernetes operator webhook can use hostPort
> 
>
> Key: FLINK-29131
> URL: https://issues.apache.org/jira/browse/FLINK-29131
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Affects Versions: kubernetes-operator-1.1.0
>Reporter: Dylan Meissner
>Assignee: Dylan Meissner
>Priority: Minor
>
> When running Flink operator on EKS cluster with Calico networking the 
> control-plane (managed by AWS) cannot reach the webhook. Requests to create 
> Flink resources fail with {_}Address is not allowed{_}.
> When the webhook listens on hostPort the requests to create Flink resources 
> are successful. However, a pod security policy is generally required to allow 
> webhook to listen on such ports.
> To support this scenario with the Helm chart make changes so that we can
>  * Specify a hostPort value for the webhook
>  * Name the port that the webhook listens on
>  * Use the named port in the webhook service
>  * Add a "use" pod security policy verb to cluster role



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-29417) One or more fetchers have encountered exception

2022-09-26 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29417?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser closed FLINK-29417.
--
Resolution: Duplicate

> One or more fetchers have encountered exception
> ---
>
> Key: FLINK-29417
> URL: https://issues.apache.org/jira/browse/FLINK-29417
> Project: Flink
>  Issue Type: Bug
>Reporter: liuchenhong
>Priority: Major
>
> One machine in the cluster goes offline, most tasks failover, and then it 
> automatically restarts successfully. But the task of writing data to kudu 
> failed to restart. The error is as follows:
>  
> {code:java}
> //代码占位符
> 2022-09-26 16:06:30,634 WARN  org.apache.flink.runtime.taskmanager.Task       
>              [] - KunlunRecLogOutputWindow (11/90)#6 
> (9e23189bdc1ba38b19f922342819dab2) switched from RUNNING to FAILED with 
> failure cause: 
> org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: 
> Error at remote task manager '/***:**'.
>     at 
> org.apache.flink.runtime.io.network.netty.CreditBasedPartitionRequestClientHandler.decodeMsg(CreditBasedPartitionRequestClientHandler.java:351)
>     at 
> org.apache.flink.runtime.io.network.netty.CreditBasedPartitionRequestClientHandler.channelRead(CreditBasedPartitionRequestClientHandler.java:240)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
>     at 
> org.apache.flink.runtime.io.network.netty.NettyMessageClientDecoderDelegate.channelRead(NettyMessageClientDecoderDelegate.java:112)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.epoll.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:795)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:480)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:378)
>     at 
> org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
>     at 
> org.apache.flink.shaded.netty4.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
>     at java.lang.Thread.run(Thread.java:745)
> Caused by: 
> org.apache.flink.runtime.io.network.partition.ProducerFailedException: 
> java.lang.RuntimeException: One or more fetchers have encountered exception
>     at 
> org.apache.flink.runtime.io.network.netty.PartitionRequestQueue.writeAndFlushNextMessageIfPossible(PartitionRequestQueue.java:285)
>     at 
> org.apache.flink.runtime.io.network.netty.PartitionRequestQueue.enqueueAvailableReader(PartitionRequestQueue.java:123)
>     at 
> org.apache.flink.runtime.io.network.netty.PartitionRequestQueue.userEventTriggered(PartitionRequestQueue.java:234)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeUserEventTriggered(AbstractChannelHandlerContext.java:346)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeUserEventTriggered(AbstractChannelHandlerContext.java:332)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireUserEventTriggered(AbstractChannelHandlerContext.java:324)
>     at 
> org.apache.flink.shaded.netty4.io.netty.channel.ChannelInboundHandlerAdapter.userEventTriggered(ChannelInboundHandlerAdapter.java:117)
>     at 
> 

[jira] [Commented] (FLINK-29199) Support blue-green deployment type

2022-09-26 Thread Gyula Fora (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17609442#comment-17609442
 ] 

Gyula Fora commented on FLINK-29199:


What I was saying is that cluster-id which must be unique within a namespace is 
currently tied to the name of the FlinkDeployment.

If you create 2 flinkdeployments, they will have different names so different 
cluster-ids. But since we cannot deploy 2 jobs using the same cluster-id for a 
single FlinkDeployment we would have to completely rework this logic whihc 
would have many implications to last-state recovery , HA etc.

> Support blue-green deployment type
> --
>
> Key: FLINK-29199
> URL: https://issues.apache.org/jira/browse/FLINK-29199
> Project: Flink
>  Issue Type: New Feature
>  Components: Kubernetes Operator
> Environment: Kubernetes
>Reporter: Oleg Vorobev
>Priority: Minor
>
> Are there any plans to support blue-green deployment/rollout mode similar to 
> *BlueGreen* in the 
> [flinkk8soperator|https://github.com/lyft/flinkk8soperator] to avoid downtime 
> while updating?
> The idea is to run a new version in parallel with an old one and remove the 
> old one only after the stability condition of the new one is satisfied (like 
> in 
> [rollbacks|https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-release-1.1/docs/custom-resource/job-management/#application-upgrade-rollbacks-experimental]).
> For stateful apps with {*}upgradeMode: savepoint{*}, this means: not 
> cancelling an old job after creating a savepoint -> starting new job from 
> that savepoint -> waiting for it to become running/one successful 
> checkpoint/timeout or something else -> cancelling and removing old job.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


  1   2   3   >