[jira] [Updated] (FLINK-33206) Verify the existence of hbase table before read/write

2023-10-09 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33206?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-33206:
---
Labels: pull-request-available stale-blocker  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as a 
Blocker but is unassigned and neither itself nor its Sub-Tasks have been 
updated for 1 days. I have gone ahead and marked it "stale-blocker". If this 
ticket is a Blocker, please either assign yourself or give an update. 
Afterwards, please remove the label or in 7 days the issue will be 
deprioritized.


> Verify the existence of hbase table before read/write
> -
>
> Key: FLINK-33206
> URL: https://issues.apache.org/jira/browse/FLINK-33206
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / HBase
>Affects Versions: hbase-3.0.1
>Reporter: tanjialiang
>Priority: Blocker
>  Labels: pull-request-available, stale-blocker
> Attachments: image-2023-10-08-16-54-05-917.png
>
>
> Currently, we do not verify the existence of hbase table before read/write, 
> and the error would make the user confused.
> The `HBaseSinkFunction` throws `TableNotFoundException` when do flush.
> The `inputFormat` throws not obvious enough.
> !image-2023-10-08-16-54-05-917.png!
> So i think we should verify the existence of hbase table when call `open` 
> function.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32714) JDBC: Add dialect for OceanBase database

2023-10-09 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32714?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32714:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> JDBC: Add dialect for OceanBase database
> 
>
> Key: FLINK-32714
> URL: https://issues.apache.org/jira/browse/FLINK-32714
> Project: Flink
>  Issue Type: New Feature
>Reporter: He Wang
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> OceanBase is a distributed relational database, the community edition of 
> OceanBase is open sourced at [https://github.com/oceanbase/oceanbase.]
> The enterprise edition of OceanBase is compatible with MySql and Oracle, 
> which means we can reuse almost all the dialect rules. 
> The difference from other databases is that we must provide the compatibility 
> mode firstly, then the connector can determine which dialect to use, so a 
> startup option like 'compatible-mode'  is needed.
> A dialect implementation for OceanBase is like below: 
> {code:java}
> package org.apache.flink.connector.jdbc.databases.oceanbase;
> import org.apache.flink.connector.jdbc.converter.JdbcRowConverter;
> import org.apache.flink.connector.jdbc.databases.mysql.dialect.MySqlDialect;
> import org.apache.flink.connector.jdbc.databases.oracle.dialect.OracleDialect;
> import org.apache.flink.connector.jdbc.dialect.AbstractDialect;
> import org.apache.flink.table.types.logical.LogicalTypeRoot;
> import org.apache.flink.table.types.logical.RowType;
> import javax.annotation.Nonnull;
> import java.util.Optional;
> import java.util.Set;
> /** JDBC dialect for OceanBase. */
> public class OceanBaseDialect extends AbstractDialect {
> private static final long serialVersionUID = 1L;
> private final AbstractDialect dialect;
> public OceanBaseDialect(@Nonnull String compatibleMode) {
> switch (compatibleMode.toLowerCase()) {
> case "mysql":
> this.dialect = new MySqlDialect();
> break;
> case "oracle":
> this.dialect = new OracleDialect();
> break;
> default:
> throw new IllegalArgumentException(
> "Unsupported compatible mode: " + compatibleMode);
> }
> }
> @Override
> public String dialectName() {
> return "OceanBase";
> }
> @Override
> public Optional defaultDriverName() {
> return Optional.of("com.oceanbase.jdbc.Driver");
> }
> @Override
> public Set supportedTypes() {
> return dialect.supportedTypes();
> }
> @Override
> public JdbcRowConverter getRowConverter(RowType rowType) {
> return dialect.getRowConverter(rowType);
> }
> @Override
> public String getLimitClause(long limit) {
> return dialect.getLimitClause(limit);
> }
> @Override
> public String quoteIdentifier(String identifier) {
> return dialect.quoteIdentifier(identifier);
> }
> @Override
> public Optional getUpsertStatement(
> String tableName, String[] fieldNames, String[] conditionFields) {
> return dialect.getUpsertStatement(tableName, fieldNames, 
> conditionFields);
> }
> @Override
> public Optional timestampPrecisionRange() {
> return dialect.timestampPrecisionRange();
> }
> @Override
> public Optional decimalPrecisionRange() {
> return dialect.decimalPrecisionRange();
> }
> @Override
> public String appendDefaultUrlProperties(String url) {
> return dialect.appendDefaultUrlProperties(url);
> }
> }
>  {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32720) Add GENERATE_SERIES support in SQL & Table API

2023-10-09 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32720?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32720:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Add GENERATE_SERIES support in SQL & Table API
> --
>
> Key: FLINK-32720
> URL: https://issues.apache.org/jira/browse/FLINK-32720
> Project: Flink
>  Issue Type: Improvement
>Reporter: Hanyu Zheng
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> GENERATE_SERIES Function
> Description
> Constructs an array of values between {{start}} and {{{}end{}}}, inclusive.
> Parameters {{start}} and {{end}} can be an {{INT}} or {{{}BIGINT{}}}.
> {{{}step{}}}, if supplied, specifies the step size. The step can be positive 
> or negative. If not supplied, {{step}} defaults to {{{}1{}}}. Parameter 
> {{step}} must be an {{{}INT{}}}.
> Syntax
> The syntax for the GENERATE_SERIES function is:
> {code:java}
> GENERATE_SERIES(start, end)
> GENERATE_SERIES(start, end, step){code}
>  
> Example
> Let's look at some  function examples and explore how to use the SPLIT 
> function.
> For example:
>  
> {code:java}
> SELECT GENERATE_SERISE(1, 5);
> Result: [1,2,3,4,5]
> SELECT GENERATE_SERISE(0, 10, 2); 
> Result: [0, 2, 4, 6, 8, 10] {code}
> see also:
> 1.PostgreSQL: PostgreSQL offers a function called {{generate_series}} which 
> generates a set of contiguous integers from a start to an end value. An 
> optional 'step' parameter is available to specify the increment between each 
> integer.
> https://www.postgresql.org/docs/current/functions-srf.html
> 2.ksqlDB: As you mentioned, ksqlDB provides a function called 
> {{GENERATE_SERIES}} that generates a series of numbers, starting from a given 
> start value, incrementing each time by a step value, until it reaches or 
> exceeds a given end value.
> https://docs.ksqldb.io/en/latest/developer-guide/ksqldb-reference/scalar-functions/#generate_series
> 3.BigQuery: BigQuery has a function called {{GENERATE_ARRAY}} that generates 
> an array consisting of integers from the start value to the end value, with 
> each integer incremented by the step value. You can find more details in the 
> https://cloud.google.com/bigquery/docs/reference/standard-sql/array_functions#generate_array
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33207) Scan hbase table will throw error when table is empty

2023-10-09 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33207?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-33207:
---
Labels: pull-request-available stale-blocker  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as a 
Blocker but is unassigned and neither itself nor its Sub-Tasks have been 
updated for 1 days. I have gone ahead and marked it "stale-blocker". If this 
ticket is a Blocker, please either assign yourself or give an update. 
Afterwards, please remove the label or in 7 days the issue will be 
deprioritized.


> Scan hbase table will throw error when table is empty
> -
>
> Key: FLINK-33207
> URL: https://issues.apache.org/jira/browse/FLINK-33207
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase
>Affects Versions: hbase-3.0.1
>Reporter: tanjialiang
>Priority: Blocker
>  Labels: pull-request-available, stale-blocker
>
> When i scan the empty hbase table, it will throw an error when 
> createInputSplits, we should return empty split instead.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32815) Add HASHCODE support in Table API

2023-10-08 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32815?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32815:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Add HASHCODE support in Table API
> -
>
> Key: FLINK-32815
> URL: https://issues.apache.org/jira/browse/FLINK-32815
> Project: Flink
>  Issue Type: Improvement
>Reporter: Hanyu Zheng
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> *This is an implementation of HASHCODE internal function*
> The {{hashcode}} function generates a hash code for a given input value, 
> including support for computing hash values of binary data types. It creates 
> a unique integer that represents the value passed to the function.
> *Brief change log*
>  * {{HASHCODE}} for Table API 
> *Syntax:*
> {code:java}
> HASHCODE(value){code}
> *Arguments:*
>  * value: the value to be hashed.
> *Returns:* The function returns a unique integer representing the hash code 
> of the value. If the input argument is NULL, the function returns NULL.
> Because it is an internal function, so it will not support sql anymore.
> *see also:*
> Java: 
> [https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html#hashCode--]
> Python: [https://docs.python.org/3/library/functions.html#hash]
> C#: [https://docs.microsoft.com/en-us/dotnet/api/system.object.gethashcode]
> SQL Server: 
> [https://docs.microsoft.com/en-us/sql/t-sql/functions/checksum-transact-sql]
> MySQL: 
> [https://dev.mysql.com/doc/refman/8.0/en/encryption-functions.html#function_md5]
> PostgreSQL: [https://www.postgresql.org/docs/current/pgcrypto-hash.html]
> Oracle: 
> [https://docs.oracle.com/en/database/oracle/oracle-database/19/sqlrf/ORA-HASH.html]
> Google Cloud BigQuery: 
> [https://cloud.google.com/bigquery/docs/reference/standard-sql/functions-and-operators#farm_fingerprint]
> AWS Redshift: [https://docs.aws.amazon.com/redshift/latest/dg/MD5.html]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32722) 239 exit code in flink-runtime

2023-10-08 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32722?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32722:
---
  Labels: auto-deprioritized-major test-stability  (was: stale-major 
test-stability)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> 239 exit code in flink-runtime
> --
>
> Key: FLINK-32722
> URL: https://issues.apache.org/jira/browse/FLINK-32722
> Project: Flink
>  Issue Type: Bug
>  Components: Test Infrastructure
>Affects Versions: 1.16.2
>Reporter: Matthias Pohl
>Priority: Minor
>  Labels: auto-deprioritized-major, test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=51852=logs=d89de3df-4600-5585-dadc-9bbc9a5e661c=be5a4b15-4b23-56b1-7582-795f58a645a2=8418
> {code:java}
> Aug 01 01:03:49 [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-surefire-plugin:3.0.0-M5:test (default-test) 
> on project flink-runtime: There are test failures.
> Aug 01 01:03:49 [ERROR] 
> Aug 01 01:03:49 [ERROR] Please refer to 
> /__w/2/s/flink-runtime/target/surefire-reports for the individual test 
> results.
> Aug 01 01:03:49 [ERROR] Please refer to dump files (if any exist) 
> [date].dump, [date]-jvmRun[N].dump and [date].dumpstream.
> Aug 01 01:03:49 [ERROR] ExecutionException The forked VM terminated without 
> properly saying goodbye. VM crash or System.exit called?
> Aug 01 01:03:49 [ERROR] Command was /bin/sh -c cd /__w/2/s/flink-runtime && 
> /usr/lib/jvm/java-8-openjdk-amd64/jre/bin/java -XX:+UseG1GC -Xms256m -Xmx768m 
> -jar 
> /__w/2/s/flink-runtime/target/surefire/surefirebooter1803120121827605294.jar 
> /__w/2/s/flink-runtime/target/surefire 2023-08-01T00-58-17_520-jvmRun1 
> surefire9107652818401825168tmp surefire_261701267003130520249tmp
> Aug 01 01:03:49 [ERROR] Error occurred in starting fork, check output in log
> Aug 01 01:03:49 [ERROR] Process Exit Code: 239
> Aug 01 01:03:49 [ERROR] 
> org.apache.maven.surefire.booter.SurefireBooterForkException: 
> ExecutionException The forked VM terminated without properly saying goodbye. 
> VM crash or System.exit called?
> Aug 01 01:03:49 [ERROR] Command was /bin/sh -c cd /__w/2/s/flink-runtime && 
> /usr/lib/jvm/java-8-openjdk-amd64/jre/bin/java -XX:+UseG1GC -Xms256m -Xmx768m 
> -jar 
> /__w/2/s/flink-runtime/target/surefire/surefirebooter1803120121827605294.jar 
> /__w/2/s/flink-runtime/target/surefire 2023-08-01T00-58-17_520-jvmRun1 
> surefire9107652818401825168tmp surefire_261701267003130520249tmp
> Aug 01 01:03:49 [ERROR] Error occurred in starting fork, check output in log
> Aug 01 01:03:49 [ERROR] Process Exit Code: 239
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.booterclient.ForkStarter.awaitResultsDone(ForkStarter.java:532)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.booterclient.ForkStarter.runSuitesForkOnceMultiple(ForkStarter.java:405)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:321)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:266)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1314)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1159)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:932)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:132)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:208)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:153)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:145)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:116)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:80)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build(SingleThreadedBuilder.java:51)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:120)
> Aug 01 01:03:49 [ERROR] at 
> 

[jira] [Updated] (FLINK-32711) Type mismatch when proctime function used as parameter

2023-10-08 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32711:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Type mismatch when proctime function used as parameter
> --
>
> Key: FLINK-32711
> URL: https://issues.apache.org/jira/browse/FLINK-32711
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.18.0
>Reporter: Aitozi
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> reproduce case:
> {code:sql}
> SELECT TYPEOF(PROCTIME())
> {code}
> this query will fail with 
> org.apache.flink.table.planner.codegen.CodeGenException: Mismatch of 
> function's argument data type 'TIMESTAMP_LTZ(3) NOT NULL' and actual argument 
> type 'TIMESTAMP_LTZ(3)'.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32665) Support read null value for csv format

2023-10-08 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32665?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32665:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Support read null value for csv format
> --
>
> Key: FLINK-32665
> URL: https://issues.apache.org/jira/browse/FLINK-32665
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.18.0
>Reporter: Fang Yong
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> when there is null column in a file with csv format, it will throw exception 
> when flink job try to parse these data



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-9631) use Files.createDirectories instead of directory.mkdirs

2023-10-07 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-9631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-9631:
--
Labels: auto-deprioritized-minor auto-unassigned stale-assigned  (was: 
auto-deprioritized-minor auto-unassigned)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> use Files.createDirectories instead of directory.mkdirs
> ---
>
> Key: FLINK-9631
> URL: https://issues.apache.org/jira/browse/FLINK-9631
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.4.2, 1.5.0
> Environment: flink1.4
> jdk1.8 latest
> linux 2.6
>Reporter: makeyang
>Assignee: Yangyang ZHANG
>Priority: Not a Priority
>  Labels: auto-deprioritized-minor, auto-unassigned, stale-assigned
>
> job can't be run due to below exception:
> {color:#6a8759}Could not create RocksDB data directory{color}
>  but with this exception, I can't tell exactly why.
> so I suggest Files.createDirectories which throw exception be used rather 
> than File.mkdirs
>  
> I have some more suggestions:
>  # should we use Files.createDirectories to relpace File.mkdirs?
>  # each time task manager throw exception to jobmanager, should IP+nodeId be 
> contained in exception, which means we should define more flink exception 
> which is used to wrap other exceptions such as jdk exceptions?



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32293) Support vector with long index

2023-10-06 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32293:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Support vector with long index
> --
>
> Key: FLINK-32293
> URL: https://issues.apache.org/jira/browse/FLINK-32293
> Project: Flink
>  Issue Type: New Feature
>  Components: Library / Machine Learning
>Reporter: Zhipeng Zhang
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> Currently in Flink ML, we only support sparse and dense vector with `int` as 
> index and `double` as value.
>  
> However, there are real-world cases that the index of a vector could exceed 
> the range of `INT.MAX`. Thus we need to support vector with `long` index.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32732) auto offset reset should be exposed to user

2023-10-06 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32732?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32732:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> auto offset reset should be exposed to user
> ---
>
> Key: FLINK-32732
> URL: https://issues.apache.org/jira/browse/FLINK-32732
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Affects Versions: 1.16.1
>Reporter: xiaogang zhou
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> {code:java}
> // code placeholder
> maybeOverride(
> ConsumerConfig.AUTO_OFFSET_RESET_CONFIG,
> 
> startingOffsetsInitializer.getAutoOffsetResetStrategy().name().toLowerCase(),
> true); {code}
> now flink override the auto.offset.reset with the scan.startup.mode config, 
> and user's explicit config does not take effect. I think maybe we should 
> expose this to customer?
>  
> I think after consuming kafka records from earliest to latest, the 
> scan.startup.mode should no longer influence the kafka scan behave. So I 
> suggest change the override to false.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31615) Fix some parts forgot to translate in "Table API" page of "Table API & SQL"

2023-10-06 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31615?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31615:
---
  Labels: auto-deprioritized-minor chinese-translation  (was: 
chinese-translation stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Fix some parts forgot to translate in "Table API" page of "Table API & SQL" 
> 
>
> Key: FLINK-31615
> URL: https://issues.apache.org/jira/browse/FLINK-31615
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Hang Ruan
>Priority: Not a Priority
>  Labels: auto-deprioritized-minor, chinese-translation
>
> The query_state_warning in "Table API" page of "Table API & SQL"  is still in 
> English. And some comments in codes are in English.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33020) OpensearchSinkTest.testAtLeastOnceSink timed out

2023-10-05 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33020?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-33020:
---
Labels: pull-request-available stale-blocker  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as a 
Blocker but is unassigned and neither itself nor its Sub-Tasks have been 
updated for 1 days. I have gone ahead and marked it "stale-blocker". If this 
ticket is a Blocker, please either assign yourself or give an update. 
Afterwards, please remove the label or in 7 days the issue will be 
deprioritized.


> OpensearchSinkTest.testAtLeastOnceSink timed out
> 
>
> Key: FLINK-33020
> URL: https://issues.apache.org/jira/browse/FLINK-33020
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Opensearch
>Affects Versions: opensearch-1.0.2
>Reporter: Martijn Visser
>Priority: Blocker
>  Labels: pull-request-available, stale-blocker
>
> https://github.com/apache/flink-connector-opensearch/actions/runs/6061205003/job/16446139552#step:13:1029
> {code:java}
> Error:  Tests run: 9, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 9.837 
> s <<< FAILURE! - in 
> org.apache.flink.streaming.connectors.opensearch.OpensearchSinkTest
> Error:  
> org.apache.flink.streaming.connectors.opensearch.OpensearchSinkTest.testAtLeastOnceSink
>   Time elapsed: 5.022 s  <<< ERROR!
> java.util.concurrent.TimeoutException: testAtLeastOnceSink() timed out after 
> 5 seconds
>   at 
> org.junit.jupiter.engine.extension.TimeoutInvocation.createTimeoutException(TimeoutInvocation.java:70)
>   at 
> org.junit.jupiter.engine.extension.TimeoutInvocation.proceed(TimeoutInvocation.java:59)
>   at 
> org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149)
>   at 
> org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140)
>   at 
> org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84)
>   at 
> org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115)
>   at 
> org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105)
>   at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106)
>   at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64)
>   at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45)
>   at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37)
>   at 
> org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104)
>   at 
> org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98)
>   at 
> org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$7(TestMethodTestDescriptor.java:214)
>   at 
> org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
>   at 
> org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:210)
>   at 
> org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:135)
>   at 
> org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:66)
>   at 
> org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:151)
>   at 
> org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
>   at 
> org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
>   at 
> org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
>   at 
> org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
>   at 
> org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
>   at 
> org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
>   at 
> org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
>   at java.base/java.util.ArrayList.forEach(ArrayList.java:1541)
>   at 
> 

[jira] [Updated] (FLINK-32667) Use standalone store and embedding writer for jobs with no-restart-strategy in session cluster

2023-10-05 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32667?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32667:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Use standalone store and embedding writer for jobs with no-restart-strategy 
> in session cluster
> --
>
> Key: FLINK-32667
> URL: https://issues.apache.org/jira/browse/FLINK-32667
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Coordination
>Affects Versions: 1.18.0
>Reporter: Fang Yong
>Assignee: Fang Yong
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> When a flink session cluster use zk or k8s high availability service, it will 
> store jobs in zk or ConfigMap. When we submit flink olap jobs to the session 
> cluster, they always turn off restart strategy. These jobs with 
> no-restart-strategy should not be stored in zk or ConfigMap in k8s



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33116) CliClientTest.testCancelExecutionInteractiveMode fails with NPE on AZP

2023-10-03 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33116?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-33116:
---
Labels: stale-critical test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Critical but is unassigned and neither itself nor its Sub-Tasks have been 
updated for 14 days. I have gone ahead and marked it "stale-critical". If this 
ticket is critical, please either assign yourself or give an update. 
Afterwards, please remove the label or in 7 days the issue will be 
deprioritized.


> CliClientTest.testCancelExecutionInteractiveMode fails with NPE on AZP
> --
>
> Key: FLINK-33116
> URL: https://issues.apache.org/jira/browse/FLINK-33116
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.19.0
>Reporter: Sergey Nuyanzin
>Priority: Critical
>  Labels: stale-critical, test-stability
>
> This build 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=53309=logs=ce3801ad-3bd5-5f06-d165-34d37e757d90=5e4d9387-1dcc-5885-a901-90469b7e6d2f=12264
> fails as
> {noformat}
> Sep 18 02:26:15   at 
> org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:86)
> Sep 18 02:26:15   at 
> org.junit.jupiter.engine.execution.InterceptingExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(InterceptingExecutableInvoker.java:103)
> Sep 18 02:26:15   at 
> org.junit.jupiter.engine.execution.InterceptingExecutableInvoker.lambda$invoke$0(InterceptingExecutableInvoker.java:93)
> Sep 18 02:26:15   at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106)
> Sep 18 02:26:15   at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64)
> Sep 18 02:26:15   at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45)
> ...
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32471) IS_NOT_NULL can add to SUITABLE_FILTER_TO_PUSH

2023-10-03 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32471?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32471:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> IS_NOT_NULL can add to SUITABLE_FILTER_TO_PUSH
> --
>
> Key: FLINK-32471
> URL: https://issues.apache.org/jira/browse/FLINK-32471
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner
>Reporter: grandfisher
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> According to FLINK-31273:
> The reason for the error is that other filters conflict with IS_NULL, but in 
> fact it won't conflict with IS_NOT_NULL, because operators in 
> SUITABLE_FILTER_TO_PUSH  such as 'SqlKind.GREATER_THAN'  has an implicit 
> filter 'IS_NOT_NULL' according to SQL Semantics.
>  
> So we think it is feasible to add  IS_NOT_NULL to the SUITABLE_FILTER_TO_PUSH 
> list.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33115) AbstractHadoopRecoverableWriterITCase is hanging with timeout on AZP

2023-10-03 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33115?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-33115:
---
Labels: stale-critical test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Critical but is unassigned and neither itself nor its Sub-Tasks have been 
updated for 14 days. I have gone ahead and marked it "stale-critical". If this 
ticket is critical, please either assign yourself or give an update. 
Afterwards, please remove the label or in 7 days the issue will be 
deprioritized.


> AbstractHadoopRecoverableWriterITCase is hanging with timeout on AZP
> 
>
> Key: FLINK-33115
> URL: https://issues.apache.org/jira/browse/FLINK-33115
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.18.0, 1.19.0
>Reporter: Sergey Nuyanzin
>Priority: Critical
>  Labels: stale-critical, test-stability
>
> This build 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=53281=logs=4eda0b4a-bd0d-521a-0916-8285b9be9bb5=2ff6d5fa-53a6-53ac-bff7-fa524ea361a9=14239
> is failing as
> {noformat}
> Sep 15 11:33:02 
> ==
> Sep 15 11:33:02 Process produced no output for 900 seconds.
> Sep 15 11:33:02 
> ==
> ...
> Sep 15 11:33:03   at 
> java.io.DataInputStream.read(DataInputStream.java:149)
> Sep 15 11:33:03   at 
> org.apache.flink.runtime.fs.hdfs.HadoopDataInputStream.read(HadoopDataInputStream.java:96)
> Sep 15 11:33:03   at 
> sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:284)
> Sep 15 11:33:03   at 
> sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:326)
> Sep 15 11:33:03   at sun.nio.cs.StreamDecoder.read(StreamDecoder.java:178)
> Sep 15 11:33:03   - locked <0xbfa7a760> (a 
> java.io.InputStreamReader)
> Sep 15 11:33:03   at 
> java.io.InputStreamReader.read(InputStreamReader.java:184)
> Sep 15 11:33:03   at java.io.BufferedReader.fill(BufferedReader.java:161)
> Sep 15 11:33:03   at 
> java.io.BufferedReader.readLine(BufferedReader.java:324)
> Sep 15 11:33:03   - locked <0xbfa7a760> (a 
> java.io.InputStreamReader)
> Sep 15 11:33:03   at 
> java.io.BufferedReader.readLine(BufferedReader.java:389)
> Sep 15 11:33:03   at 
> org.apache.flink.runtime.fs.hdfs.AbstractHadoopRecoverableWriterITCase.getContentsOfFile(AbstractHadoopRecoverableWriterITCase.java:387)
> Sep 15 11:33:03   at 
> org.apache.flink.runtime.fs.hdfs.AbstractHadoopRecoverableWriterITCase.testResumeAfterMultiplePersist(AbstractHadoopRecoverableWriterITCase.java:377)
> Sep 15 11:33:03   at 
> org.apache.flink.runtime.fs.hdfs.AbstractHadoopRecoverableWriterITCase.testResumeAfterMultiplePersistWithMultiPartUploads(AbstractHadoopRecoverableWriterITCase.java:330)
> Sep 15 11:33:03   at 
> org.apache.flink.runtime.fs.hdfs.AbstractHadoopRecoverableWri
> ...
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33114) RightOuterJoinTaskTest.testCancelRightOuterJoinTaskWhileBuilding:232 Task thread did not finish within 60 seconds

2023-10-03 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33114?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-33114:
---
Labels: stale-critical test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Critical but is unassigned and neither itself nor its Sub-Tasks have been 
updated for 14 days. I have gone ahead and marked it "stale-critical". If this 
ticket is critical, please either assign yourself or give an update. 
Afterwards, please remove the label or in 7 days the issue will be 
deprioritized.


> RightOuterJoinTaskTest.testCancelRightOuterJoinTaskWhileBuilding:232 Task 
> thread did not finish within 60 seconds
> -
>
> Key: FLINK-33114
> URL: https://issues.apache.org/jira/browse/FLINK-33114
> Project: Flink
>  Issue Type: Bug
>  Components: API / Core
>Affects Versions: 1.19.0
>Reporter: Sergey Nuyanzin
>Priority: Critical
>  Labels: stale-critical, test-stability
>
> This build 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=53250=logs=0da23115-68bb-5dcd-192c-bd4c8adebde1=24c3384f-1bcb-57b3-224f-51bf973bbee8
> failed as 
> {noformat}
> Sep 14 22:34:21   at 
> org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37)
> Sep 14 22:34:21   at 
> org.junit.jupiter.engine.execution.InterceptingExecutableInvoker.invoke(InterceptingExecutableInvoker.java:92)
> Sep 14 22:34:21   at 
> org.junit.jupiter.engine.execution.InterceptingExecutableInvoker.invoke(InterceptingExecutableInvoker.java:86)
> Sep 14 22:34:21   at 
> org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.invokeMethodInExtensionContext(ClassBasedTestDescriptor.java:520)
> Sep 14 22:34:21   at 
> org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$synthesizeAfterEachMethodAdapter$24(ClassBasedTestDescriptor.java:510)
> Sep 14 22:34:21   at 
> org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeAfterEachMethods$10(TestMethodTestDescriptor.java:243)
> Sep 14 22:34:21   at 
> org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeAllAfterMethodsOrCallbacks$13(TestMethodTestDescriptor.java:276)
> Sep 14 22:34:21   at 
> org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
> Sep 14 22:34:21   at 
> org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeAllAfterMethodsOrCallbacks$14(TestMethodTestDescriptor.java:276)
> Sep 14 22:34:21   at 
> java.util.ArrayList.forEach(ArrayList.java:1259)
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32743) Flink kafka connector source can directly parse data collected from kafka-connect

2023-10-03 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32743?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32743:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Flink kafka connector source can directly parse data collected from 
> kafka-connect
> -
>
> Key: FLINK-32743
> URL: https://issues.apache.org/jira/browse/FLINK-32743
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Reporter: Xiaojian Sun
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> If the data in Kafka is collected through the Kafka connect system, it is 
> hoped that the data collected through Flink has already excluded the schema



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-20886) Add the option to get a threaddump on checkpoint timeouts

2023-10-03 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20886?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-20886:
---
Labels: auto-deprioritized-major auto-deprioritized-minor stale-assigned 
usability  (was: auto-deprioritized-major auto-deprioritized-minor usability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Add the option to get a threaddump on checkpoint timeouts
> -
>
> Key: FLINK-20886
> URL: https://issues.apache.org/jira/browse/FLINK-20886
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Checkpointing
>Reporter: Nico Kruber
>Assignee: Zakelly Lan
>Priority: Minor
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> stale-assigned, usability
>
> For debugging checkpoint timeouts, I was thinking about the following 
> addition to Flink:
> When a checkpoint times out and the async thread is still running, create a 
> thread dump [1] and either add this to the checkpoint stats, log it, or write 
> it out.
> This may help identifying where the checkpoint is stuck (maybe a lock, could 
> also be in a third party lib like the FS connectors,...). It would give us 
> some insights into what the thread is currently doing.
> Limiting the scope of the threads would be nice but may not be possible in 
> the general case since additional threads (spawned by the FS connector lib, 
> or otherwise connected) may interact with the async thread(s) by e.g. going 
> through the same locks. Maybe we can reduce the thread dumps to all async 
> threads of the failed checkpoint + all thready that interact with it, e.g. 
> via locks?
> I'm also not sure whether the ability to have thread dumps or not should be 
> user-configurable (Could it contain sensitive information from other jobs if 
> you run a session cluster? Is that even relevant since we don't give 
> isolation guarantees anyway?). If it is configurable, it should be on by 
> default.
> [1] https://crunchify.com/how-to-generate-java-thread-dump-programmatically/



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32733) Add AlgoOperator for FpGrowth

2023-10-02 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32733?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32733:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Add AlgoOperator for FpGrowth
> -
>
> Key: FLINK-32733
> URL: https://issues.apache.org/jira/browse/FLINK-32733
> Project: Flink
>  Issue Type: New Feature
>  Components: Library / Machine Learning
>Affects Versions: ml-2.4.0
>Reporter: Yindi Wang
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> _Frequent Pattern Growth_ Algorithm is the method of finding frequent 
> patterns. 
> The distributed version of FpGrowth algorithm is described in Li et al., PFP: 
> Parallel FP-Growth for Query Recommendation. This algorithm will be an 
> addition to the 
> _org.apache.flink.ml.recommendation_ package.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32523) NotifyCheckpointAbortedITCase.testNotifyCheckpointAborted fails with timeout on AZP

2023-10-01 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32523?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32523:
---
Labels: pull-request-available stale-assigned test-stability  (was: 
pull-request-available test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> NotifyCheckpointAbortedITCase.testNotifyCheckpointAborted fails with timeout 
> on AZP
> ---
>
> Key: FLINK-32523
> URL: https://issues.apache.org/jira/browse/FLINK-32523
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.16.2, 1.18.0, 1.17.1
>Reporter: Sergey Nuyanzin
>Assignee: Hangxiang Yu
>Priority: Critical
>  Labels: pull-request-available, stale-assigned, test-stability
> Fix For: 1.18.0, 1.17.2
>
> Attachments: failure.log
>
>
> This build
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=50795=logs=39d5b1d5-3b41-54dc-6458-1e2ddd1cdcf3=0c010d0c-3dec-5bf1-d408-7b18988b1b2b=8638
>  fails with timeout
> {noformat}
> Jul 03 01:26:35 org.junit.runners.model.TestTimedOutException: test timed out 
> after 10 milliseconds
> Jul 03 01:26:35   at java.lang.Object.wait(Native Method)
> Jul 03 01:26:35   at java.lang.Object.wait(Object.java:502)
> Jul 03 01:26:35   at 
> org.apache.flink.core.testutils.OneShotLatch.await(OneShotLatch.java:61)
> Jul 03 01:26:35   at 
> org.apache.flink.test.checkpointing.NotifyCheckpointAbortedITCase.verifyAllOperatorsNotifyAborted(NotifyCheckpointAbortedITCase.java:198)
> Jul 03 01:26:35   at 
> org.apache.flink.test.checkpointing.NotifyCheckpointAbortedITCase.testNotifyCheckpointAborted(NotifyCheckpointAbortedITCase.java:189)
> Jul 03 01:26:35   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> Jul 03 01:26:35   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> Jul 03 01:26:35   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> Jul 03 01:26:35   at java.lang.reflect.Method.invoke(Method.java:498)
> Jul 03 01:26:35   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> Jul 03 01:26:35   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> Jul 03 01:26:35   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> Jul 03 01:26:35   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> Jul 03 01:26:35   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
> Jul 03 01:26:35   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
> Jul 03 01:26:35   at 
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> Jul 03 01:26:35   at java.lang.Thread.run(Thread.java:748)
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32714) JDBC: Add dialect for OceanBase database

2023-10-01 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32714?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32714:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> JDBC: Add dialect for OceanBase database
> 
>
> Key: FLINK-32714
> URL: https://issues.apache.org/jira/browse/FLINK-32714
> Project: Flink
>  Issue Type: New Feature
>Reporter: He Wang
>Priority: Major
>  Labels: pull-request-available, stale-major
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> OceanBase is a distributed relational database, the community edition of 
> OceanBase is open sourced at [https://github.com/oceanbase/oceanbase.]
> The enterprise edition of OceanBase is compatible with MySql and Oracle, 
> which means we can reuse almost all the dialect rules. 
> The difference from other databases is that we must provide the compatibility 
> mode firstly, then the connector can determine which dialect to use, so a 
> startup option like 'compatible-mode'  is needed.
> A dialect implementation for OceanBase is like below: 
> {code:java}
> package org.apache.flink.connector.jdbc.databases.oceanbase;
> import org.apache.flink.connector.jdbc.converter.JdbcRowConverter;
> import org.apache.flink.connector.jdbc.databases.mysql.dialect.MySqlDialect;
> import org.apache.flink.connector.jdbc.databases.oracle.dialect.OracleDialect;
> import org.apache.flink.connector.jdbc.dialect.AbstractDialect;
> import org.apache.flink.table.types.logical.LogicalTypeRoot;
> import org.apache.flink.table.types.logical.RowType;
> import javax.annotation.Nonnull;
> import java.util.Optional;
> import java.util.Set;
> /** JDBC dialect for OceanBase. */
> public class OceanBaseDialect extends AbstractDialect {
> private static final long serialVersionUID = 1L;
> private final AbstractDialect dialect;
> public OceanBaseDialect(@Nonnull String compatibleMode) {
> switch (compatibleMode.toLowerCase()) {
> case "mysql":
> this.dialect = new MySqlDialect();
> break;
> case "oracle":
> this.dialect = new OracleDialect();
> break;
> default:
> throw new IllegalArgumentException(
> "Unsupported compatible mode: " + compatibleMode);
> }
> }
> @Override
> public String dialectName() {
> return "OceanBase";
> }
> @Override
> public Optional defaultDriverName() {
> return Optional.of("com.oceanbase.jdbc.Driver");
> }
> @Override
> public Set supportedTypes() {
> return dialect.supportedTypes();
> }
> @Override
> public JdbcRowConverter getRowConverter(RowType rowType) {
> return dialect.getRowConverter(rowType);
> }
> @Override
> public String getLimitClause(long limit) {
> return dialect.getLimitClause(limit);
> }
> @Override
> public String quoteIdentifier(String identifier) {
> return dialect.quoteIdentifier(identifier);
> }
> @Override
> public Optional getUpsertStatement(
> String tableName, String[] fieldNames, String[] conditionFields) {
> return dialect.getUpsertStatement(tableName, fieldNames, 
> conditionFields);
> }
> @Override
> public Optional timestampPrecisionRange() {
> return dialect.timestampPrecisionRange();
> }
> @Override
> public Optional decimalPrecisionRange() {
> return dialect.decimalPrecisionRange();
> }
> @Override
> public String appendDefaultUrlProperties(String url) {
> return dialect.appendDefaultUrlProperties(url);
> }
> }
>  {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-28879) New File Sink s3 end-to-end test failed with Output hash mismatch

2023-10-01 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28879?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-28879:
---
  Labels: auto-deprioritized-major test-stability  (was: stale-major 
test-stability)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> New File Sink s3 end-to-end test failed with Output hash mismatch
> -
>
> Key: FLINK-28879
> URL: https://issues.apache.org/jira/browse/FLINK-28879
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream, Connectors / FileSystem, Tests
>Affects Versions: 1.16.0, 1.17.2
>Reporter: Huang Xingbo
>Priority: Minor
>  Labels: auto-deprioritized-major, test-stability
>
> {code:java}
> 2022-08-09T00:50:02.8229585Z Aug 09 00:50:02 FAIL File Streaming Sink: Output 
> hash mismatch.  Got 6037b01ca0ffc61a95c12cb475c661a8, expected 
> 6727342fdd3aae2129e61fc8f433fb6f.
> 2022-08-09T00:50:02.8230700Z Aug 09 00:50:02 head hexdump of actual:
> 2022-08-09T00:50:02.8477319Z Aug 09 00:50:02 000   E   r   r   o   r  
>  e   x   e   c   u   t   i   n   g
> 2022-08-09T00:50:02.8478206Z Aug 09 00:50:02 010   a   w   s   c   o  
>  m   m   a   n   d   :   s   3
> 2022-08-09T00:50:02.8479475Z Aug 09 00:50:02 020   c   p   -   -   q  
>  u   i   e   t   s   3   :   /   /
> 2022-08-09T00:50:02.8480205Z Aug 09 00:50:02 030   f   l   i   n   k   -  
>  i   n   t   e   g   r   a   t   i   o
> 2022-08-09T00:50:02.8480924Z Aug 09 00:50:02 040   n   -   t   e   s   t  
>  s   /   t   e   m   p   /   t   e   s
> 2022-08-09T00:50:02.8481612Z Aug 09 00:50:02 050   t   _   f   i   l   e  
>  _   s   i   n   k   -   1   d   3   d
> 2022-08-09T00:50:02.8483048Z Aug 09 00:50:02 060   4   0   0   8   -   b  
>  0   b   f   -   4   2   6   5   -   b
> 2022-08-09T00:50:02.8483618Z Aug 09 00:50:02 070   e   0   e   -   3   b  
>  9   f   7   8   2   c   5   5   2   d
> 2022-08-09T00:50:02.8484222Z Aug 09 00:50:02 080   /   h   o   s   t  
>  d   i   r   /   /   t   e   m   p   -
> 2022-08-09T00:50:02.8484831Z Aug 09 00:50:02 090   t   e   s   t   -   d  
>  i   r   e   c   t   o   r   y   -   2
> 2022-08-09T00:50:02.8485719Z Aug 09 00:50:02 0a0   3   9   3   7   7   8  
>  2   6   8   0   /   t   e   m   p   /
> 2022-08-09T00:50:02.8486427Z Aug 09 00:50:02 0b0   t   e   s   t   _   f  
>  i   l   e   _   s   i   n   k   -   1
> 2022-08-09T00:50:02.8487134Z Aug 09 00:50:02 0c0   d   3   d   4   0   0  
>  8   -   b   0   b   f   -   4   2   6
> 2022-08-09T00:50:02.8487826Z Aug 09 00:50:02 0d0   5   -   b   e   0   e  
>  -   3   b   9   f   7   8   2   c   5
> 2022-08-09T00:50:02.8488511Z Aug 09 00:50:02 0e0   5   2   d   -   -  
>  e   x   c   l   u   d   e   '   *
> 2022-08-09T00:50:02.8489202Z Aug 09 00:50:02 0f0   '   -   -   i   n  
>  c   l   u   d   e   '   *   /   p
> 2022-08-09T00:50:02.8489891Z Aug 09 00:50:02 100   a   r   t   -   [   !  
>  /   ]   *   '   -   -   r   e   c
> 2022-08-09T00:50:02.8490385Z Aug 09 00:50:02 110   u   r   s   i   v   e  
> \n
> 2022-08-09T00:50:02.8490822Z Aug 09 00:50:02 117
> 2022-08-09T00:50:02.8502212Z Aug 09 00:50:02 Stopping job timeout watchdog 
> (with pid=141134)
> 2022-08-09T00:50:06.8430959Z rm: cannot remove 
> '/home/vsts/work/1/s/flink-dist/target/flink-1.16-SNAPSHOT-bin/flink-1.16-SNAPSHOT/lib/flink-shaded-netty-tcnative-static-*.jar':
>  No such file or directory
> 2022-08-09T00:50:06.9278248Z Aug 09 00:50:06 
> 5ccfeb22307c2a88625a38b9537acc79001d1b29094ef40fd70692ce11407502
> 2022-08-09T00:50:06.9618147Z Aug 09 00:50:06 
> 5ccfeb22307c2a88625a38b9537acc79001d1b29094ef40fd70692ce11407502
> 2022-08-09T00:50:06.9645077Z Aug 09 00:50:06 [FAIL] Test script contains 
> errors.
> 2022-08-09T00:50:06.9666227Z Aug 09 00:50:06 Checking of logs skipped.
> 2022-08-09T00:50:06.9671891Z Aug 09 00:50:06 
> 2022-08-09T00:50:06.9673050Z Aug 09 00:50:06 [FAIL] 'New File Sink s3 
> end-to-end test' failed after 3 minutes and 42 seconds! Test exited with exit 
> code 1
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39667=logs=af184cdd-c6d8-5084-0b69-7e9c67b35f7a=160c9ae5-96fd-516e-1c91-deb81f59292a=4136



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32570) Deprecate API that uses Flink's Time implementation (related to FLINK-14638)

2023-10-01 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32570:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Deprecate API that uses Flink's Time implementation (related to FLINK-14638)
> 
>
> Key: FLINK-32570
> URL: https://issues.apache.org/jira/browse/FLINK-32570
> Project: Flink
>  Issue Type: Technical Debt
>Affects Versions: 1.19.0
>Reporter: Matthias Pohl
>Assignee: Matthias Pohl
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> The plan is to resolve FLINK-14038 with Flink 2.0. As a preparation, we have 
> to deprecate related @Public API .



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31825) Stopping minikube fails with timeout

2023-09-30 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31825?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31825:
---
  Labels: auto-deprioritized-major test-stability  (was: stale-major 
test-stability)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Stopping minikube fails with timeout
> 
>
> Key: FLINK-31825
> URL: https://issues.apache.org/jira/browse/FLINK-31825
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hadoop Compatibility, Tests
>Affects Versions: 1.18.0
>Reporter: Sergey Nuyanzin
>Priority: Minor
>  Labels: auto-deprioritized-major, test-stability
>
> Currently there is not so much information in logs...
> {noformat}
> Apr 13 01:27:02 * Stopping node "minikube"  ...
> ==
> === WARNING: This task took already 95% of the available time budget of 286 
> minutes ===
> ==
> ==
> The following Java processes are running (JPS)
> ==
> 243413 Jps
> ==
> Printing stack trace of Java process 243413
> ==
> 243413: No such process
> ==
> The following Java processes are running (JPS)
> ==
> 243516 Jps
> ==
> Printing stack trace of Java process 243516
> ==
> 243516: No such process
> =
> === WARNING: Killing task ===
> =
> Terminated
> Apr 13 05:48:53 [FAIL] Test script contains errors.
> {noformat}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=48096=logs=bbb1e2a2-a43c-55c8-fb48-5cfe7a8a0ca6=ba24ad14-6ea3-5ee3-c4ec-9e7cd2c9e754=5290



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32722) 239 exit code in flink-runtime

2023-09-30 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32722?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32722:
---
Labels: stale-major test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> 239 exit code in flink-runtime
> --
>
> Key: FLINK-32722
> URL: https://issues.apache.org/jira/browse/FLINK-32722
> Project: Flink
>  Issue Type: Bug
>  Components: Test Infrastructure
>Affects Versions: 1.16.2
>Reporter: Matthias Pohl
>Priority: Major
>  Labels: stale-major, test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=51852=logs=d89de3df-4600-5585-dadc-9bbc9a5e661c=be5a4b15-4b23-56b1-7582-795f58a645a2=8418
> {code:java}
> Aug 01 01:03:49 [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-surefire-plugin:3.0.0-M5:test (default-test) 
> on project flink-runtime: There are test failures.
> Aug 01 01:03:49 [ERROR] 
> Aug 01 01:03:49 [ERROR] Please refer to 
> /__w/2/s/flink-runtime/target/surefire-reports for the individual test 
> results.
> Aug 01 01:03:49 [ERROR] Please refer to dump files (if any exist) 
> [date].dump, [date]-jvmRun[N].dump and [date].dumpstream.
> Aug 01 01:03:49 [ERROR] ExecutionException The forked VM terminated without 
> properly saying goodbye. VM crash or System.exit called?
> Aug 01 01:03:49 [ERROR] Command was /bin/sh -c cd /__w/2/s/flink-runtime && 
> /usr/lib/jvm/java-8-openjdk-amd64/jre/bin/java -XX:+UseG1GC -Xms256m -Xmx768m 
> -jar 
> /__w/2/s/flink-runtime/target/surefire/surefirebooter1803120121827605294.jar 
> /__w/2/s/flink-runtime/target/surefire 2023-08-01T00-58-17_520-jvmRun1 
> surefire9107652818401825168tmp surefire_261701267003130520249tmp
> Aug 01 01:03:49 [ERROR] Error occurred in starting fork, check output in log
> Aug 01 01:03:49 [ERROR] Process Exit Code: 239
> Aug 01 01:03:49 [ERROR] 
> org.apache.maven.surefire.booter.SurefireBooterForkException: 
> ExecutionException The forked VM terminated without properly saying goodbye. 
> VM crash or System.exit called?
> Aug 01 01:03:49 [ERROR] Command was /bin/sh -c cd /__w/2/s/flink-runtime && 
> /usr/lib/jvm/java-8-openjdk-amd64/jre/bin/java -XX:+UseG1GC -Xms256m -Xmx768m 
> -jar 
> /__w/2/s/flink-runtime/target/surefire/surefirebooter1803120121827605294.jar 
> /__w/2/s/flink-runtime/target/surefire 2023-08-01T00-58-17_520-jvmRun1 
> surefire9107652818401825168tmp surefire_261701267003130520249tmp
> Aug 01 01:03:49 [ERROR] Error occurred in starting fork, check output in log
> Aug 01 01:03:49 [ERROR] Process Exit Code: 239
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.booterclient.ForkStarter.awaitResultsDone(ForkStarter.java:532)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.booterclient.ForkStarter.runSuitesForkOnceMultiple(ForkStarter.java:405)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:321)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:266)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1314)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1159)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:932)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:132)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:208)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:153)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:145)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:116)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:80)
> Aug 01 01:03:49 [ERROR] at 
> org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build(SingleThreadedBuilder.java:51)
> Aug 01 01:03:49 

[jira] [Updated] (FLINK-29390) Pulsar SQL Connector: SQLClient E2E testing

2023-09-30 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-29390:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Pulsar SQL Connector: SQLClient E2E testing
> ---
>
> Key: FLINK-29390
> URL: https://issues.apache.org/jira/browse/FLINK-29390
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Pulsar
>Affects Versions: 1.17.0
>Reporter: Yufei Zhang
>Assignee: Zili Chen
>Priority: Minor
>  Labels: pull-request-available, stale-assigned
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32665) Support read null value for csv format

2023-09-30 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32665?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32665:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Support read null value for csv format
> --
>
> Key: FLINK-32665
> URL: https://issues.apache.org/jira/browse/FLINK-32665
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.18.0
>Reporter: Fang Yong
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> when there is null column in a file with csv format, it will throw exception 
> when flink job try to parse these data



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-28440) EventTimeWindowCheckpointingITCase failed with restore

2023-09-30 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28440?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-28440:
---
Labels: auto-deprioritized-critical pull-request-available stale-assigned 
test-stability  (was: auto-deprioritized-critical pull-request-available 
test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> EventTimeWindowCheckpointingITCase failed with restore
> --
>
> Key: FLINK-28440
> URL: https://issues.apache.org/jira/browse/FLINK-28440
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing, Runtime / State Backends
>Affects Versions: 1.16.0, 1.17.0, 1.18.0
>Reporter: Huang Xingbo
>Assignee: Yanfei Lei
>Priority: Critical
>  Labels: auto-deprioritized-critical, pull-request-available, 
> stale-assigned, test-stability
> Fix For: 1.18.0
>
> Attachments: image-2023-02-01-00-51-54-506.png, 
> image-2023-02-01-01-10-01-521.png, image-2023-02-01-01-19-12-182.png, 
> image-2023-02-01-16-47-23-756.png, image-2023-02-01-16-57-43-889.png, 
> image-2023-02-02-10-52-56-599.png, image-2023-02-03-10-09-07-586.png, 
> image-2023-02-03-12-03-16-155.png, image-2023-02-03-12-03-56-614.png
>
>
> {code:java}
> Caused by: java.lang.Exception: Exception while creating 
> StreamOperatorStateContext.
>   at 
> org.apache.flink.streaming.api.operators.StreamTaskStateInitializerImpl.streamOperatorStateContext(StreamTaskStateInitializerImpl.java:256)
>   at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:268)
>   at 
> org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.initializeStateAndOpenOperators(RegularOperatorChain.java:106)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.restoreGates(StreamTask.java:722)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$1.call(StreamTaskActionExecutor.java:55)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.restoreInternal(StreamTask.java:698)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:665)
>   at 
> org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:935)
>   at 
> org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:904)
>   at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:728)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:550)
>   at java.lang.Thread.run(Thread.java:748)
> Caused by: org.apache.flink.util.FlinkException: Could not restore keyed 
> state backend for WindowOperator_0a448493b4782967b150582570326227_(2/4) from 
> any of the 1 provided restore options.
>   at 
> org.apache.flink.streaming.api.operators.BackendRestorerProcedure.createAndRestore(BackendRestorerProcedure.java:160)
>   at 
> org.apache.flink.streaming.api.operators.StreamTaskStateInitializerImpl.keyedStatedBackend(StreamTaskStateInitializerImpl.java:353)
>   at 
> org.apache.flink.streaming.api.operators.StreamTaskStateInitializerImpl.streamOperatorStateContext(StreamTaskStateInitializerImpl.java:165)
>   ... 11 more
> Caused by: java.lang.RuntimeException: java.io.FileNotFoundException: 
> /tmp/junit1835099326935900400/junit1113650082510421526/52ee65b7-033f-4429-8ddd-adbe85e27ced
>  (No such file or directory)
>   at org.apache.flink.util.ExceptionUtils.rethrow(ExceptionUtils.java:321)
>   at 
> org.apache.flink.runtime.state.changelog.StateChangelogHandleStreamHandleReader$1.advance(StateChangelogHandleStreamHandleReader.java:87)
>   at 
> org.apache.flink.runtime.state.changelog.StateChangelogHandleStreamHandleReader$1.hasNext(StateChangelogHandleStreamHandleReader.java:69)
>   at 
> org.apache.flink.state.changelog.restore.ChangelogBackendRestoreOperation.readBackendHandle(ChangelogBackendRestoreOperation.java:96)
>   at 
> org.apache.flink.state.changelog.restore.ChangelogBackendRestoreOperation.restore(ChangelogBackendRestoreOperation.java:75)
>   at 
> org.apache.flink.state.changelog.ChangelogStateBackend.restore(ChangelogStateBackend.java:92)
>   at 
> 

[jira] [Updated] (FLINK-32910) Silence curls in test code

2023-09-30 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32910?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32910:
---
Labels: pull-request-available stale-assigned starter  (was: 
pull-request-available starter)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Silence curls in test code
> --
>
> Key: FLINK-32910
> URL: https://issues.apache.org/jira/browse/FLINK-32910
> Project: Flink
>  Issue Type: Improvement
>  Components: Test Infrastructure
>Affects Versions: 1.16.2, 1.18.0, 1.17.1
>Reporter: Matthias Pohl
>Assignee: Matthias Pohl
>Priority: Major
>  Labels: pull-request-available, stale-assigned, starter
>
> We use {{curl}} in several locations to download artifacts. Usually, the a 
> progress bar is printed which spams the console output of the test execution. 
> This issue is about cleaning this up.
> Parameters to consider (depending on the usecase):
>  * {{\-L}}/{{\-\-location}} redirects the curl command and retries if the 
> server reported that the artifact was moved
>  * {{\-O}}/{{\-\-remote-name}} writes output to file matching the remote name 
> (which was extracted from the URL) instead of stdout; alternative: 
> {{\-o}}/{{\-\-output}} writes output to a file with the given name instead of 
> stdout
> * {{\-f}}/{{\-\-fail}} makes curl command fail with non-0 exit code for HTTP 
> error codes
> * {{\-s \-S}}/{{\-\-silent \-\-show-error}} doesn't print progress bar but 
> shows error
> * {{\-r}}/{{\-\-retry}} Retries certain errors
> {{curl}} uses a default config file {{${user.home}/.curlrc}}. But one could 
> make it more explicit using {{\-K}}/{{\-\-config}}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33056) NettyClientServerSslTest#testValidSslConnection fails on AZP

2023-09-29 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33056?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-33056:
---
  Labels: auto-deprioritized-critical test-stability  (was: stale-critical 
test-stability)
Priority: Major  (was: Critical)

This issue was labeled "stale-critical" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Critical, 
please raise the priority and ask a committer to assign you the issue or revive 
the public discussion.


> NettyClientServerSslTest#testValidSslConnection fails on AZP
> 
>
> Key: FLINK-33056
> URL: https://issues.apache.org/jira/browse/FLINK-33056
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Configuration, Runtime / Coordination
>Affects Versions: 1.19.0
>Reporter: Sergey Nuyanzin
>Priority: Major
>  Labels: auto-deprioritized-critical, test-stability
> Attachments: logs-cron_azure-test_cron_azure_core-1694048924.zip
>
>
> This build 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=53020=logs=77a9d8e1-d610-59b3-fc2a-4766541e0e33=125e07e7-8de0-5c6c-a541-a567415af3ef=8592
> fails with 
> {noformat}
> Test testValidSslConnection[SSL provider = 
> JDK](org.apache.flink.runtime.io.network.netty.NettyClientServerSslTest) is 
> running.
> 
> 01:20:31,479 [main] INFO  
> org.apache.flink.runtime.io.network.netty.NettyConfig[] - NettyConfig 
> [server address: localhost/127.0.0.1, server port range: 36717, ssl enabled: 
> true, memory segment size (bytes): 1024, transport type: AUTO, number of 
> server threads: 1 (manual), number of client thr
> eads: 1 (manual), server connect backlog: 0 (use Netty's default), client 
> connect timeout (sec): 120, send/receive buffer size (bytes): 0 (use Netty's 
> default)]
> 01:20:31,479 [main] INFO  
> org.apache.flink.runtime.io.network.netty.NettyServer[] - Transport 
> type 'auto': using EPOLL.
> 01:20:31,475 [Flink Netty Client (42359) Thread 0] WARN  
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline [] - 
> An exceptionCaught() event was fired, and it reached at the tail of the 
> pipeline. It usually means the last handler in the pipeline did not handle 
> the exception.
> org.apache.flink.shaded.netty4.io.netty.handler.codec.DecoderException: 
> javax.net.ssl.SSLHandshakeException: server certificate with unknown 
> fingerprint: CN=Unknown, OU=Unknown, O=Unknown, L=Unknown, ST=Unknown, 
> C=Unknown
> at 
> org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:499)
>  ~[flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:290)
>  ~[flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.epoll.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:800)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:499)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> 

[jira] [Updated] (FLINK-31406) Do not delete jobgraph on scale only last-state upgrades

2023-09-29 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31406?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31406:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Do not delete jobgraph on scale only last-state upgrades
> 
>
> Key: FLINK-31406
> URL: https://issues.apache.org/jira/browse/FLINK-31406
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Reporter: Gyula Fora
>Assignee: Nicholas Jiang
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> Currently the operator always deletes the jobgraph from HA metadata so that 
> it's regenerated for last-state upgrades. 
> This is unnecessary for scale only operations. Keeping the jobgraph can 
> greately speed up startup time for some jobs.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-16987) FLIP-95: Add new table source and sink interfaces

2023-09-29 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16987?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-16987:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> FLIP-95: Add new table source and sink interfaces
> -
>
> Key: FLINK-16987
> URL: https://issues.apache.org/jira/browse/FLINK-16987
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API, Table SQL / Planner
>Reporter: Timo Walther
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> Proper support for handling changelogs, more efficient processing of data 
> through the new Blink planner, and unified interfaces that are DataStream API 
> agnostic make it necessary to rework the table source and sink interfaces.
> The goals of this FLIP are:
>  * *Simplify the current interface architecture*:
>  ** Merge upsert, retract, and append sinks.
>  ** Unify batch and streaming sources.
>  ** Unify batch and streaming sinks.
>  * *Allow sources to produce a changelog*:
>  ** UpsertTableSources have been requested a lot by users. Now is the time to 
> open the internal planner capabilities via the new interfaces.
>  ** According to FLIP-105, we would like to support changelogs for processing 
> formats such as [Debezium|https://debezium.io/].
>  * *Don't rely on DataStream API for source and sinks*:
>  ** According to FLIP-32, the Table API and SQL should be independent of the 
> DataStream API which is why the `table-common` module has no dependencies on 
> `flink-streaming-java`.
>  ** Source and sink implementations should only depend on the `table-common` 
> module after FLIP-27.
>  ** Until FLIP-27 is ready, we still put most of the interfaces in 
> `table-common` and strictly separate interfaces that communicate with a 
> planner and actual runtime reader/writers.
>  * *Implement efficient sources and sinks without planner dependencies*:
>  ** Make Blink's internal data structures available to connectors.
>  ** Introduce stable interfaces for data structures that can be marked as 
> `@PublicEvolving`.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31615) Fix some parts forgot to translate in "Table API" page of "Table API & SQL"

2023-09-28 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31615?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31615:
---
Labels: chinese-translation stale-minor  (was: chinese-translation)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Minor but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 180 days. I have gone ahead and marked it "stale-minor". If this ticket is 
still Minor, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Fix some parts forgot to translate in "Table API" page of "Table API & SQL" 
> 
>
> Key: FLINK-31615
> URL: https://issues.apache.org/jira/browse/FLINK-31615
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Hang Ruan
>Priority: Minor
>  Labels: chinese-translation, stale-minor
>
> The query_state_warning in "Table API" page of "Table API & SQL"  is still in 
> English. And some comments in codes are in English.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-13876) Remove ExecutionConfig field from PojoSerializer

2023-09-27 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-13876:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor  (was: 
auto-deprioritized-major stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Remove ExecutionConfig field from PojoSerializer
> 
>
> Key: FLINK-13876
> URL: https://issues.apache.org/jira/browse/FLINK-13876
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Type Serialization System
>Affects Versions: 1.7.2, 1.8.1, 1.9.0, 1.10.0
>Reporter: Dawid Wysakowicz
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor
>
> The PojoSerializers stores an instance of ExecutionConfig as internal field, 
> even though that the only information it may ever need are the registered 
> kryo serializers.
> This has a few drawbacks:
> * It blocks the evolution of {{ExecutionConfig}} as serializers where stored 
> in a state. Therefore any change to ExecutionConfig must be backwards 
> compatible in respect to java serialization
> * It probably already introduced a bug, as upon restore the Snapshot actually 
> recreates the serializer with an empty ExecutionConfig (see 
> org.apache.flink.api.java.typeutils.runtime.PojoSerializerSnapshot#restoreSerializer)
> I suggest to remove the field completely and adjust corresponding usages.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32866) Unify the TestLoggerExtension config of junit5

2023-09-27 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32866?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32866:
---
Labels: pull-request-available stale-assigned starter  (was: 
pull-request-available starter)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Unify the TestLoggerExtension config of junit5
> --
>
> Key: FLINK-32866
> URL: https://issues.apache.org/jira/browse/FLINK-32866
> Project: Flink
>  Issue Type: Sub-task
>  Components: Tests
>Reporter: Rui Fan
>Assignee: Rui Fan
>Priority: Minor
>  Labels: pull-request-available, stale-assigned, starter
>
> Some modules added the {{TestLoggerExtension}} to the 
> {{org.junit.jupiter.api.extension.Extension}} resource file. All test classes 
> of these modules don't need add the 
> {{@ExtendWith(TestLoggerExtension.class)}}  at class level.
> This JIRA propose clean up the {{@ExtendWith(TestLoggerExtension.class)}}  
> for modules that added the {{TestLoggerExtension}} to the 
> {{org.junit.jupiter.api.extension.Extension}} resource file.
> Update: We could also investigate in what extend we could remove the 
> {{org.junit.jupiter.api.extension.Extension}} resource files from certain 
> modules. It only has to be present once on the classpath. A single location 
> for this configuration would be nice (e.g. putting it in 
> {{flink-test-utils-parent/flink-test-utils-junit}}). But I think not all 
> modules have this one as there dependency. So, having a single resource file 
> for configuring this might be not possible. But at least, we can reduce the 
> number of files to a minimum.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32582) Move TypeSerializerUpgradeTestBase from Kafka connector into flink-connector-common

2023-09-26 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32582?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32582:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Move TypeSerializerUpgradeTestBase from Kafka connector into 
> flink-connector-common
> ---
>
> Key: FLINK-32582
> URL: https://issues.apache.org/jira/browse/FLINK-32582
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Common
>Reporter: Matthias Pohl
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> The externalization of connectors made caused problems with the Flink's test 
> data generation. The Kafka connector relied on TypeSerializerUpgradeTestBase 
> for some test cases which was fine prior to FLINK-27518 where the test data 
> generation was handled individually.
> With FLINK-27518 the process was automated in Flink 1.18. For now, the 
> TypeSerializerUpgradeTestBase class was just copied over into the Kafka 
> connector since it was the only connector that would utilize this test base.
> But we might want to provide a more generalized solution where the test base 
> is provided by {{flink-connector-common}} to offer a generalized approach for 
> any connector.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33025) BatchArrowPythonOverWindowAggregateFunctionOperatorTest.testFinishBundleTriggeredByCount fails on AZP

2023-09-26 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33025?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-33025:
---
  Labels: auto-deprioritized-critical test-stability  (was: stale-critical 
test-stability)
Priority: Major  (was: Critical)

This issue was labeled "stale-critical" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Critical, 
please raise the priority and ask a committer to assign you the issue or revive 
the public discussion.


> BatchArrowPythonOverWindowAggregateFunctionOperatorTest.testFinishBundleTriggeredByCount
>  fails on AZP
> -
>
> Key: FLINK-33025
> URL: https://issues.apache.org/jira/browse/FLINK-33025
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.3, 1.19.0
>Reporter: Sergey Nuyanzin
>Priority: Major
>  Labels: auto-deprioritized-critical, test-stability
>
> This build 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=52958=logs=5cae8624-c7eb-5c51-92d3-4d2dacedd221=5acec1b4-945b-59ca-34f8-168928ce5199=22618
>  fails on AZP as
> {noformat}
> Sep 03 05:05:38 05:05:38.220 [ERROR] Failures: 
> Sep 03 05:05:38 05:05:38.220 [ERROR]   
> BatchArrowPythonOverWindowAggregateFunctionOperatorTest.testFinishBundleTriggeredByCount:122->ArrowPythonAggregateFunctionOperatorTestBase.assertOutputEquals:62
>  
> Sep 03 05:05:38 Expected size: 4 but was: 3 in:
> Sep 03 05:05:38 [Record @ (undef) : +I(c1,c2,0,0,0),
> Sep 03 05:05:38 Record @ (undef) : +I(c1,c4,1,0,0),
> Sep 03 05:05:38 Record @ (undef) : +I(c1,c6,2,10,2)]
> {noformat}
> probably related to FLINK-26990



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32471) IS_NOT_NULL can add to SUITABLE_FILTER_TO_PUSH

2023-09-25 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32471?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32471:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> IS_NOT_NULL can add to SUITABLE_FILTER_TO_PUSH
> --
>
> Key: FLINK-32471
> URL: https://issues.apache.org/jira/browse/FLINK-32471
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner
>Reporter: grandfisher
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> According to FLINK-31273:
> The reason for the error is that other filters conflict with IS_NULL, but in 
> fact it won't conflict with IS_NOT_NULL, because operators in 
> SUITABLE_FILTER_TO_PUSH  such as 'SqlKind.GREATER_THAN'  has an implicit 
> filter 'IS_NOT_NULL' according to SQL Semantics.
>  
> So we think it is feasible to add  IS_NOT_NULL to the SUITABLE_FILTER_TO_PUSH 
> list.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-30519) Add e2e tests for operator dynamic config

2023-09-24 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30519?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-30519:
---
  Labels: auto-deprioritized-critical stale-assigned starter  (was: 
stale-assigned stale-critical starter)
Priority: Major  (was: Critical)

This issue was labeled "stale-critical" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Critical, 
please raise the priority and ask a committer to assign you the issue or revive 
the public discussion.


> Add e2e tests for operator dynamic config
> -
>
> Key: FLINK-30519
> URL: https://issues.apache.org/jira/browse/FLINK-30519
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Kubernetes Operator
>Reporter: Gyula Fora
>Priority: Major
>  Labels: auto-deprioritized-critical, stale-assigned, starter
>
> The dynamic config feature is currently not covered by e2e tests and is 
> subject to accidental regressions, as shown in:
> https://issues.apache.org/jira/browse/FLINK-30329
> We should add an e2e test that covers this



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-26974) Python EmbeddedThreadDependencyTests.test_add_python_file failed on azure

2023-09-23 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-26974?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-26974:
---
Labels: auto-deprioritized-major stale-assigned test-stability  (was: 
auto-deprioritized-major test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Python EmbeddedThreadDependencyTests.test_add_python_file failed on azure
> -
>
> Key: FLINK-26974
> URL: https://issues.apache.org/jira/browse/FLINK-26974
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.15.0, 1.16.0, 1.17.0
>Reporter: Yun Gao
>Assignee: Huang Xingbo
>Priority: Critical
>  Labels: auto-deprioritized-major, stale-assigned, test-stability
>
> {code:java}
> Mar 31 10:49:17 === FAILURES 
> ===
> Mar 31 10:49:17 __ 
> EmbeddedThreadDependencyTests.test_add_python_file __
> Mar 31 10:49:17 
> Mar 31 10:49:17 self = 
>  testMethod=test_add_python_file>
> Mar 31 10:49:17 
> Mar 31 10:49:17 def test_add_python_file(self):
> Mar 31 10:49:17 python_file_dir = os.path.join(self.tempdir, 
> "python_file_dir_" + str(uuid.uuid4()))
> Mar 31 10:49:17 os.mkdir(python_file_dir)
> Mar 31 10:49:17 python_file_path = os.path.join(python_file_dir, 
> "test_dependency_manage_lib.py")
> Mar 31 10:49:17 with open(python_file_path, 'w') as f:
> Mar 31 10:49:17 f.write("def add_two(a):\nraise 
> Exception('This function should not be called!')")
> Mar 31 10:49:17 self.t_env.add_python_file(python_file_path)
> Mar 31 10:49:17 
> Mar 31 10:49:17 python_file_dir_with_higher_priority = os.path.join(
> Mar 31 10:49:17 self.tempdir, "python_file_dir_" + 
> str(uuid.uuid4()))
> Mar 31 10:49:17 os.mkdir(python_file_dir_with_higher_priority)
> Mar 31 10:49:17 python_file_path_higher_priority = 
> os.path.join(python_file_dir_with_higher_priority,
> Mar 31 10:49:17 
> "test_dependency_manage_lib.py")
> Mar 31 10:49:17 with open(python_file_path_higher_priority, 'w') as f:
> Mar 31 10:49:17 f.write("def add_two(a):\nreturn a + 2")
> Mar 31 10:49:17 
> self.t_env.add_python_file(python_file_path_higher_priority)
> Mar 31 10:49:17 
> Mar 31 10:49:17 def plus_two(i):
> Mar 31 10:49:17 from test_dependency_manage_lib import add_two
> Mar 31 10:49:17 return add_two(i)
> Mar 31 10:49:17 
> Mar 31 10:49:17 self.t_env.create_temporary_system_function(
> Mar 31 10:49:17 "add_two", udf(plus_two, DataTypes.BIGINT(), 
> DataTypes.BIGINT()))
> Mar 31 10:49:17 table_sink = source_sink_utils.TestAppendSink(
> Mar 31 10:49:17 ['a', 'b'], [DataTypes.BIGINT(), 
> DataTypes.BIGINT()])
> Mar 31 10:49:17 self.t_env.register_table_sink("Results", table_sink)
> Mar 31 10:49:17 t = self.t_env.from_elements([(1, 2), (2, 5), (3, 
> 1)], ['a', 'b'])
> Mar 31 10:49:17 >   t.select(expr.call("add_two", t.a), 
> t.a).execute_insert("Results").wait()
> Mar 31 10:49:17 
> Mar 31 10:49:17 pyflink/table/tests/test_dependency.py:63: 
> Mar 31 10:49:17 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ 
> Mar 31 10:49:17 pyflink/table/table_result.py:76: in wait
> Mar 31 10:49:17 get_method(self._j_table_result, "await")()
> Mar 31 10:49:17 
> .tox/py38-cython/lib/python3.8/site-packages/py4j/java_gateway.py:1321: in 
> __call__
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=34001=logs=821b528f-1eed-5598-a3b4-7f748b13f261=6bb545dd-772d-5d8c-f258-f5085fba3295=27239



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32605) JoinITCase.testFullOuterJoinWithMultipleKeys fails with TimeoutException: Futures timed out after [20 seconds]

2023-09-23 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32605?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32605:
---
  Labels: auto-deprioritized-major test-stability  (was: stale-major 
test-stability)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> JoinITCase.testFullOuterJoinWithMultipleKeys fails with TimeoutException: 
> Futures timed out after [20 seconds]
> --
>
> Key: FLINK-32605
> URL: https://issues.apache.org/jira/browse/FLINK-32605
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / RPC, Table SQL / API
>Affects Versions: 1.18.0
>Reporter: Sergey Nuyanzin
>Priority: Minor
>  Labels: auto-deprioritized-major, test-stability
>
> While execution of 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=51254=logs=a9db68b9-a7e0-54b6-0f98-010e0aff39e2=cdd32e0b-6047-565b-c58f-14054472f1be=11681
> there was an exception 
> {noformat}
> Jul 14 04:35:32 Caused by: java.lang.Exception: Could not create actor system
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaBootstrapTools.startLocalActorSystem(AkkaBootstrapTools.java:238)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils$AkkaRpcServiceBuilder.createAndStart(AkkaRpcServiceUtils.java:349)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils$AkkaRpcServiceBuilder.createAndStart(AkkaRpcServiceUtils.java:327)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils$AkkaRpcServiceBuilder.createAndStart(AkkaRpcServiceUtils.java:247)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.minicluster.MiniCluster.createLocalRpcService(MiniCluster.java:1188)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.minicluster.MiniCluster.start(MiniCluster.java:355)
> Jul 14 04:35:32   at 
> org.apache.flink.client.program.PerJobMiniClusterFactory.submitJob(PerJobMiniClusterFactory.java:77)
> Jul 14 04:35:32   at 
> org.apache.flink.client.deployment.executors.LocalExecutor.execute(LocalExecutor.java:85)
> Jul 14 04:35:32   at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:2218)
> Jul 14 04:35:32   at 
> org.apache.flink.table.planner.delegation.DefaultExecutor.executeAsync(DefaultExecutor.java:110)
> Jul 14 04:35:32   at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.executeQueryOperation(TableEnvironmentImpl.java:992)
> Jul 14 04:35:32   ... 102 more
> Jul 14 04:35:32 Caused by: java.util.concurrent.TimeoutException: Futures 
> timed out after [20 seconds]
> Jul 14 04:35:32   at 
> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:259)
> Jul 14 04:35:32   at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:263)
> Jul 14 04:35:32   at 
> scala.concurrent.Await$.$anonfun$result$1(package.scala:223)
> Jul 14 04:35:32   at 
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:57)
> Jul 14 04:35:32   at scala.concurrent.Await$.result(package.scala:146)
> Jul 14 04:35:32   at 
> akka.stream.SystemMaterializer.(SystemMaterializer.scala:90)
> Jul 14 04:35:32   at 
> akka.stream.SystemMaterializer$.createExtension(SystemMaterializer.scala:39)
> Jul 14 04:35:32   at 
> akka.stream.SystemMaterializer$.createExtension(SystemMaterializer.scala:32)
> Jul 14 04:35:32   at 
> akka.actor.ActorSystemImpl.registerExtension(ActorSystem.scala:1165)
> Jul 14 04:35:32   at 
> akka.actor.ActorSystemImpl.$anonfun$loadExtensions$1(ActorSystem.scala:1208)
> Jul 14 04:35:32   at scala.collection.Iterator.foreach(Iterator.scala:943)
> Jul 14 04:35:32   at 
> scala.collection.Iterator.foreach$(Iterator.scala:943)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.RobustActorSystem.create(RobustActorSystem.java:54)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaUtils.createActorSystem(AkkaUtils.java:421)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaBootstrapTools.startActorSystem(AkkaBootstrapTools.java:253)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaBootstrapTools.startLocalActorSystem(AkkaBootstrapTools.java:236)
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-28879) New File Sink s3 end-to-end test failed with Output hash mismatch

2023-09-23 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28879?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-28879:
---
Labels: stale-major test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> New File Sink s3 end-to-end test failed with Output hash mismatch
> -
>
> Key: FLINK-28879
> URL: https://issues.apache.org/jira/browse/FLINK-28879
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream, Connectors / FileSystem, Tests
>Affects Versions: 1.16.0, 1.17.2
>Reporter: Huang Xingbo
>Priority: Major
>  Labels: stale-major, test-stability
>
> {code:java}
> 2022-08-09T00:50:02.8229585Z Aug 09 00:50:02 FAIL File Streaming Sink: Output 
> hash mismatch.  Got 6037b01ca0ffc61a95c12cb475c661a8, expected 
> 6727342fdd3aae2129e61fc8f433fb6f.
> 2022-08-09T00:50:02.8230700Z Aug 09 00:50:02 head hexdump of actual:
> 2022-08-09T00:50:02.8477319Z Aug 09 00:50:02 000   E   r   r   o   r  
>  e   x   e   c   u   t   i   n   g
> 2022-08-09T00:50:02.8478206Z Aug 09 00:50:02 010   a   w   s   c   o  
>  m   m   a   n   d   :   s   3
> 2022-08-09T00:50:02.8479475Z Aug 09 00:50:02 020   c   p   -   -   q  
>  u   i   e   t   s   3   :   /   /
> 2022-08-09T00:50:02.8480205Z Aug 09 00:50:02 030   f   l   i   n   k   -  
>  i   n   t   e   g   r   a   t   i   o
> 2022-08-09T00:50:02.8480924Z Aug 09 00:50:02 040   n   -   t   e   s   t  
>  s   /   t   e   m   p   /   t   e   s
> 2022-08-09T00:50:02.8481612Z Aug 09 00:50:02 050   t   _   f   i   l   e  
>  _   s   i   n   k   -   1   d   3   d
> 2022-08-09T00:50:02.8483048Z Aug 09 00:50:02 060   4   0   0   8   -   b  
>  0   b   f   -   4   2   6   5   -   b
> 2022-08-09T00:50:02.8483618Z Aug 09 00:50:02 070   e   0   e   -   3   b  
>  9   f   7   8   2   c   5   5   2   d
> 2022-08-09T00:50:02.8484222Z Aug 09 00:50:02 080   /   h   o   s   t  
>  d   i   r   /   /   t   e   m   p   -
> 2022-08-09T00:50:02.8484831Z Aug 09 00:50:02 090   t   e   s   t   -   d  
>  i   r   e   c   t   o   r   y   -   2
> 2022-08-09T00:50:02.8485719Z Aug 09 00:50:02 0a0   3   9   3   7   7   8  
>  2   6   8   0   /   t   e   m   p   /
> 2022-08-09T00:50:02.8486427Z Aug 09 00:50:02 0b0   t   e   s   t   _   f  
>  i   l   e   _   s   i   n   k   -   1
> 2022-08-09T00:50:02.8487134Z Aug 09 00:50:02 0c0   d   3   d   4   0   0  
>  8   -   b   0   b   f   -   4   2   6
> 2022-08-09T00:50:02.8487826Z Aug 09 00:50:02 0d0   5   -   b   e   0   e  
>  -   3   b   9   f   7   8   2   c   5
> 2022-08-09T00:50:02.8488511Z Aug 09 00:50:02 0e0   5   2   d   -   -  
>  e   x   c   l   u   d   e   '   *
> 2022-08-09T00:50:02.8489202Z Aug 09 00:50:02 0f0   '   -   -   i   n  
>  c   l   u   d   e   '   *   /   p
> 2022-08-09T00:50:02.8489891Z Aug 09 00:50:02 100   a   r   t   -   [   !  
>  /   ]   *   '   -   -   r   e   c
> 2022-08-09T00:50:02.8490385Z Aug 09 00:50:02 110   u   r   s   i   v   e  
> \n
> 2022-08-09T00:50:02.8490822Z Aug 09 00:50:02 117
> 2022-08-09T00:50:02.8502212Z Aug 09 00:50:02 Stopping job timeout watchdog 
> (with pid=141134)
> 2022-08-09T00:50:06.8430959Z rm: cannot remove 
> '/home/vsts/work/1/s/flink-dist/target/flink-1.16-SNAPSHOT-bin/flink-1.16-SNAPSHOT/lib/flink-shaded-netty-tcnative-static-*.jar':
>  No such file or directory
> 2022-08-09T00:50:06.9278248Z Aug 09 00:50:06 
> 5ccfeb22307c2a88625a38b9537acc79001d1b29094ef40fd70692ce11407502
> 2022-08-09T00:50:06.9618147Z Aug 09 00:50:06 
> 5ccfeb22307c2a88625a38b9537acc79001d1b29094ef40fd70692ce11407502
> 2022-08-09T00:50:06.9645077Z Aug 09 00:50:06 [FAIL] Test script contains 
> errors.
> 2022-08-09T00:50:06.9666227Z Aug 09 00:50:06 Checking of logs skipped.
> 2022-08-09T00:50:06.9671891Z Aug 09 00:50:06 
> 2022-08-09T00:50:06.9673050Z Aug 09 00:50:06 [FAIL] 'New File Sink s3 
> end-to-end test' failed after 3 minutes and 42 seconds! Test exited with exit 
> code 1
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39667=logs=af184cdd-c6d8-5084-0b69-7e9c67b35f7a=160c9ae5-96fd-516e-1c91-deb81f59292a=4136



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31860) FlinkDeployments never finalize when namespace is deleted

2023-09-23 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31860?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31860:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> FlinkDeployments never finalize when namespace is deleted
> -
>
> Key: FLINK-31860
> URL: https://issues.apache.org/jira/browse/FLINK-31860
> Project: Flink
>  Issue Type: Bug
>  Components: Kubernetes Operator
>Affects Versions: kubernetes-operator-1.3.1
> Environment: Apache Flink Kubernetes Operator 1.3.1
> Kubernetes 1.24.9
>Reporter: Jayme Howard
>Assignee: Jayme Howard
>Priority: Blocker
>  Labels: pull-request-available, stale-assigned
>
> This appears to be a pretty straightforward issue, but I don't know the 
> codebase well enough to propose a fix.  When a FlinkDeployment is present in 
> a namespace, and the namespace is deleted, the FlinkDeployment never 
> reconciles and fails to complete its finalizer.  This leads to the namespace 
> being blocked from deletion indefinitely, requiring manual manipulation to 
> remove the finalizer on the FlinkDeployment.
>  
> Namespace conditions:
> {code:java}
> conditions:
> - lastTransitionTime: '2023-04-18T22:17:48Z'
>   message: All resources successfully discovered
>   reason: ResourcesDiscovered
>   status: 'False'
>   type: NamespaceDeletionDiscoveryFailure
> - lastTransitionTime: '2023-03-23T18:27:37Z'
>   message: All legacy kube types successfully parsed
>   reason: ParsedGroupVersions
>   status: 'False'
>   type: NamespaceDeletionGroupVersionParsingFailure
> - lastTransitionTime: '2023-03-23T18:27:37Z'
>   message: All content successfully deleted, may be waiting on finalization
>   reason: ContentDeleted
>   status: 'False'
>   type: NamespaceDeletionContentFailure
> - lastTransitionTime: '2023-03-23T18:27:37Z'
>   message: 'Some resources are remaining: flinkdeployments.flink.apache.org 
> has 2
> resource instances'
>   reason: SomeResourcesRemain
>   status: 'True'
>   type: NamespaceContentRemaining
> - lastTransitionTime: '2023-03-23T18:27:37Z'
>   message: 'Some content in the namespace has finalizers remaining: 
> flinkdeployments.flink.apache.org/finalizer
> in 2 resource instances'
>   reason: SomeFinalizersRemain
>   status: 'True'
>   type: NamespaceFinalizersRemaining
> phase: Terminating {code}
> FlinkDeployment example (some fields redacted):
> {code:java}
> apiVersion: flink.apache.org/v1beta1
> kind: FlinkDeployment
> metadata:
>   creationTimestamp: '2023-03-23T18:27:02Z'
>   deletionGracePeriodSeconds: 0
>   deletionTimestamp: '2023-03-23T18:27:35Z'
>   finalizers:
>   - flinkdeployments.flink.apache.org/finalizer
>   generation: 3
>   name: 
>   namespace: 
>   resourceVersion: '10565277081'
>   uid: e50d2683-6c0c-467e-b10c-fe0f4e404692
> spec:
>   flinkConfiguration:
>     taskmanager.numberOfTaskSlots: '2'
>   flinkVersion: v1_16
>   image: 
>   job:
>     args: []
>     entryClass: 
>     jarURI: 
>     parallelism: 2
>     state: running
>     upgradeMode: stateless
>   jobManager:
>     replicas: 1
>     resource:
>       cpu: 1
>       memory: 2048m
>   logConfiguration:
>     log4j-console.properties: '# This affects logging for both user code and 
> Flink      rootLogger.level = INFO      rootLogger.appenderRef.console.ref = 
> ConsoleAppender      rootLogger.appenderRef.rolling.ref = RollingFileAppender 
>      # Uncomment this if you want to _only_ change Flink''s logging      
> #logger.flink.name = org.apache.flink      #logger.flink.level = INFO      # 
> The following lines keep the log level of common libraries/connectors on      
> # log level INFO. The root logger does not override this. You have to 
> manually      # change the log levels here.      logger.akka.name = akka      
> logger.akka.level = INFO      logger.kafka.name= org.apache.kafka      
> logger.kafka.level = INFO      logger.hadoop.name = org.apache.hadoop      
> logger.hadoop.level = INFO      logger.zookeeper.name = org.apache.zookeeper  
>     logger.zookeeper.level = INFO      # Log all infos to the console      
> appender.console.name = ConsoleAppender      

[jira] [Updated] (FLINK-31951) Mix schema record source creates corrupt record

2023-09-22 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31951?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31951:
---
Labels: fix-provided pull-request-available stale-assigned  (was: 
fix-provided pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Mix schema record source creates corrupt record
> ---
>
> Key: FLINK-31951
> URL: https://issues.apache.org/jira/browse/FLINK-31951
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.17.0, 1.18.0
>Reporter: Kevin Tseng
>Assignee: Kevin Tseng
>Priority: Minor
>  Labels: fix-provided, pull-request-available, stale-assigned
> Fix For: 1.18.0
>
>
> This seems to be an unexpected side effect with how AvroDeserializationSchema 
> class was written.
> Sometimes we do not have control over what record comes through a Kafka Topic.
> In current implementation, if AvroDeserializationSchema encountered a record 
> byte array that does not conform to the specified Schema / SpecificRecord 
> type, it will cause future record to be deserialized incorrectly.
> Origin of the issue is with how
> {code:java}
> AvroDeserializationSchema.deserialize{code}
> handles exception, and how
> {code:java}
> AvroDeserializationSchema.checkAvroInitialized{code}
> handles initialization of Decoder object



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31825) Stopping minikube fails with timeout

2023-09-22 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31825?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31825:
---
Labels: stale-major test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Stopping minikube fails with timeout
> 
>
> Key: FLINK-31825
> URL: https://issues.apache.org/jira/browse/FLINK-31825
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hadoop Compatibility, Tests
>Affects Versions: 1.18.0
>Reporter: Sergey Nuyanzin
>Priority: Major
>  Labels: stale-major, test-stability
>
> Currently there is not so much information in logs...
> {noformat}
> Apr 13 01:27:02 * Stopping node "minikube"  ...
> ==
> === WARNING: This task took already 95% of the available time budget of 286 
> minutes ===
> ==
> ==
> The following Java processes are running (JPS)
> ==
> 243413 Jps
> ==
> Printing stack trace of Java process 243413
> ==
> 243413: No such process
> ==
> The following Java processes are running (JPS)
> ==
> 243516 Jps
> ==
> Printing stack trace of Java process 243516
> ==
> 243516: No such process
> =
> === WARNING: Killing task ===
> =
> Terminated
> Apr 13 05:48:53 [FAIL] Test script contains errors.
> {noformat}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=48096=logs=bbb1e2a2-a43c-55c8-fb48-5cfe7a8a0ca6=ba24ad14-6ea3-5ee3-c4ec-9e7cd2c9e754=5290



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-16987) FLIP-95: Add new table source and sink interfaces

2023-09-22 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16987?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-16987:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> FLIP-95: Add new table source and sink interfaces
> -
>
> Key: FLINK-16987
> URL: https://issues.apache.org/jira/browse/FLINK-16987
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API, Table SQL / Planner
>Reporter: Timo Walther
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> Proper support for handling changelogs, more efficient processing of data 
> through the new Blink planner, and unified interfaces that are DataStream API 
> agnostic make it necessary to rework the table source and sink interfaces.
> The goals of this FLIP are:
>  * *Simplify the current interface architecture*:
>  ** Merge upsert, retract, and append sinks.
>  ** Unify batch and streaming sources.
>  ** Unify batch and streaming sinks.
>  * *Allow sources to produce a changelog*:
>  ** UpsertTableSources have been requested a lot by users. Now is the time to 
> open the internal planner capabilities via the new interfaces.
>  ** According to FLIP-105, we would like to support changelogs for processing 
> formats such as [Debezium|https://debezium.io/].
>  * *Don't rely on DataStream API for source and sinks*:
>  ** According to FLIP-32, the Table API and SQL should be independent of the 
> DataStream API which is why the `table-common` module has no dependencies on 
> `flink-streaming-java`.
>  ** Source and sink implementations should only depend on the `table-common` 
> module after FLIP-27.
>  ** Until FLIP-27 is ready, we still put most of the interfaces in 
> `table-common` and strictly separate interfaces that communicate with a 
> planner and actual runtime reader/writers.
>  * *Implement efficient sources and sinks without planner dependencies*:
>  ** Make Blink's internal data structures available to connectors.
>  ** Introduce stable interfaces for data structures that can be marked as 
> `@PublicEvolving`.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32851) [JUnit5 Migration] The rest package of flink-runtime module

2023-09-21 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32851:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> [JUnit5 Migration] The rest package of flink-runtime module
> ---
>
> Key: FLINK-32851
> URL: https://issues.apache.org/jira/browse/FLINK-32851
> Project: Flink
>  Issue Type: Sub-task
>  Components: Tests
>Reporter: Rui Fan
>Assignee: Matt Wang
>Priority: Minor
>  Labels: pull-request-available, stale-assigned
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33056) NettyClientServerSslTest#testValidSslConnection fails on AZP

2023-09-21 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33056?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-33056:
---
Labels: stale-critical test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Critical but is unassigned and neither itself nor its Sub-Tasks have been 
updated for 14 days. I have gone ahead and marked it "stale-critical". If this 
ticket is critical, please either assign yourself or give an update. 
Afterwards, please remove the label or in 7 days the issue will be 
deprioritized.


> NettyClientServerSslTest#testValidSslConnection fails on AZP
> 
>
> Key: FLINK-33056
> URL: https://issues.apache.org/jira/browse/FLINK-33056
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Configuration, Runtime / Coordination
>Affects Versions: 1.19.0
>Reporter: Sergey Nuyanzin
>Priority: Critical
>  Labels: stale-critical, test-stability
> Attachments: logs-cron_azure-test_cron_azure_core-1694048924.zip
>
>
> This build 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=53020=logs=77a9d8e1-d610-59b3-fc2a-4766541e0e33=125e07e7-8de0-5c6c-a541-a567415af3ef=8592
> fails with 
> {noformat}
> Test testValidSslConnection[SSL provider = 
> JDK](org.apache.flink.runtime.io.network.netty.NettyClientServerSslTest) is 
> running.
> 
> 01:20:31,479 [main] INFO  
> org.apache.flink.runtime.io.network.netty.NettyConfig[] - NettyConfig 
> [server address: localhost/127.0.0.1, server port range: 36717, ssl enabled: 
> true, memory segment size (bytes): 1024, transport type: AUTO, number of 
> server threads: 1 (manual), number of client thr
> eads: 1 (manual), server connect backlog: 0 (use Netty's default), client 
> connect timeout (sec): 120, send/receive buffer size (bytes): 0 (use Netty's 
> default)]
> 01:20:31,479 [main] INFO  
> org.apache.flink.runtime.io.network.netty.NettyServer[] - Transport 
> type 'auto': using EPOLL.
> 01:20:31,475 [Flink Netty Client (42359) Thread 0] WARN  
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline [] - 
> An exceptionCaught() event was fired, and it reached at the tail of the 
> pipeline. It usually means the last handler in the pipeline did not handle 
> the exception.
> org.apache.flink.shaded.netty4.io.netty.handler.codec.DecoderException: 
> javax.net.ssl.SSLHandshakeException: server certificate with unknown 
> fingerprint: CN=Unknown, OU=Unknown, O=Unknown, L=Unknown, ST=Unknown, 
> C=Unknown
> at 
> org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:499)
>  ~[flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:290)
>  ~[flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> org.apache.flink.shaded.netty4.io.netty.channel.epoll.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:800)
>  [flink-shaded-netty-4.1.91.Final-17.0.jar:?]
> at 
> 

[jira] [Updated] (FLINK-32995) TPC-DS end-to-end test fails with chmod: cannot access '../target/generator/dsdgen_linux':

2023-09-21 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32995?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32995:
---
  Labels: auto-deprioritized-critical test-stability  (was: stale-critical 
test-stability)
Priority: Major  (was: Critical)

This issue was labeled "stale-critical" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Critical, 
please raise the priority and ask a committer to assign you the issue or revive 
the public discussion.


> TPC-DS end-to-end test fails with chmod: cannot access 
> '../target/generator/dsdgen_linux': 
> ---
>
> Key: FLINK-32995
> URL: https://issues.apache.org/jira/browse/FLINK-32995
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.19.0
>Reporter: Sergey Nuyanzin
>Priority: Major
>  Labels: auto-deprioritized-critical, test-stability
>
> This build 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=52773=logs=af184cdd-c6d8-5084-0b69-7e9c67b35f7a=0f3adb59-eefa-51c6-2858-3654d9e0749d=5504
>  fails as
> {noformat}
> Aug 29 10:03:20 [INFO] 10:03:20 Generating TPC-DS qualification data, this 
> need several minutes, please wait...
> chmod: cannot access '../target/generator/dsdgen_linux': No such file or 
> directory
> Aug 29 10:03:20 [FAIL] Test script contains errors.
> Aug 29 10:03:20 Checking for errors...
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33018) GCP Pubsub PubSubConsumingTest.testStoppingConnectorWhenDeserializationSchemaIndicatesEndOfStream failed

2023-09-20 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33018?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-33018:
---
  Labels: auto-deprioritized-blocker pull-request-available  (was: 
pull-request-available stale-blocker)
Priority: Critical  (was: Blocker)

This issue was labeled "stale-blocker" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually a Blocker, 
please raise the priority and ask a committer to assign you the issue or revive 
the public discussion.


> GCP Pubsub 
> PubSubConsumingTest.testStoppingConnectorWhenDeserializationSchemaIndicatesEndOfStream
>  failed
> 
>
> Key: FLINK-33018
> URL: https://issues.apache.org/jira/browse/FLINK-33018
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Google Cloud PubSub
>Affects Versions: gcp-pubsub-3.0.2
>Reporter: Martijn Visser
>Priority: Critical
>  Labels: auto-deprioritized-blocker, pull-request-available
>
> https://github.com/apache/flink-connector-gcp-pubsub/actions/runs/6061318336/job/16446392844#step:13:507
> {code:java}
> [INFO] 
> [INFO] Results:
> [INFO] 
> Error:  Failures: 
> Error:
> PubSubConsumingTest.testStoppingConnectorWhenDeserializationSchemaIndicatesEndOfStream:119
>  
> expected: ["1", "2", "3"]
>  but was: ["1", "2"]
> [INFO] 
> Error:  Tests run: 30, Failures: 1, Errors: 0, Skipped: 0
> [INFO] 
> [INFO] 
> 
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32223) Add Hive delegation token support

2023-09-20 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32223:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Add Hive delegation token support 
> --
>
> Key: FLINK-32223
> URL: https://issues.apache.org/jira/browse/FLINK-32223
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Hadoop Compatibility
>Reporter: qingbo jiao
>Assignee: qingbo jiao
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Attachments: image-2023-08-03-10-18-56-029.png, 
> image-2023-08-03-10-24-53-860.png, screenshot-1.png, screenshot-2.png, 
> screenshot-3.png
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32418) ClassNotFoundException when using flink-protobuf with sql-client

2023-09-20 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32418?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32418:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> ClassNotFoundException when using flink-protobuf with sql-client
> 
>
> Key: FLINK-32418
> URL: https://issues.apache.org/jira/browse/FLINK-32418
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table 
> SQL / Client
>Affects Versions: 1.16.2
>Reporter: Michael Kreis
>Assignee: Michael Kreis
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Attachments: full-stacktrace.log
>
>
> When the protobuf format in the kafka connector is used via the sql-client it 
> is not able to load the generated protobuf classes which are either passed 
> via `-j /protobuf-classes.jar` or added in the script via ADD JAR 
> '/protobuf-classes.jar'. The SHOW JARS command prints that the jar is loaded 
> but when the protobuf classes are loaded a ClassNotFoundException occurs.
> executed command:
> {code:java}
> sql-client.sh -f protobuf-table.sql -j /protobuf-classes.jar
> {code}
> protobuf-table.sql
> {code:sql}
> ADD JAR '/opt/sql-client/lib/flink-sql-connector-kafka-1.16.2.jar';
> ADD JAR '/opt/sql-client/lib/flink-protobuf-1.16.2.jar';
> SHOW JARS;
> CREATE TABLE POSITIONS(id BIGINT) WITH (
>   'connector' = 'kafka',
>   'format' = 'protobuf',
>   'topic' = 'protbuf-topic',
>   'properties.bootstrap.servers' = 'kafka:9092',
>   'properties.group.id' = 'flink-protobuf',
>   'properties.security.protocol' = 'SASL_PLAINTEXT',
>   'properties.sasl.mechanism' = 'SCRAM-SHA-512',
>   'properties.sasl.jaas.config' = 
> 'org.apache.kafka.common.security.scram.ScramLoginModule required 
> username="user" password="";',
>   'scan.startup.mode' = 'earliest-offset',
>   'protobuf.message-class-name' = 'com.example.protobuf.ProtoMessage',
>   'protobuf.ignore-parse-errors' = 'true'
>   );
> SELECT * FROM POSITIONS;
> {code}
> exception in the log:
> {code:java}
> Caused by: java.lang.ClassNotFoundException: com.example.protobuf.ProtoMessage
> at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(Unknown 
> Source)
> at 
> java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(Unknown 
> Source)
> at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
> at java.base/java.lang.Class.forName0(Native Method)
> at java.base/java.lang.Class.forName(Unknown Source)
> at 
> org.apache.flink.formats.protobuf.util.PbFormatUtils.getDescriptor(PbFormatUtils.java:89)
> ... 36 more
> {code}
> This also seems somehow related to FLINK-30318



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-24241) test_table_environment_api.py fail with NPE

2023-09-19 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-24241?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-24241:
---
  Labels: auto-deprioritized-major test-stability  (was: stale-major 
test-stability)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> test_table_environment_api.py fail with NPE
> ---
>
> Key: FLINK-24241
> URL: https://issues.apache.org/jira/browse/FLINK-24241
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python, Table SQL / Planner
>Affects Versions: 1.14.0, 1.15.0
>Reporter: Xintong Song
>Priority: Minor
>  Labels: auto-deprioritized-major, test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=23876=logs=821b528f-1eed-5598-a3b4-7f748b13f261=6bb545dd-772d-5d8c-f258-f5085fba3295=23263
> {code}
> Sep 10 03:03:39 E   py4j.protocol.Py4JJavaError: An error 
> occurred while calling o16211.execute.
> Sep 10 03:03:39 E   : java.lang.NullPointerException
> Sep 10 03:03:39 E at 
> java.util.Objects.requireNonNull(Objects.java:203)
> Sep 10 03:03:39 E at 
> org.apache.calcite.rel.metadata.RelMetadataQuery.(RelMetadataQuery.java:144)
> Sep 10 03:03:39 E at 
> org.apache.calcite.rel.metadata.RelMetadataQuery.(RelMetadataQuery.java:108)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.metadata.FlinkRelMetadataQuery.(FlinkRelMetadataQuery.java:78)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.metadata.FlinkRelMetadataQuery.instance(FlinkRelMetadataQuery.java:59)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.calcite.FlinkRelOptClusterFactory$$anon$1.get(FlinkRelOptClusterFactory.scala:39)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.calcite.FlinkRelOptClusterFactory$$anon$1.get(FlinkRelOptClusterFactory.scala:38)
> Sep 10 03:03:39 E at 
> org.apache.calcite.plan.RelOptCluster.getMetadataQuery(RelOptCluster.java:178)
> Sep 10 03:03:39 E at 
> org.apache.calcite.rel.metadata.RelMdUtil.clearCache(RelMdUtil.java:965)
> Sep 10 03:03:39 E at 
> org.apache.calcite.plan.hep.HepPlanner.buildFinalPlan(HepPlanner.java:942)
> Sep 10 03:03:39 E at 
> org.apache.calcite.plan.hep.HepPlanner.buildFinalPlan(HepPlanner.java:939)
> Sep 10 03:03:39 E at 
> org.apache.calcite.plan.hep.HepPlanner.findBestExp(HepPlanner.java:194)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkHepProgram.optimize(FlinkHepProgram.scala:69)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkHepRuleSetProgram.optimize(FlinkHepRuleSetProgram.scala:87)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkGroupProgram$$anonfun$optimize$1$$anonfun$apply$1.apply(FlinkGroupProgram.scala:63)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkGroupProgram$$anonfun$optimize$1$$anonfun$apply$1.apply(FlinkGroupProgram.scala:60)
> Sep 10 03:03:39 E at 
> scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
> Sep 10 03:03:39 E at 
> scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
> Sep 10 03:03:39 E at 
> scala.collection.Iterator$class.foreach(Iterator.scala:891)
> Sep 10 03:03:39 E at 
> scala.collection.AbstractIterator.foreach(Iterator.scala:1334)
> Sep 10 03:03:39 E at 
> scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
> Sep 10 03:03:39 E at 
> scala.collection.AbstractIterable.foreach(Iterable.scala:54)
> Sep 10 03:03:39 E at 
> scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157)
> Sep 10 03:03:39 E at 
> scala.collection.AbstractTraversable.foldLeft(Traversable.scala:104)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkGroupProgram$$anonfun$optimize$1.apply(FlinkGroupProgram.scala:60)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkGroupProgram$$anonfun$optimize$1.apply(FlinkGroupProgram.scala:55)
> Sep 10 03:03:39 E at 

[jira] [Updated] (FLINK-13876) Remove ExecutionConfig field from PojoSerializer

2023-09-19 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-13876:
---
Labels: auto-deprioritized-major stale-minor  (was: 
auto-deprioritized-major)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Minor but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 180 days. I have gone ahead and marked it "stale-minor". If this ticket is 
still Minor, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Remove ExecutionConfig field from PojoSerializer
> 
>
> Key: FLINK-13876
> URL: https://issues.apache.org/jira/browse/FLINK-13876
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Type Serialization System
>Affects Versions: 1.7.2, 1.8.1, 1.9.0, 1.10.0
>Reporter: Dawid Wysakowicz
>Priority: Minor
>  Labels: auto-deprioritized-major, stale-minor
>
> The PojoSerializers stores an instance of ExecutionConfig as internal field, 
> even though that the only information it may ever need are the registered 
> kryo serializers.
> This has a few drawbacks:
> * It blocks the evolution of {{ExecutionConfig}} as serializers where stored 
> in a state. Therefore any change to ExecutionConfig must be backwards 
> compatible in respect to java serialization
> * It probably already introduced a bug, as upon restore the Snapshot actually 
> recreates the serializer with an empty ExecutionConfig (see 
> org.apache.flink.api.java.typeutils.runtime.PojoSerializerSnapshot#restoreSerializer)
> I suggest to remove the field completely and adjust corresponding usages.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32582) Move TypeSerializerUpgradeTestBase from Kafka connector into flink-connector-common

2023-09-18 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32582?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32582:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Move TypeSerializerUpgradeTestBase from Kafka connector into 
> flink-connector-common
> ---
>
> Key: FLINK-32582
> URL: https://issues.apache.org/jira/browse/FLINK-32582
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Common
>Reporter: Matthias Pohl
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> The externalization of connectors made caused problems with the Flink's test 
> data generation. The Kafka connector relied on TypeSerializerUpgradeTestBase 
> for some test cases which was fine prior to FLINK-27518 where the test data 
> generation was handled individually.
> With FLINK-27518 the process was automated in Flink 1.18. For now, the 
> TypeSerializerUpgradeTestBase class was just copied over into the Kafka 
> connector since it was the only connector that would utilize this test base.
> But we might want to provide a more generalized solution where the test base 
> is provided by {{flink-connector-common}} to offer a generalized approach for 
> any connector.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33025) BatchArrowPythonOverWindowAggregateFunctionOperatorTest.testFinishBundleTriggeredByCount fails on AZP

2023-09-18 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33025?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-33025:
---
Labels: stale-critical test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Critical but is unassigned and neither itself nor its Sub-Tasks have been 
updated for 14 days. I have gone ahead and marked it "stale-critical". If this 
ticket is critical, please either assign yourself or give an update. 
Afterwards, please remove the label or in 7 days the issue will be 
deprioritized.


> BatchArrowPythonOverWindowAggregateFunctionOperatorTest.testFinishBundleTriggeredByCount
>  fails on AZP
> -
>
> Key: FLINK-33025
> URL: https://issues.apache.org/jira/browse/FLINK-33025
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.3, 1.19.0
>Reporter: Sergey Nuyanzin
>Priority: Critical
>  Labels: stale-critical, test-stability
>
> This build 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=52958=logs=5cae8624-c7eb-5c51-92d3-4d2dacedd221=5acec1b4-945b-59ca-34f8-168928ce5199=22618
>  fails on AZP as
> {noformat}
> Sep 03 05:05:38 05:05:38.220 [ERROR] Failures: 
> Sep 03 05:05:38 05:05:38.220 [ERROR]   
> BatchArrowPythonOverWindowAggregateFunctionOperatorTest.testFinishBundleTriggeredByCount:122->ArrowPythonAggregateFunctionOperatorTestBase.assertOutputEquals:62
>  
> Sep 03 05:05:38 Expected size: 4 but was: 3 in:
> Sep 03 05:05:38 [Record @ (undef) : +I(c1,c2,0,0,0),
> Sep 03 05:05:38 Record @ (undef) : +I(c1,c4,1,0,0),
> Sep 03 05:05:38 Record @ (undef) : +I(c1,c6,2,10,2)]
> {noformat}
> probably related to FLINK-26990



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32701) Potential Memory Leak in Flink CEP due to Persistent Starting States in NFAState

2023-09-18 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32701?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32701:
---
  Labels: CEP auto-deprioritized-critical cep  (was: CEP cep stale-critical)
Priority: Major  (was: Critical)

This issue was labeled "stale-critical" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Critical, 
please raise the priority and ask a committer to assign you the issue or revive 
the public discussion.


> Potential Memory Leak in Flink CEP due to Persistent Starting States in 
> NFAState
> 
>
> Key: FLINK-32701
> URL: https://issues.apache.org/jira/browse/FLINK-32701
> Project: Flink
>  Issue Type: Bug
>  Components: Library / CEP
>Affects Versions: 1.17.0, 1.16.1, 1.16.2, 1.17.1
>Reporter: Puneet Duggal
>Priority: Major
>  Labels: CEP, auto-deprioritized-critical, cep
> Attachments: Screenshot 2023-07-26 at 11.45.06 AM.png, Screenshot 
> 2023-07-26 at 11.50.28 AM.png
>
>
> Our team has encountered a potential memory leak issue while working with the 
> Complex Event Processing (CEP) library in Flink v1.17.
> h2. Context
> The CEP Operator maintains a keyed state called NFAState, which holds two 
> queues: one for partial matches and one for completed matches. When a key is 
> first encountered, the CEP creates a starting computation state and stores it 
> in the partial matches queue. As more events occur that match the defined 
> conditions (e.g., a TAKE condition), additional computation states get added 
> to the queue, with their specific type (normal, pending, end) depending on 
> the pattern sequence.
> However, I have noticed that the starting computation state remains in the 
> partial matches queue even after the pattern sequence has been completely 
> matched. This is also the case for keys that have already timed out. As a 
> result, the state gets stored for all keys that the CEP ever encounters, 
> leading to a continual increase in the checkpoint size.
> h2.  How to reproduce this
>  # Pattern Sequence - A not_followed_by B within 5 mins
>  # Time Characteristic - EventTime
>  # StateBackend - HashMapStateBackend
> On my local machine, I started this pipeline and started sending events at 
> the rate of 10 events per second (only A) and as expected after 5 mins, CEP 
> started sending pattern matched output with the same rate. But the issue was 
> that after every 2 mins (checkpoint interval), checkpoint size kept on 
> increasing. Expectation was that after 5 mins (2-3 checkpoints), checkpoint 
> size will remain constant since any window of 5 mins will consist of the same 
> number of unique keys (older ones will get matched or timed out hence removed 
> from state). But as you can see below attached images, checkpoint size kept 
> on increasing till 40 checkpoints (around 1.5hrs).
> P.S. - After 3 checkpoints (6 mins), the checkpoint size was around 1.78MB. 
> Hence assumption is that ideal checkpoint size for a 5 min window should be 
> less than 1.78MB.
> As you can see after 39 checkpoints, I triggered a savepoint for this 
> pipeline. After that I used a savepoint reader to investigate what all is 
> getting stored in CEP states. Below code investigates NFAState of CEPOperator 
> for potential memory leak.
> {code:java}
> import lombok.AllArgsConstructor;
> import lombok.Data;
> import lombok.NoArgsConstructor;
> import org.apache.flink.api.common.state.ValueState;
> import org.apache.flink.api.common.state.ValueStateDescriptor;
> import org.apache.flink.cep.nfa.NFAState;
> import org.apache.flink.cep.nfa.NFAStateSerializer;
> import org.apache.flink.configuration.Configuration;
> import org.apache.flink.runtime.state.filesystem.FsStateBackend;
> import org.apache.flink.state.api.OperatorIdentifier;
> import org.apache.flink.state.api.SavepointReader;
> import org.apache.flink.state.api.functions.KeyedStateReaderFunction;
> import org.apache.flink.streaming.api.datastream.DataStream;
> import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
> import org.apache.flink.util.Collector;
> import org.junit.jupiter.api.Test;
> import java.io.Serializable;
> import java.util.Objects;
> public class NFAStateReaderTest {
> private static final String NFA_STATE_NAME = "nfaStateName";
> @Test
> public void testNfaStateReader() throws Exception {
> StreamExecutionEnvironment environment = 
> StreamExecutionEnvironment.getExecutionEnvironment();
> SavepointReader savepointReader =
> SavepointReader.read(environment, 
> "file:///opt/flink/savepoints/savepoint-093404-9bc0a38654df", new 
> FsStateBackend("file:///abc"));
> DataStream stream = 
> 

[jira] [Updated] (FLINK-32512) SHOW JARS should not show the jars for temporary function

2023-09-18 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32512?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32512:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> SHOW JARS should not show the jars for temporary function
> -
>
> Key: FLINK-32512
> URL: https://issues.apache.org/jira/browse/FLINK-32512
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Gateway
>Affects Versions: 1.19.0
>Reporter: Fang Yong
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> According to 
> https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/show/#show-jars,
>  `SHOW JARS` should only list the jars added by `ADD JAR` statement, but 
> currently it also show the jars for `CREATE TEMPORARY FUNCTION`



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-21648) FLIP-151: Incremental snapshots for heap-based state backend

2023-09-17 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21648?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-21648:
---
Labels: auto-deprioritized-major auto-unassigned stale-assigned  (was: 
auto-deprioritized-major auto-unassigned)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> FLIP-151: Incremental snapshots for heap-based state backend
> 
>
> Key: FLINK-21648
> URL: https://issues.apache.org/jira/browse/FLINK-21648
> Project: Flink
>  Issue Type: New Feature
>  Components: Runtime / State Backends
>Reporter: Roman Khachatryan
>Assignee: Roman Khachatryan
>Priority: Major
>  Labels: auto-deprioritized-major, auto-unassigned, stale-assigned
> Fix For: 1.18.0
>
>
> Umbrella ticket for 
> [https://cwiki.apache.org/confluence/display/FLINK/FLIP-151%3A+Incremental+snapshots+for+heap-based+state+backend]
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32658) State should not be silently removed when ignore-unclaimed-state is false

2023-09-17 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32658?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32658:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> State should not be silently removed when ignore-unclaimed-state is false
> -
>
> Key: FLINK-32658
> URL: https://issues.apache.org/jira/browse/FLINK-32658
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.18.0, 1.17.1
>Reporter: Rui Fan
>Assignee: Rui Fan
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Attachments: screenshot-1.png
>
>
> When ignore-unclaimed-state is false and the old state is removed, flink 
> should throw exception. It's similar to removing a stateful operator.
> This case occurs not only when the user removes state, but also when the 
> operator is replaced. 
> For example: upgrade FlinkKafkaConsumer to KafkaSource. All logical are not 
> changed, so the operator id isn't changed. The KafkaSource cannot resume from 
> the state of FlinkKafkaConsumer. However, the new flink job can start, and 
> the state is silently removed in the new job.(The old state is not physically 
> discarded, it is still stored in the state backend, but the new code will 
> never use it.)
> It also brings an additional problem: the KafkaSource will snapshot 2 states, 
> it includes the new state of KafkaSource, and the union list state of 
> FlinkKafkaConsumer. Whenever a job resumes from checkpoint, the union List 
> state is inflated. Eventually the state size of kafka offset exceeded 200MB.
>  !screenshot-1.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-30699) Improved getRandomString method code format in the StringUtils class

2023-09-17 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-30699:
---
  Labels: auto-deprioritized-minor pull-request-available  (was: 
pull-request-available stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Improved getRandomString method code format in the StringUtils class
> 
>
> Key: FLINK-30699
> URL: https://issues.apache.org/jira/browse/FLINK-30699
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Core
>Reporter: Bingye Chen
>Priority: Not a Priority
>  Labels: auto-deprioritized-minor, pull-request-available
> Attachments: image-2023-01-16-18-13-56-912.png, 
> image-2023-01-16-18-14-12-939.png
>
>
>  
> !image-2023-01-16-18-13-56-912.png|width=398,height=148!
> !image-2023-01-16-18-14-12-939.png|width=398,height=114!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32575) Unified the Cpu of JobManager Name

2023-09-17 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32575?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32575:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Unified the Cpu of JobManager Name
> --
>
> Key: FLINK-32575
> URL: https://issues.apache.org/jira/browse/FLINK-32575
> Project: Flink
>  Issue Type: New Feature
>Reporter: Bo Cui
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> To set the jm CPU of the Yarn, use the `yarn.appmaster.vcores`. To set the jm 
> cpu of the k8s, use `kubernetes.jobmanager.cpu`. when there are yarn and k8s 
> clusters, managing these configurations is difficult. Add a unified name for 
> them for ease of use.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32806) EmbeddedJobResultStore keeps the non-dirty job entries forever

2023-09-17 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32806?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32806:
---
Labels: stale-assigned starter  (was: starter)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> EmbeddedJobResultStore keeps the non-dirty job entries forever
> --
>
> Key: FLINK-32806
> URL: https://issues.apache.org/jira/browse/FLINK-32806
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.18.0, 1.17.1, 1.19.0
>Reporter: Matthias Pohl
>Assignee: hk__lrzy
>Priority: Major
>  Labels: stale-assigned, starter
>
> The {{EmbeddedJobResultStore}} keeps the entries of cleaned-up jobs in-memory 
> forever. We might want to add a TTL to have those entries be removed after a 
> certain amount of time to allow maintaining the memory footprint of the 
> {{EmbeddedJobResultStore}}.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-30519) Add e2e tests for operator dynamic config

2023-09-16 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30519?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-30519:
---
Labels: stale-assigned stale-critical starter  (was: stale-assigned starter)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Critical but is unassigned and neither itself nor its Sub-Tasks have been 
updated for 14 days. I have gone ahead and marked it "stale-critical". If this 
ticket is critical, please either assign yourself or give an update. 
Afterwards, please remove the label or in 7 days the issue will be 
deprioritized.


> Add e2e tests for operator dynamic config
> -
>
> Key: FLINK-30519
> URL: https://issues.apache.org/jira/browse/FLINK-30519
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Kubernetes Operator
>Reporter: Gyula Fora
>Priority: Critical
>  Labels: stale-assigned, stale-critical, starter
>
> The dynamic config feature is currently not covered by e2e tests and is 
> subject to accidental regressions, as shown in:
> https://issues.apache.org/jira/browse/FLINK-30329
> We should add an e2e test that covers this



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32522) Kafka connector should depend on commons-collections instead of inherit from flink

2023-09-16 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32522:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Kafka connector should depend on commons-collections instead of inherit from 
> flink
> --
>
> Key: FLINK-32522
> URL: https://issues.apache.org/jira/browse/FLINK-32522
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Affects Versions: 1.17.1
>Reporter: Ran Tao
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
> Attachments: image-2023-07-03-20-15-47-608.png, 
> image-2023-07-03-20-16-03-031.png
>
>
> Currently, externalized sql connector rely on flink main repo. but flink main 
> repo has many test cases(especially in flink-python) reference 
> flink-sql-kafka-connector.
> If we change the dependencies(e.g.  commons-collections) in flink main repo, 
> it cause exception:
> !image-2023-07-03-20-15-47-608.png!
>  
> !https://user-images.githubusercontent.com/11287509/250120522-6b096a4f-83f0-4287-b7ad-d46b9371de4c.png!
>  
> So must add this dependency explicitly. Otherwise, it will cause external 
> connectors block the upgrade of flink main. Connectors shouldn't rely on 
> dependencies that may or may not be
> available in Flink itself. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32605) JoinITCase.testFullOuterJoinWithMultipleKeys fails with TimeoutException: Futures timed out after [20 seconds]

2023-09-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32605?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32605:
---
Labels: stale-major test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> JoinITCase.testFullOuterJoinWithMultipleKeys fails with TimeoutException: 
> Futures timed out after [20 seconds]
> --
>
> Key: FLINK-32605
> URL: https://issues.apache.org/jira/browse/FLINK-32605
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / RPC, Table SQL / API
>Affects Versions: 1.18.0
>Reporter: Sergey Nuyanzin
>Priority: Major
>  Labels: stale-major, test-stability
>
> While execution of 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=51254=logs=a9db68b9-a7e0-54b6-0f98-010e0aff39e2=cdd32e0b-6047-565b-c58f-14054472f1be=11681
> there was an exception 
> {noformat}
> Jul 14 04:35:32 Caused by: java.lang.Exception: Could not create actor system
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaBootstrapTools.startLocalActorSystem(AkkaBootstrapTools.java:238)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils$AkkaRpcServiceBuilder.createAndStart(AkkaRpcServiceUtils.java:349)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils$AkkaRpcServiceBuilder.createAndStart(AkkaRpcServiceUtils.java:327)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils$AkkaRpcServiceBuilder.createAndStart(AkkaRpcServiceUtils.java:247)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.minicluster.MiniCluster.createLocalRpcService(MiniCluster.java:1188)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.minicluster.MiniCluster.start(MiniCluster.java:355)
> Jul 14 04:35:32   at 
> org.apache.flink.client.program.PerJobMiniClusterFactory.submitJob(PerJobMiniClusterFactory.java:77)
> Jul 14 04:35:32   at 
> org.apache.flink.client.deployment.executors.LocalExecutor.execute(LocalExecutor.java:85)
> Jul 14 04:35:32   at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:2218)
> Jul 14 04:35:32   at 
> org.apache.flink.table.planner.delegation.DefaultExecutor.executeAsync(DefaultExecutor.java:110)
> Jul 14 04:35:32   at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.executeQueryOperation(TableEnvironmentImpl.java:992)
> Jul 14 04:35:32   ... 102 more
> Jul 14 04:35:32 Caused by: java.util.concurrent.TimeoutException: Futures 
> timed out after [20 seconds]
> Jul 14 04:35:32   at 
> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:259)
> Jul 14 04:35:32   at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:263)
> Jul 14 04:35:32   at 
> scala.concurrent.Await$.$anonfun$result$1(package.scala:223)
> Jul 14 04:35:32   at 
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:57)
> Jul 14 04:35:32   at scala.concurrent.Await$.result(package.scala:146)
> Jul 14 04:35:32   at 
> akka.stream.SystemMaterializer.(SystemMaterializer.scala:90)
> Jul 14 04:35:32   at 
> akka.stream.SystemMaterializer$.createExtension(SystemMaterializer.scala:39)
> Jul 14 04:35:32   at 
> akka.stream.SystemMaterializer$.createExtension(SystemMaterializer.scala:32)
> Jul 14 04:35:32   at 
> akka.actor.ActorSystemImpl.registerExtension(ActorSystem.scala:1165)
> Jul 14 04:35:32   at 
> akka.actor.ActorSystemImpl.$anonfun$loadExtensions$1(ActorSystem.scala:1208)
> Jul 14 04:35:32   at scala.collection.Iterator.foreach(Iterator.scala:943)
> Jul 14 04:35:32   at 
> scala.collection.Iterator.foreach$(Iterator.scala:943)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.RobustActorSystem.create(RobustActorSystem.java:54)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaUtils.createActorSystem(AkkaUtils.java:421)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaBootstrapTools.startActorSystem(AkkaBootstrapTools.java:253)
> Jul 14 04:35:32   at 
> org.apache.flink.runtime.rpc.akka.AkkaBootstrapTools.startLocalActorSystem(AkkaBootstrapTools.java:236)
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-28475) kafka connector won't stop when the stopping offset is zero

2023-09-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-28475:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> kafka connector won't stop when the stopping offset is zero
> ---
>
> Key: FLINK-28475
> URL: https://issues.apache.org/jira/browse/FLINK-28475
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Affects Versions: 1.15.0
>Reporter: Leo zhang
>Assignee: Leo zhang
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Fix For: 1.18.0
>
>
> when use kafka connector in bounded mode,and the stopping offset hapends to 
> be 0,the kafka connector won't stop,which is not expected.
> I had traced the code, and found the stopping offset will be set to empty 
> when it is zero, and an empty stopping offset means no stopping offset when 
> serialized. This leads to a wrong execution.
> I had fixed this in my personal branch,now I am logging this issue in Jira so 
> that I can make merge request.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29050) [JUnit5 Migration] Module: flink-hadoop-compatibility

2023-09-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29050?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-29050:
---
Labels: pull-request-available stale-assigned starter  (was: 
pull-request-available starter)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> [JUnit5 Migration] Module: flink-hadoop-compatibility
> -
>
> Key: FLINK-29050
> URL: https://issues.apache.org/jira/browse/FLINK-29050
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hadoop Compatibility, Tests
>Reporter: RocMarshal
>Assignee: RocMarshal
>Priority: Major
>  Labels: pull-request-available, stale-assigned, starter
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31390) Optimize the FlinkChangelogModeInferenceProgram by avoiding unnecessary traversals.

2023-09-14 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31390:
---
  Labels: auto-deprioritized-minor pull-request-available  (was: 
pull-request-available stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Optimize the FlinkChangelogModeInferenceProgram by avoiding unnecessary 
> traversals.
> ---
>
> Key: FLINK-31390
> URL: https://issues.apache.org/jira/browse/FLINK-31390
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner
>Reporter: Aitozi
>Priority: Not a Priority
>  Labels: auto-deprioritized-minor, pull-request-available
>
> We can avoid the unnecessary traversals of the RelNode tree, since we are 
> only interested in the first satisfied plan.
>  
> FlinkChangelogModeInferenceProgram
> {code:java}
> val updateKindTraitVisitor = new SatisfyUpdateKindTraitVisitor(context)
> val finalRoot = requiredUpdateKindTraits.flatMap {
>   requiredUpdateKindTrait =>
> updateKindTraitVisitor.visit(rootWithModifyKindSet, 
> requiredUpdateKindTrait)
> }
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32941) Table API Bridge `toDataStream(targetDataType)` function not working correctly for Java List

2023-09-14 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32941?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32941:
---
  Labels: auto-deprioritized-critical bridge  (was: bridge stale-critical)
Priority: Major  (was: Critical)

This issue was labeled "stale-critical" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Critical, 
please raise the priority and ask a committer to assign you the issue or revive 
the public discussion.


> Table API Bridge `toDataStream(targetDataType)` function not working 
> correctly for Java List
> 
>
> Key: FLINK-32941
> URL: https://issues.apache.org/jira/browse/FLINK-32941
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Reporter: Tan Kim
>Priority: Major
>  Labels: auto-deprioritized-critical, bridge
>
> When the code below is executed, only the first element of the list is 
> assigned to the List variable in MyPoJo repeatedly.
> {code:java}
> case class Item(
>   name: String
> )
> case class MyPojo(
>   @DataTypeHist("RAW") items: java.util.List[Item]
> )
> ...
> tableEnv
>   .sqlQuery("select items from table")
>   .toDataStream(DataTypes.of(classOf[MyPoJo])) {code}
>  
> For example, if you have the following list coming in as input,
> ["a","b","c"]
> The value actually stored in MyPojo's list variable is
> ["a","a","a"] 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32862) Support INIT operation type to be compatible with DTS on Alibaba Cloud

2023-09-14 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32862?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32862:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Support INIT operation type to be compatible with DTS on Alibaba Cloud
> --
>
> Key: FLINK-32862
> URL: https://issues.apache.org/jira/browse/FLINK-32862
> Project: Flink
>  Issue Type: Improvement
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Reporter: Hang Ruan
>Assignee: Hang Ruan
>Priority: Minor
>  Labels: pull-request-available, stale-assigned
>
> The operation type of canal json messages from DTS on Alibaba Cloud may 
> contain a new type `INIT`. We cannot handle these messages.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32995) TPC-DS end-to-end test fails with chmod: cannot access '../target/generator/dsdgen_linux':

2023-09-13 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32995?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32995:
---
Labels: stale-critical test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Critical but is unassigned and neither itself nor its Sub-Tasks have been 
updated for 14 days. I have gone ahead and marked it "stale-critical". If this 
ticket is critical, please either assign yourself or give an update. 
Afterwards, please remove the label or in 7 days the issue will be 
deprioritized.


> TPC-DS end-to-end test fails with chmod: cannot access 
> '../target/generator/dsdgen_linux': 
> ---
>
> Key: FLINK-32995
> URL: https://issues.apache.org/jira/browse/FLINK-32995
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.19.0
>Reporter: Sergey Nuyanzin
>Priority: Critical
>  Labels: stale-critical, test-stability
>
> This build 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=52773=logs=af184cdd-c6d8-5084-0b69-7e9c67b35f7a=0f3adb59-eefa-51c6-2858-3654d9e0749d=5504
>  fails as
> {noformat}
> Aug 29 10:03:20 [INFO] 10:03:20 Generating TPC-DS qualification data, this 
> need several minutes, please wait...
> chmod: cannot access '../target/generator/dsdgen_linux': No such file or 
> directory
> Aug 29 10:03:20 [FAIL] Test script contains errors.
> Aug 29 10:03:20 Checking for errors...
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32753) Print JVM flags on AZP

2023-09-13 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32753?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32753:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Print JVM flags on AZP
> --
>
> Key: FLINK-32753
> URL: https://issues.apache.org/jira/browse/FLINK-32753
> Project: Flink
>  Issue Type: Improvement
>  Components: Build System / Azure Pipelines
>Reporter: Zakelly Lan
>Assignee: Zakelly Lan
>Priority: Minor
>  Labels: pull-request-available, stale-assigned
> Fix For: 1.18.0
>
>
> I suggest printing JVM flags before the tests run, which could help 
> investigate the test failures (especially memory or GC related issue). An 
> example of pipeline output 
> [here|https://dev.azure.com/lzq82555906/flink-for-Zakelly/_build/results?buildId=122=logs=9dc1b5dc-bcfa-5f83-eaa7-0cb181ddc267=511d2595-ec54-5ab7-86ce-92f328796f20=165].
>  You may search 'JVM information' in this log.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33018) GCP Pubsub PubSubConsumingTest.testStoppingConnectorWhenDeserializationSchemaIndicatesEndOfStream failed

2023-09-12 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33018?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-33018:
---
Labels: pull-request-available stale-blocker  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as a 
Blocker but is unassigned and neither itself nor its Sub-Tasks have been 
updated for 1 days. I have gone ahead and marked it "stale-blocker". If this 
ticket is a Blocker, please either assign yourself or give an update. 
Afterwards, please remove the label or in 7 days the issue will be 
deprioritized.


> GCP Pubsub 
> PubSubConsumingTest.testStoppingConnectorWhenDeserializationSchemaIndicatesEndOfStream
>  failed
> 
>
> Key: FLINK-33018
> URL: https://issues.apache.org/jira/browse/FLINK-33018
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Google Cloud PubSub
>Affects Versions: gcp-pubsub-3.0.2
>Reporter: Martijn Visser
>Priority: Blocker
>  Labels: pull-request-available, stale-blocker
>
> https://github.com/apache/flink-connector-gcp-pubsub/actions/runs/6061318336/job/16446392844#step:13:507
> {code:java}
> [INFO] 
> [INFO] Results:
> [INFO] 
> Error:  Failures: 
> Error:
> PubSubConsumingTest.testStoppingConnectorWhenDeserializationSchemaIndicatesEndOfStream:119
>  
> expected: ["1", "2", "3"]
>  but was: ["1", "2"]
> [INFO] 
> Error:  Tests run: 30, Failures: 1, Errors: 0, Skipped: 0
> [INFO] 
> [INFO] 
> 
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-30774) Introduce flink-utils module

2023-09-12 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30774?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-30774:
---
  Labels: auto-deprioritized-major starter  (was: stale-major starter)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Introduce flink-utils module
> 
>
> Key: FLINK-30774
> URL: https://issues.apache.org/jira/browse/FLINK-30774
> Project: Flink
>  Issue Type: Improvement
>  Components: Build System
>Affects Versions: 1.17.0
>Reporter: Matthias Pohl
>Priority: Minor
>  Labels: auto-deprioritized-major, starter
>
> Currently, utility methods generic utility classes like {{Preconditions}} or 
> {{AbstractAutoCloseableRegistry}} are collected in {{flink-core}}. The flaw 
> of this approach is that we cannot use those classes in modules like 
> {{fink-migration-test-utils}}, {{flink-test-utils-junit}}, 
> {{flink-metrics-core}} or {{flink-annotations}}.
> We might want to have a generic {{flink-utils}} analogously to 
> {{flink-test-utils}} that collects Flink-independent utility functionality 
> that can be access by any module {{flink-core}} is depending on to make this 
> utility functionality available in any Flink-related module.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32564) Support cast from BYTES to NUMBER

2023-09-12 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32564:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Support cast from BYTES to NUMBER
> -
>
> Key: FLINK-32564
> URL: https://issues.apache.org/jira/browse/FLINK-32564
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Hanyu Zheng
>Assignee: Hanyu Zheng
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> We are dealing with a task that requires casting from the BYTES type to 
> BIGINT. Specifically, we have a string '00T1p'. Our approach is to convert 
> this string to BYTES and then cast the result to BIGINT with the following 
> SQL query:
> {code:java}
> SELECT CAST((CAST('00T1p' as BYTES)) as BIGINT);{code}
> However, an issue arises when executing this query, likely due to an error in 
> the conversion between BYTES and BIGINT. We aim to identify and rectify this 
> issue so our query can run correctly. The tasks involved are:
>  # Investigate and identify the specific reason for the failure of conversion 
> from BYTES to BIGINT.
>  # Design and implement a solution to ensure our query can function correctly.
>  # Test this solution across all required scenarios to guarantee its 
> functionality.
>  
> see also
> 1. PostgreSQL: PostgreSQL supports casting from BYTES type (BYTEA) to NUMBER 
> types (INTEGER, BIGINT, DECIMAL, etc.). In PostgreSQL, you can use CAST or 
> type conversion operator (::) for performing the conversion. URL: 
> [https://www.postgresql.org/docs/current/sql-expressions.html#SQL-SYNTAX-TYPE-CASTS]
> 2. MySQL: MySQL supports casting from BYTES type (BLOB or BINARY) to NUMBER 
> types (INTEGER, BIGINT, DECIMAL, etc.). In MySQL, you can use CAST or CONVERT 
> functions for performing the conversion. URL: 
> [https://dev.mysql.com/doc/refman/8.0/en/cast-functions.html]
> 3. Microsoft SQL Server: SQL Server supports casting from BYTES type 
> (VARBINARY, IMAGE) to NUMBER types (INT, BIGINT, NUMERIC, etc.). You can use 
> CAST or CONVERT functions for performing the conversion. URL: 
> [https://docs.microsoft.com/en-us/sql/t-sql/functions/cast-and-convert-transact-sql]
> 4. Oracle Database: Oracle supports casting from RAW type (equivalent to 
> BYTES) to NUMBER types (NUMBER, INTEGER, FLOAT, etc.). You can use the 
> TO_NUMBER function for performing the conversion. URL: 
> [https://docs.oracle.com/en/database/oracle/oracle-database/19/sqlrf/TO_NUMBER.html]
> 5. Apache Spark: Spark DataFrame supports casting binary types (BinaryType or 
> ByteType) to numeric types (IntegerType, LongType, DecimalType, etc.) by 
> using the {{cast}} function. URL: 
> [https://spark.apache.org/docs/latest/api/sql/#cast]
>  
> for the problem of bytes order may arise (little vs big endian). 
>  
> 1. Apache Hadoop: Hadoop, being an open-source framework, has to deal with 
> byte order issues across different platforms and architectures. The Hadoop 
> File System (HDFS) uses a technique called "sequence files," which include 
> metadata to describe the byte order of the data. This metadata ensures that 
> data is read and written correctly, regardless of the endianness of the 
> platform.
> 2. Apache Avro: Avro is a data serialization system used by various big data 
> frameworks like Hadoop and Apache Kafka. Avro uses a compact binary encoding 
> format that includes a marker for the byte order. This allows Avro to handle 
> endianness issues seamlessly when data is exchanged between systems with 
> different byte orders.
> 3. Apache Parquet: Parquet is a columnar storage format used in big data 
> processing frameworks like Apache Spark. Parquet uses a little-endian format 
> for encoding numeric values, which is the most common format on modern 
> systems. When reading or writing Parquet data, data processing engines 
> typically handle any necessary byte order conversions transparently.
> 4. Apache Spark: Spark is a popular big data processing engine that can 
> handle data on distributed systems. It relies on the underlying data formats 
> it reads (e.g., Avro, Parquet, ORC) to manage byte order issues. 

[jira] [Updated] (FLINK-31689) Filesystem sink fails when parallelism of compactor operator changed

2023-09-12 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31689?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31689:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Filesystem sink fails when parallelism of compactor operator changed
> 
>
> Key: FLINK-31689
> URL: https://issues.apache.org/jira/browse/FLINK-31689
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem
>Affects Versions: 1.16.1
>Reporter: jirawech.s
>Assignee: jirawech.s
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Attachments: HelloFlinkHadoopSink.java
>
>
> I encounter this error when i tried to use Filesystem sink with Table SQL. I 
> have not tested with Datastream API tho. You may refers to the error as below
> {code:java}
> // code placeholder
> java.util.NoSuchElementException
>   at java.util.ArrayList$Itr.next(ArrayList.java:864)
>   at 
> org.apache.flink.connector.file.table.stream.compact.CompactOperator.initializeState(CompactOperator.java:119)
>   at 
> org.apache.flink.streaming.api.operators.StreamOperatorStateHandler.initializeOperatorState(StreamOperatorStateHandler.java:122)
>   at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:286)
>   at 
> org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.initializeStateAndOpenOperators(RegularOperatorChain.java:106)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.restoreGates(StreamTask.java:700)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$1.call(StreamTaskActionExecutor.java:55)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.restoreInternal(StreamTask.java:676)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:643)
>   at 
> org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:948)
>   at 
> org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:917)
>   at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:741)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:563)
>   at java.lang.Thread.run(Thread.java:750) {code}
> I cannot attach the full reproducible code here, but you may follow my pseudo 
> code in attachment and reproducible steps below
> 1. Create Kafka source
> 2. Set state.savepoints.dir
> 3. Set Job parallelism to 1
> 4. Create FileSystem Sink
> 5. Run the job and trigger savepoint with API
> {noformat}
> curl -X POST localhost:8081/jobs/:jobId/savepoints -d '{"cancel-job": 
> false}'{noformat}
> {color:#172b4d}6. Cancel job, change parallelism to 2, and resume job from 
> savepoint{color}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-24241) test_table_environment_api.py fail with NPE

2023-09-11 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-24241?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-24241:
---
Labels: stale-major test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> test_table_environment_api.py fail with NPE
> ---
>
> Key: FLINK-24241
> URL: https://issues.apache.org/jira/browse/FLINK-24241
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python, Table SQL / Planner
>Affects Versions: 1.14.0, 1.15.0
>Reporter: Xintong Song
>Priority: Major
>  Labels: stale-major, test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=23876=logs=821b528f-1eed-5598-a3b4-7f748b13f261=6bb545dd-772d-5d8c-f258-f5085fba3295=23263
> {code}
> Sep 10 03:03:39 E   py4j.protocol.Py4JJavaError: An error 
> occurred while calling o16211.execute.
> Sep 10 03:03:39 E   : java.lang.NullPointerException
> Sep 10 03:03:39 E at 
> java.util.Objects.requireNonNull(Objects.java:203)
> Sep 10 03:03:39 E at 
> org.apache.calcite.rel.metadata.RelMetadataQuery.(RelMetadataQuery.java:144)
> Sep 10 03:03:39 E at 
> org.apache.calcite.rel.metadata.RelMetadataQuery.(RelMetadataQuery.java:108)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.metadata.FlinkRelMetadataQuery.(FlinkRelMetadataQuery.java:78)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.metadata.FlinkRelMetadataQuery.instance(FlinkRelMetadataQuery.java:59)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.calcite.FlinkRelOptClusterFactory$$anon$1.get(FlinkRelOptClusterFactory.scala:39)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.calcite.FlinkRelOptClusterFactory$$anon$1.get(FlinkRelOptClusterFactory.scala:38)
> Sep 10 03:03:39 E at 
> org.apache.calcite.plan.RelOptCluster.getMetadataQuery(RelOptCluster.java:178)
> Sep 10 03:03:39 E at 
> org.apache.calcite.rel.metadata.RelMdUtil.clearCache(RelMdUtil.java:965)
> Sep 10 03:03:39 E at 
> org.apache.calcite.plan.hep.HepPlanner.buildFinalPlan(HepPlanner.java:942)
> Sep 10 03:03:39 E at 
> org.apache.calcite.plan.hep.HepPlanner.buildFinalPlan(HepPlanner.java:939)
> Sep 10 03:03:39 E at 
> org.apache.calcite.plan.hep.HepPlanner.findBestExp(HepPlanner.java:194)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkHepProgram.optimize(FlinkHepProgram.scala:69)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkHepRuleSetProgram.optimize(FlinkHepRuleSetProgram.scala:87)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkGroupProgram$$anonfun$optimize$1$$anonfun$apply$1.apply(FlinkGroupProgram.scala:63)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkGroupProgram$$anonfun$optimize$1$$anonfun$apply$1.apply(FlinkGroupProgram.scala:60)
> Sep 10 03:03:39 E at 
> scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
> Sep 10 03:03:39 E at 
> scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
> Sep 10 03:03:39 E at 
> scala.collection.Iterator$class.foreach(Iterator.scala:891)
> Sep 10 03:03:39 E at 
> scala.collection.AbstractIterator.foreach(Iterator.scala:1334)
> Sep 10 03:03:39 E at 
> scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
> Sep 10 03:03:39 E at 
> scala.collection.AbstractIterable.foreach(Iterable.scala:54)
> Sep 10 03:03:39 E at 
> scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157)
> Sep 10 03:03:39 E at 
> scala.collection.AbstractTraversable.foldLeft(Traversable.scala:104)
> Sep 10 03:03:39 E at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkGroupProgram$$anonfun$optimize$1.apply(FlinkGroupProgram.scala:60)
> Sep 10 03:03:39 E at 
> 

[jira] [Updated] (FLINK-31857) Support pluginable observers mechanism

2023-09-11 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31857?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31857:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Support pluginable observers mechanism
> --
>
> Key: FLINK-31857
> URL: https://issues.apache.org/jira/browse/FLINK-31857
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Affects Versions: kubernetes-operator-1.5.0
>Reporter: Daren Wong
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> Currently, Kubernetes Operator uses AbstractFlinkResourceObserver to observe 
> Flink Job state, cluster state, JM/TM state, etc and update Custom Resource: 
> FlinkDeployment, FlinkSessionJob with the gathered info.
> This Jira is to introduce a Flink plugin to allow user to implement custom 
> observer logic after the default observer logic is run.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32525) Update commons-beanutils to 1.9.4

2023-09-10 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32525?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32525:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Update commons-beanutils to 1.9.4
> -
>
> Key: FLINK-32525
> URL: https://issues.apache.org/jira/browse/FLINK-32525
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Deployment / YARN
>Reporter: Martijn Visser
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> YARN still tests with commons-beanutils 1.8.3 with a remark that beanutil 
> 1.9+ doesn't work with Hadoop, but Hadoop 2.10.2 (which is our minimum 
> supported version) uses beanutils 1.9.4 itself, per 
> https://github.com/apache/hadoop/blob/rel/release-2.10.2/hadoop-project/pom.xml#L861-L863



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32044) Improve catalog name check to keep consistent about human-readable exception log in FunctionCatalog

2023-09-10 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32044?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32044:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Improve catalog name check to keep consistent about human-readable exception 
> log in FunctionCatalog 
> 
>
> Key: FLINK-32044
> URL: https://issues.apache.org/jira/browse/FLINK-32044
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / API
>Affects Versions: 1.17.0
>Reporter: Ran Tao
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> {code:java}
> Catalog catalog = catalogManager.getCatalog(catalogName).get(); {code}
>  
> We can do an improvement to check optional#get and throw more friendly log to 
> users like other list operations.  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-26115) Multiple Kafka connector tests failed due to The topic metadata failed to propagate to Kafka broker

2023-09-10 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-26115?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-26115:
---
  Labels: auto-deprioritized-critical auto-deprioritized-major 
auto-deprioritized-minor test-stability  (was: auto-deprioritized-critical 
auto-deprioritized-major stale-minor test-stability)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Multiple Kafka connector tests failed due to The topic metadata failed to 
> propagate to Kafka broker
> ---
>
> Key: FLINK-26115
> URL: https://issues.apache.org/jira/browse/FLINK-26115
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Affects Versions: 1.13.5, 1.14.3, 1.15.0
>Reporter: Yun Gao
>Priority: Not a Priority
>  Labels: auto-deprioritized-critical, auto-deprioritized-major, 
> auto-deprioritized-minor, test-stability
>
> This issues tracks all the related issues with "The topic metadata failed to 
> propagate to Kafka broker"



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-28881) PubSubConsumingTest.testStoppingConnectorWhenDeserializationSchemaIndicatesEndOfStream test failure

2023-09-10 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28881?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-28881:
---
  Labels: auto-deprioritized-major test-stability  (was: stale-major 
test-stability)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> PubSubConsumingTest.testStoppingConnectorWhenDeserializationSchemaIndicatesEndOfStream
>  test failure
> ---
>
> Key: FLINK-28881
> URL: https://issues.apache.org/jira/browse/FLINK-28881
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Google Cloud PubSub
>Affects Versions: 1.16.0, gcp-pubsub-3.0.0
>Reporter: Leonard Xu
>Priority: Minor
>  Labels: auto-deprioritized-major, test-stability
>
> {code:java}
> 2022-08-09T03:14:22.0113691Z Aug 09 03:14:22 [ERROR] 
> org.apache.flink.streaming.connectors.gcp.pubsub.PubSubConsumingTest.testStoppingConnectorWhenDeserializationSchemaIndicatesEndOfStream
>   Time elapsed: 1.867 s  <<< FAILURE!
> 2022-08-09T03:14:22.0114504Z Aug 09 03:14:22 java.lang.AssertionError: 
> 2022-08-09T03:14:22.0114903Z Aug 09 03:14:22 
> 2022-08-09T03:14:22.0115263Z Aug 09 03:14:22 Expected: <[1, 2, 3]>
> 2022-08-09T03:14:22.0115679Z Aug 09 03:14:22  but: was <[1, 2]>
> 2022-08-09T03:14:22.0116232Z Aug 09 03:14:22  at 
> org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
> 2022-08-09T03:14:22.0116871Z Aug 09 03:14:22  at 
> org.junit.Assert.assertThat(Assert.java:964)
> 2022-08-09T03:14:22.0117580Z Aug 09 03:14:22  at 
> org.junit.Assert.assertThat(Assert.java:930)
> 2022-08-09T03:14:22.0118460Z Aug 09 03:14:22  at 
> org.apache.flink.streaming.connectors.gcp.pubsub.PubSubConsumingTest.testStoppingConnectorWhenDeserializationSchemaIndicatesEndOfStream(PubSubConsumingTest.java:119)
> {code}
> CI link: 
> https://dev.azure.com/leonardBang/Azure_CI/_build/results?buildId=713=logs=3796201e-ea88-5776-0ea8-9ccca648a70c=8ca54b76-085e-5cf1-8060-2c500a258258



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-30935) Add KafkaSerializer deserialize check when using SimpleVersionedSerializer

2023-09-10 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-30935:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Add KafkaSerializer deserialize check when using SimpleVersionedSerializer
> --
>
> Key: FLINK-30935
> URL: https://issues.apache.org/jira/browse/FLINK-30935
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Reporter: Ran Tao
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> Current kafka many implemented serializers do not deal with version check 
> while other implementations of SimpleVersionedSerializer supports it.
> we can add it like many other connectors's implementation in case of 
> incompatible or corrupt state when restoring from checkpoint.
>  
> {code:java}
> @Override
> public int getVersion() {
> return CURRENT_VERSION;
> }
> @Override
> public KafkaPartitionSplit deserialize(int version, byte[] serialized) throws 
> IOException {
> try (ByteArrayInputStream bais = new ByteArrayInputStream(serialized);
> DataInputStream in = new DataInputStream(bais)) {
> String topic = in.readUTF();
> int partition = in.readInt();
> long offset = in.readLong();
> long stoppingOffset = in.readLong();
> return new KafkaPartitionSplit(
> new TopicPartition(topic, partition), offset, stoppingOffset);
> }
> } {code}
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32344) MongoDB connector support unbounded streaming read via ChangeStream feature

2023-09-10 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32344:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> MongoDB connector support unbounded streaming read via ChangeStream feature
> ---
>
> Key: FLINK-32344
> URL: https://issues.apache.org/jira/browse/FLINK-32344
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / MongoDB
>Affects Versions: mongodb-1.0.1
>Reporter: Jiabao Sun
>Assignee: Jiabao Sun
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> Change streams allow applications to access real-time data changes without 
> the complexity and risk of tailing the oplog. Applications can use change 
> streams to subscribe to all data changes on a single collection, a database, 
> or an entire deployment, and immediately react to them. Because change 
> streams use the aggregation framework, applications can also filter for 
> specific changes or transform the notifications at will.
> We can use MongoDB change streams feature to support unbounded streaming read 
> for mongodb connector.
> [Change 
> Streams|https://www.mongodb.com/docs/manual/changeStreams/#change-streams]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32512) SHOW JARS should not show the jars for temporary function

2023-09-10 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32512?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32512:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> SHOW JARS should not show the jars for temporary function
> -
>
> Key: FLINK-32512
> URL: https://issues.apache.org/jira/browse/FLINK-32512
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Gateway
>Affects Versions: 1.19.0
>Reporter: Fang Yong
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> According to 
> https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/show/#show-jars,
>  `SHOW JARS` should only list the jars added by `ADD JAR` statement, but 
> currently it also show the jars for `CREATE TEMPORARY FUNCTION`



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32575) Unified the Cpu of JobManager Name

2023-09-09 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32575?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32575:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Unified the Cpu of JobManager Name
> --
>
> Key: FLINK-32575
> URL: https://issues.apache.org/jira/browse/FLINK-32575
> Project: Flink
>  Issue Type: New Feature
>Reporter: Bo Cui
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> To set the jm CPU of the Yarn, use the `yarn.appmaster.vcores`. To set the jm 
> cpu of the k8s, use `kubernetes.jobmanager.cpu`. when there are yarn and k8s 
> clusters, managing these configurations is difficult. Add a unified name for 
> them for ease of use.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-30699) Improved getRandomString method code format in the StringUtils class

2023-09-09 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-30699:
---
Labels: pull-request-available stale-minor  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Minor but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 180 days. I have gone ahead and marked it "stale-minor". If this ticket is 
still Minor, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Improved getRandomString method code format in the StringUtils class
> 
>
> Key: FLINK-30699
> URL: https://issues.apache.org/jira/browse/FLINK-30699
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Core
>Reporter: Bingye Chen
>Priority: Minor
>  Labels: pull-request-available, stale-minor
> Attachments: image-2023-01-16-18-13-56-912.png, 
> image-2023-01-16-18-14-12-939.png
>
>
>  
> !image-2023-01-16-18-13-56-912.png|width=398,height=148!
> !image-2023-01-16-18-14-12-939.png|width=398,height=114!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-25447) batch query cannot generate plan when a sorted view into multi sinks

2023-09-09 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-25447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-25447:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> batch query cannot generate plan when a sorted view into multi sinks
> 
>
> Key: FLINK-25447
> URL: https://issues.apache.org/jira/browse/FLINK-25447
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.14.2
>Reporter: lincoln lee
>Assignee: Zheng yunhong
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Fix For: 1.18.0, 1.17.2, 1.19.0
>
>
> A batch query  write a sorted view into multi sinks will get a cannot plan 
> exception
> {code}
>   @Test
>   def testSortedResultIntoMultiSinks(): Unit = {
> util.tableEnv.executeSql(
>   s"""
>  |CREATE TABLE Src (
>  |  `a` INT,
>  |  `b` BIGINT,
>  |  `c` STRING,
>  |  `d` STRING,
>  |  `e` STRING
>  |) WITH (
>  |  'connector' = 'values',
>  |  'bounded' = 'true'
>  |)
>""".stripMargin)
> val query = "SELECT * FROM Src order by c"
> val table = util.tableEnv.sqlQuery(query)
> util.tableEnv.registerTable("sortedTable", table)
> util.tableEnv.executeSql(
>   s"""
>  |CREATE TABLE sink1 (
>  |  `a` INT,
>  |  `b` BIGINT,
>  |  `c` STRING
>  |) WITH (
>  |  'connector' = 'filesystem',
>  |  'format' = 'testcsv',
>  |  'path' = '/tmp/test'
>  |)
>""".stripMargin)
> util.tableEnv.executeSql(
>   s"""
>  |CREATE TABLE sink2 (
>  |  `a` INT,
>  |  `b` BIGINT,
>  |  `c` STRING,
>  |  `d` STRING
>  |) WITH (
>  |  'connector' = 'filesystem',
>  |  'format' = 'testcsv',
>  |  'path' = '/tmp/test'
>  |)
>   """.stripMargin)
> val stmtSet= util.tableEnv.createStatementSet()
> stmtSet.addInsertSql(
>   "insert into sink1 select a, b, listagg(d) from sortedTable group by a, 
> b")
> stmtSet.addInsertSql(
>   "insert into sink2 select a, b, c, d from sortedTable")
> util.verifyExecPlan(stmtSet)
>   }
> {code}
> {code}
>   org.apache.flink.table.api.TableException: Cannot generate a valid 
> execution plan for the given query: 
>   LogicalSink(table=[default_catalog.default_database.sink2], fields=[a, b, 
> c, d])
>   +- LogicalProject(inputs=[0..3])
>  +- LogicalTableScan(table=[[IntermediateRelTable_0]])
>   This exception indicates that the query uses an unsupported SQL feature.
>   Please check the documentation for the set of currently supported SQL 
> features.
>   at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkVolcanoProgram.optimize(FlinkVolcanoProgram.scala:76)
>   at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkChainedProgram.$anonfun$optimize$1(FlinkChainedProgram.scala:62)
>   at 
> scala.collection.TraversableOnce.$anonfun$foldLeft$1(TraversableOnce.scala:156)
>   at 
> scala.collection.TraversableOnce.$anonfun$foldLeft$1$adapted(TraversableOnce.scala:156)
>   at scala.collection.Iterator.foreach(Iterator.scala:937)
>   at scala.collection.Iterator.foreach$(Iterator.scala:937)
>   at scala.collection.AbstractIterator.foreach(Iterator.scala:1425)
>   at scala.collection.IterableLike.foreach(IterableLike.scala:70)
>   at scala.collection.IterableLike.foreach$(IterableLike.scala:69)
>   at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
>   at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:156)
>   at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:154)
>   at scala.collection.AbstractTraversable.foldLeft(Traversable.scala:104)
>   at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkChainedProgram.optimize(FlinkChainedProgram.scala:58)
>   at 
> 

[jira] [Updated] (FLINK-32522) Kafka connector should depend on commons-collections instead of inherit from flink

2023-09-08 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32522:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Kafka connector should depend on commons-collections instead of inherit from 
> flink
> --
>
> Key: FLINK-32522
> URL: https://issues.apache.org/jira/browse/FLINK-32522
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Affects Versions: 1.17.1
>Reporter: Ran Tao
>Priority: Major
>  Labels: pull-request-available, stale-major
> Attachments: image-2023-07-03-20-15-47-608.png, 
> image-2023-07-03-20-16-03-031.png
>
>
> Currently, externalized sql connector rely on flink main repo. but flink main 
> repo has many test cases(especially in flink-python) reference 
> flink-sql-kafka-connector.
> If we change the dependencies(e.g.  commons-collections) in flink main repo, 
> it cause exception:
> !image-2023-07-03-20-15-47-608.png!
>  
> !https://user-images.githubusercontent.com/11287509/250120522-6b096a4f-83f0-4287-b7ad-d46b9371de4c.png!
>  
> So must add this dependency explicitly. Otherwise, it will cause external 
> connectors block the upgrade of flink main. Connectors shouldn't rely on 
> dependencies that may or may not be
> available in Flink itself. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32396) Support timestamp for jdbc driver and gateway

2023-09-08 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32396?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32396:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Support timestamp for jdbc driver and gateway
> -
>
> Key: FLINK-32396
> URL: https://issues.apache.org/jira/browse/FLINK-32396
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / JDBC
>Affects Versions: 1.18.0
>Reporter: Fang Yong
>Assignee: Fang Yong
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> Support timestamp and timestamp_ltz data type for jdbc driver and sql-gateway



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32706) Add SPLIT(STRING) support in SQL & Table API

2023-09-08 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32706?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32706:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Add SPLIT(STRING) support in SQL & Table API
> 
>
> Key: FLINK-32706
> URL: https://issues.apache.org/jira/browse/FLINK-32706
> Project: Flink
>  Issue Type: Improvement
>Reporter: Hanyu Zheng
>Assignee: Hanyu Zheng
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> SPLIT Function
> Description
> Splits a string into an array of substrings, based on a delimiter.
> Syntax
> The syntax for the SPLIT function is:
> {code:java}
> SPLIT(col1, delimiter){code}
> Splits a string into an array of substrings based on a delimiter. If the 
> delimiter is not found, then the original string is returned as the only 
> element in the array. If the delimiter is empty, then all characters in the 
> string are split. If either, string or delimiter, are NULL, then a NULL value 
> is returned.
> If the delimiter is found at the beginning or end of the string, or there are 
> contiguous delimiters, then an empty space is added to the array.
> Example
> Let's look at some  function examples and explore how to use the SPLIT 
> function.
> For example:
>  
> {code:java}
> SELECT SPLIT('abcdefg', 'c');
> Result: ['ab', 'defg']
> {code}
> see also:
> 1. ksqlDB Split function
> ksqlDB provides a scalar function named {{SPLIT}} which splits a string into 
> an array of substrings based on a delimiter.
> Syntax: {{SPLIT(string, delimiter)}}
> For example: {{SPLIT('a,b,c', ',')}} will return {{{}['a', 'b', 'c']{}}}.
> [https://docs.ksqldb.io/en/0.8.1-ksqldb/developer-guide/ksqldb-reference/scalar-functions/#split]
> 2. Apache Hive Split function
> Hive offers a function named {{split}} which splits a string around a 
> specified delimiter and returns an array of strings.
> Syntax: {{array split(string str, string pat)}}
> For example: {{split('a,b,c', ',')}} will return {{{}["a", "b", "c"]{}}}.
> https://cwiki.apache.org/confluence/display/Hive/LanguageManual+UDF
> 3. Spark SQL Split function
> Spark SQL also offers a function named {{{}split{}}}, similar to the one in 
> Hive.
> Syntax: {{split(str, pattern[, limit])}}
> Here, {{limit}} is an optional parameter to specify the maximum length of the 
> returned array.
> For example: {{split('oneAtwoBthreeC', '[ABC]', 2)}} will return {{{}["one", 
> "twoBthreeC"]{}}}.
> [https://spark.apache.org/docs/latest/api/sql/index.html#split]
> 4. Presto Split function
> Presto offers a function named {{split}} which splits a string around a 
> regular expression and returns an array of strings.
> Syntax: {{array split(string str, string regex)}}
> For example: {{split('a.b.c', '\.')}} will return {{{}["a", "b", "c"]{}}}.
> [https://prestodb.io/docs/current/functions/string.html]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31335) using sql-gateway to submit job to yarn cluster, sql-gateway don't support kerberos

2023-09-08 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31335?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31335:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> using sql-gateway to submit job to yarn cluster, sql-gateway don't support 
> kerberos
> ---
>
> Key: FLINK-31335
> URL: https://issues.apache.org/jira/browse/FLINK-31335
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Gateway
>Affects Versions: 1.16.0
>Reporter: felixzh
>Assignee: felixzh
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Fix For: 1.17.2
>
>
> when submit job to yarn cluster, sql-gateway don't support kerberos.
> 1. yarn-per-job mode
> -Dexecution.target=yarn-per-job
> 2. yarn-session mode
> -Dexecution.target=yarn-session -Dyarn.application.id=yarnSessionAppID(eg: 
> application_1677479737242_0052)
> sql-gateway need to use SecurityUtils Modules.
> default use flink-conf.yaml(security.kerberos.login.principal and 
> security.kerberos.login.keytab), also support 
> -Dsecurity.kerberos.login.keytab and -Dsecurity.kerberos.login.principal (eg: 
> 1/2)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-20628) Port RabbitMQ Sources to FLIP-27 API

2023-09-07 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20628?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-20628:
---
Labels: auto-deprioritized-major pull-request-available stale-assigned  
(was: auto-deprioritized-major pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Port RabbitMQ Sources to FLIP-27 API
> 
>
> Key: FLINK-20628
> URL: https://issues.apache.org/jira/browse/FLINK-20628
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors/ RabbitMQ
>Reporter: Jan Westphal
>Assignee: RocMarshal
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available, 
> stale-assigned
>
> *Structure*
> The new RabbitMQ Source will have three components:
>  * RabbitMQ enumerator that receives one RabbitMQ Channel Config.
>  * RabbitMQ splits contain the RabbitMQ Channel Config
>  * RabbitMQ Readers which subscribe to the same RabbitMQ channel and receive 
> the messages (automatically load balanced by RabbitMQ).
> *Checkpointing Enumerators*
> The enumerator only needs to checkpoint the RabbitMQ channel config since the 
> continuous discovery of new unread/unhandled messages is taken care of by the 
> subscribed RabbitMQ readers and RabbitMQ itself.
> *Checkpointing Readers*
> The new RabbitMQ Source needs to ensure that every reader can be checkpointed.
> Since RabbitMQ is non-persistent and cannot be read by offset, a combined 
> usage of checkpoints and message acknowledgments is necessary. Until a 
> received message is checkpointed by a reader, it will stay in an 
> un-acknowledge state. As soon as the checkpoint is created, the messages from 
> the last checkpoint can be acknowledged as handled against RabbitMQ and thus 
> will be deleted only then. Messages need to be acknowledged one by one as 
> messages are handled by each SourceReader individually.
> When deserializing the messages we will make use of the implementation in the 
> existing RabbitMQ Source.
> *Message Delivery Guarantees* 
> Unacknowledged messages of a reader will be redelivered by RabbitMQ 
> automatically to other consumers of the same channel if the reader goes down.
>  
> This Source is going to only support at-least-once as this is the default 
> RabbitMQ behavior and thus everything else would require changes to RabbitMQ 
> itself or would impair the idea of parallelizing SourceReaders.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32509) avoid using skip in InputStreamFSInputWrapper.seek

2023-09-07 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32509?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32509:
---
  Labels: auto-deprioritized-major pull-request-available  (was: 
pull-request-available stale-major)
Priority: Minor  (was: Major)

This issue was labeled "stale-major" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Major, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> avoid using skip in InputStreamFSInputWrapper.seek
> --
>
> Key: FLINK-32509
> URL: https://issues.apache.org/jira/browse/FLINK-32509
> Project: Flink
>  Issue Type: Bug
>  Components: API / Core
>Affects Versions: 1.18.0
>Reporter: Libin Qin
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
>
> The implementation of  InputStream does not return -1  for eof.
> The java doc of InputStream said "The skip method may, for a variety of 
> reasons, end up skipping over some smaller number of bytes, possibly 0." 
> For FileInputStream, it allows skipping any number of bytes past the end of 
> the file.
> So the method "seek" of InputStreamFSInputWrapper will cause infinite loop if 
> desired exceed end of file
>  
> I reproduced with following case
>  
> {code:java}
> byte[] bytes = "flink".getBytes();
> try (InputStream inputStream = new ByteArrayInputStream(bytes)){ 
> InputStreamFSInputWrapper wrapper = new 
> InputStreamFSInputWrapper(inputStream); 
> wrapper.seek(20); 
> } {code}
> I  found an issue of commons-io talks about the problem of skip
> https://issues.apache.org/jira/browse/IO-203
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-12869) Add yarn acls capability to flink containers

2023-09-07 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-12869?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-12869:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Add yarn acls capability to flink containers
> 
>
> Key: FLINK-12869
> URL: https://issues.apache.org/jira/browse/FLINK-12869
> Project: Flink
>  Issue Type: New Feature
>  Components: Deployment / YARN
>Reporter: Nicolas Fraison
>Assignee: Archit Goyal
>Priority: Minor
>  Labels: pull-request-available, stale-assigned
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Yarn provide application acls mechanism to be able to provide specific rights 
> to other users than the one running the job (view logs through the 
> resourcemanager/job history, kill the application)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31238) Use IngestDB to speed up Rocksdb rescaling recovery

2023-09-07 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31238?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31238:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Use IngestDB to speed up Rocksdb rescaling recovery 
> 
>
> Key: FLINK-31238
> URL: https://issues.apache.org/jira/browse/FLINK-31238
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.16.1
>Reporter: Yue Ma
>Assignee: Yue Ma
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Attachments: image-2023-02-27-16-41-18-552.png, 
> image-2023-02-27-16-57-18-435.png, image-2023-03-07-14-27-10-260.png, 
> image-2023-03-09-15-23-30-581.png, image-2023-03-09-15-26-12-314.png, 
> image-2023-03-09-15-28-32-363.png, image-2023-03-09-15-41-03-074.png, 
> image-2023-03-09-15-41-08-379.png, image-2023-03-09-15-45-56-081.png, 
> image-2023-03-09-15-46-01-176.png, image-2023-03-09-15-50-04-281.png, 
> image-2023-03-29-15-25-21-868.png, image-2023-07-17-14-37-38-864.png, 
> image-2023-07-17-14-38-56-946.png, image-2023-07-22-14-16-31-856.png, 
> image-2023-07-22-14-19-01-390.png, image-2023-08-08-21-32-43-783.png, 
> image-2023-08-08-21-34-39-008.png, image-2023-08-08-21-39-39-135.png, 
> screenshot-1.png
>
>
> (The detailed design is in this document
> [https://docs.google.com/document/d/10MNVytTsyiDLZQSR89kDkVdmK_YjbM6jh0teerfDFfI|https://docs.google.com/document/d/10MNVytTsyiDLZQSR89kDkVdmK_YjbM6jh0teerfDFfI])
> There have been many discussions and optimizations in the community about 
> optimizing rocksdb scaling and recovery.
> https://issues.apache.org/jira/browse/FLINK-17971
> https://issues.apache.org/jira/browse/FLINK-8845
> https://issues.apache.org/jira/browse/FLINK-21321
> We hope to discuss some of our explorations under this ticket
> The process of scaling and recovering in rocksdb simply requires two steps
>  # Insert the valid keyGroup data of the new task.
>  # Delete the invalid data in the old stateHandle.
> The current method for data writing is to specify the main Db first and then 
> insert data using writeBatch.In addition, the method of deleteRange is 
> currently used to speed up the ClipDB. But in our production environment, we 
> found that the speed of rescaling is still very slow, especially when the 
> state of a single Task is large. 
>  
> We hope that the previous sst file can be reused directly when restoring 
> state, instead of retraversing the data. So we made some attempts to optimize 
> it in our internal version of flink and frocksdb.
>  
> We added two APIs *ClipDb* and *IngestDb* in frocksdb. 
>  * ClipDB is used to clip the data of a DB. Different from db.DeteleRange and 
> db.Delete, DeleteValue and RangeTombstone will not be generated for parts 
> beyond the key range. We will iterate over the FileMetaData of db. Process 
> each sst file. There are three situations here. 
> If all the keys of a file are required, we will keep the sst file and do 
> nothing 
> If all the keys of the sst file exceed the specified range, we will delete 
> the file directly. 
> If we only need some part of the sst file, we will rewrite the required keys 
> to generate a new sst file。
> All sst file changes will be placed in a VersionEdit, and the current 
> versions will LogAndApply this edit to ensure that these changes can take 
> effect
>  * IngestDb is used to directly ingest all sst files of one DB into another 
> DB. But it is necessary to strictly ensure that the keys of the two DBs do 
> not overlap, which is easy to do in the Flink scenario. The hard link method 
> will be used in the process of ingesting files, so it will be very fast. At 
> the same time, the file number of the main DB will be incremented 
> sequentially, and the SequenceNumber of the main DB will be updated to the 
> larger SequenceNumber of the two DBs.
> When IngestDb and ClipDb are supported, the state restoration logic is as 
> follows
>  * Open the first StateHandle as the main DB and pause the compaction.
>  * Clip the main DB according to the KeyGroup range of the Task with ClipDB

[jira] [Updated] (FLINK-27402) Unexpected boolean expression simplification for AND expression

2023-09-07 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-27402:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Unexpected boolean expression simplification  for AND expression 
> -
>
> Key: FLINK-27402
> URL: https://issues.apache.org/jira/browse/FLINK-27402
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: luoyuxia
>Assignee: Yunhong Zheng
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> Flink supports to compare between string and boolean, so the following sql 
> can work fine
>  
> {code:java}
> create table (c1 int, c2 string);
> select * from c2 = true;
> {code}
> But the following sql will throw excpetion 
>  
> {code:java}
> select * from c1 = 1 and c2 = true; {code}
> The reason it that Flink will try to simplify BOOLEAN expressions if possible 
> in 
> "c1 = 1 and c2 = true".
> So "c2 = true" will be simplified to "c2" by the following code in Flink:
>  
> {code:java}
> RexSimplify#simplifyAnd2ForUnknownAsFalse
> // Simplify BOOLEAN expressions if possible
> while (term.getKind() == SqlKind.EQUALS) {
> RexCall call = (RexCall) term;
> if (call.getOperands().get(0).isAlwaysTrue()) {
> term = call.getOperands().get(1);
> terms.set(i, term);
> continue;
> } else if (call.getOperands().get(1).isAlwaysTrue()) {
> term = call.getOperands().get(0);
> terms.set(i, term);
> continue;
> }
> break;
> } {code}
> So the expression will be reduced to ""c1 = 1 and c2". But AND requries both 
> sides are boolean expression and c2 is not a boolean expression for it 
> actually is a string.
> Then the exception "Boolean expression type expected" is thrown.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


  1   2   3   4   5   6   7   8   9   10   >