[jira] [Commented] (FLINK-32604) PyFlink end-to-end fails with kafka-server-stop.sh: No such file or directory

2023-07-26 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-32604?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17747430#comment-17747430
 ] 

Huang Xingbo commented on FLINK-32604:
--

It is a downloading error. We can keep observing the frequency of this failed 
case.


> PyFlink end-to-end fails  with kafka-server-stop.sh: No such file or 
> directory 
> ---
>
> Key: FLINK-32604
> URL: https://issues.apache.org/jira/browse/FLINK-32604
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.18.0
>Reporter: Sergey Nuyanzin
>Priority: Critical
>  Labels: test-stability
>
> This build 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=51253=logs=af184cdd-c6d8-5084-0b69-7e9c67b35f7a=0f3adb59-eefa-51c6-2858-3654d9e0749d=7883
> fails as
> {noformat}
> /home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/kafka-common.sh: line 
> 117: 
> /home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-27379214502/kafka_2.12-3.2.3/bin/kafka-server-stop.sh:
>  No such file or directory
> /home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/kafka-common.sh: line 
> 121: 
> /home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-27379214502/kafka_2.12-3.2.3/bin/zookeeper-server-stop.sh:
>  No such file or directory
> Jul 13 19:43:07 [FAIL] Test script contains errors.
> Jul 13 19:43:07 Checking of logs skipped.
> Jul 13 19:43:07 
> Jul 13 19:43:07 [FAIL] 'PyFlink end-to-end test' failed after 0 minutes and 
> 40 seconds! Test exited with exit code 1
> Jul 13 19:43:07 
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-31099) Chained WindowOperator throws NPE in PyFlink ThreadMode

2023-02-16 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31099?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo resolved FLINK-31099.
--
Resolution: Fixed

Merged into master via ca770b3d905936d8a93071210bd6542b6733221d
Merged into release-1.17 via c7c035a2413c04cd75948d8364e0770b97499901
Merged into release-1.16 via e3c0060e7fca53e0e01cb91e00607c8146b85604

> Chained WindowOperator throws NPE in PyFlink ThreadMode
> ---
>
> Key: FLINK-31099
> URL: https://issues.apache.org/jira/browse/FLINK-31099
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.17.0, 1.16.1
>Reporter: Huang Xingbo
>Assignee: Huang Xingbo
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0, 1.16.2
>
>
> Test case
> {code:python}
> config = Configuration()
> config.set_string("python.execution-mode", "process")
> env = StreamExecutionEnvironment.get_execution_environment(config)
> class MyTimestampAssigner(TimestampAssigner, ABC):
> def extract_timestamp(self, value: tuple, record_timestamp: int) -> int:
> return value[0]
> ds = env.from_collection(
> [(167646168, "a1", "b1", 1), (167646168, "a1", "b1", 1),
>  (167646168, "a2", "b2", 1), (167646168, "a1", "b2", 1),
>  (167646174, "a1", "b1", 1), (167646174, "a2", "b2", 1)]
> ).assign_timestamps_and_watermarks(
> 
> WatermarkStrategy.for_monotonous_timestamps().with_timestamp_assigner(MyTimestampAssigner())
> )
> ds.key_by(
> lambda x: (x[0], x[1], x[2])
> ).window(
> TumblingEventTimeWindows.of(Time.minutes(1))
> ).reduce(
> lambda x, y: (x[0], x[1], x[2], x[3] + y[3]),
> output_type=Types.TUPLE([Types.LONG(), Types.STRING(), Types.STRING(), 
> Types.INT()])
> # ).filter(
> # lambda x: x[1] == "a1"
> ).map(
> lambda x: (x[0], x[1], x[3]),
> output_type=Types.TUPLE([Types.LONG(), Types.STRING(), Types.INT()])
> ).print()
> env.execute()
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31099) Chained WindowOperator throws NPE in PyFlink ThreadMode

2023-02-15 Thread Huang Xingbo (Jira)
Huang Xingbo created FLINK-31099:


 Summary: Chained WindowOperator throws NPE in PyFlink ThreadMode
 Key: FLINK-31099
 URL: https://issues.apache.org/jira/browse/FLINK-31099
 Project: Flink
  Issue Type: Bug
  Components: API / Python
Affects Versions: 1.16.1, 1.17.0
Reporter: Huang Xingbo
Assignee: Huang Xingbo
 Fix For: 1.17.0, 1.16.2


Test case
{code:python}
config = Configuration()
config.set_string("python.execution-mode", "process")
env = StreamExecutionEnvironment.get_execution_environment(config)

class MyTimestampAssigner(TimestampAssigner, ABC):
def extract_timestamp(self, value: tuple, record_timestamp: int) -> int:
return value[0]

ds = env.from_collection(
[(167646168, "a1", "b1", 1), (167646168, "a1", "b1", 1),
 (167646168, "a2", "b2", 1), (167646168, "a1", "b2", 1),
 (167646174, "a1", "b1", 1), (167646174, "a2", "b2", 1)]
).assign_timestamps_and_watermarks(

WatermarkStrategy.for_monotonous_timestamps().with_timestamp_assigner(MyTimestampAssigner())
)
ds.key_by(
lambda x: (x[0], x[1], x[2])
).window(
TumblingEventTimeWindows.of(Time.minutes(1))
).reduce(
lambda x, y: (x[0], x[1], x[2], x[3] + y[3]),
output_type=Types.TUPLE([Types.LONG(), Types.STRING(), Types.STRING(), 
Types.INT()])
# ).filter(
# lambda x: x[1] == "a1"
).map(
lambda x: (x[0], x[1], x[3]),
output_type=Types.TUPLE([Types.LONG(), Types.STRING(), Types.INT()])
).print()
env.execute()
{code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-28786) Cannot run PyFlink 1.16 on MacOS with M1 chip

2023-02-14 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28786?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo resolved FLINK-28786.
--
Resolution: Fixed

Merged into master via 1e8d528e293da7ab990f1406fcd209ff4fb177b3
Merged into release-1.17 via 556a3faae9f147edf1c263e57a6661fc88b8f7ab
Merged into release-1.16 via 1e1461945d633486ee41f7137743837d0756e9b6

> Cannot run PyFlink 1.16 on MacOS with M1 chip
> -
>
> Key: FLINK-28786
> URL: https://issues.apache.org/jira/browse/FLINK-28786
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
>Reporter: Ran Tao
>Assignee: Huang Xingbo
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0, 1.16.2
>
>
> I have tested it with 2 m1 machines. i will reproduce the bug 100%.
> 1.m1 machine
> macos bigsur 11.5.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> 1.m1 machine
> macos monterey 12.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> reproduce step:
> 1.python -m pip install -r flink-python/dev/dev-requirements.txt
> 2.cd flink-python; python setup.py sdist bdist_wheel; cd 
> apache-flink-libraries; python setup.py sdist; cd ..;
> 3.python -m pip install apache-flink-libraries/dist/*.tar.gz
> 4.python -m pip install dist/*.whl
> when run 
> [word_count.py|https://nightlies.apache.org/flink/flink-docs-master/docs/dev/python/table_api_tutorial/]
>  it will cause
> {code:java}
> :219: RuntimeWarning: 
> apache_beam.coders.coder_impl.StreamCoderImpl size changed, may indicate 
> binary incompatibility. Expected 24 from C header, got 32 from PyObject
> Traceback (most recent call last):
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 129, in 
> word_count(known_args.input, known_args.output)
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 49, in word_count
> t_env = TableEnvironment.create(EnvironmentSettings.in_streaming_mode())
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 121, in create
> return TableEnvironment(j_tenv)
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 100, in __init__
> self._open()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1637, in _open
> startup_loopback_server()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1628, in startup_loopback_server
> from pyflink.fn_execution.beam.beam_worker_pool_service import \
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_worker_pool_service.py",
>  line 44, in 
> from pyflink.fn_execution.beam import beam_sdk_worker_main  # noqa # 
> pylint: disable=unused-import
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_sdk_worker_main.py",
>  line 21, in 
> import pyflink.fn_execution.beam.beam_operations # noqa # pylint: 
> disable=unused-import
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_operations.py",
>  line 27, in 
> from pyflink.fn_execution.state_impl import RemoteKeyedStateBackend, 
> RemoteOperatorStateBackend
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/state_impl.py",
>  line 33, in 
> from pyflink.fn_execution.beam.beam_coders import FlinkCoder
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_coders.py",
>  line 27, in 
> from pyflink.fn_execution.beam import beam_coder_impl_fast as 
> beam_coder_impl
>   File "pyflink/fn_execution/beam/beam_coder_impl_fast.pyx", line 1, in init 
> pyflink.fn_execution.beam.beam_coder_impl_fast
> KeyError: '__pyx_vtable__'
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-28786) Cannot run PyFlink 1.16 on MacOS with M1 chip

2023-02-07 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28786?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-28786:
-
Fix Version/s: 1.16.2
   (was: 1.16.1)

> Cannot run PyFlink 1.16 on MacOS with M1 chip
> -
>
> Key: FLINK-28786
> URL: https://issues.apache.org/jira/browse/FLINK-28786
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
>Reporter: Ran Tao
>Assignee: Huang Xingbo
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0, 1.16.2
>
>
> I have tested it with 2 m1 machines. i will reproduce the bug 100%.
> 1.m1 machine
> macos bigsur 11.5.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> 1.m1 machine
> macos monterey 12.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> reproduce step:
> 1.python -m pip install -r flink-python/dev/dev-requirements.txt
> 2.cd flink-python; python setup.py sdist bdist_wheel; cd 
> apache-flink-libraries; python setup.py sdist; cd ..;
> 3.python -m pip install apache-flink-libraries/dist/*.tar.gz
> 4.python -m pip install dist/*.whl
> when run 
> [word_count.py|https://nightlies.apache.org/flink/flink-docs-master/docs/dev/python/table_api_tutorial/]
>  it will cause
> {code:java}
> :219: RuntimeWarning: 
> apache_beam.coders.coder_impl.StreamCoderImpl size changed, may indicate 
> binary incompatibility. Expected 24 from C header, got 32 from PyObject
> Traceback (most recent call last):
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 129, in 
> word_count(known_args.input, known_args.output)
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 49, in word_count
> t_env = TableEnvironment.create(EnvironmentSettings.in_streaming_mode())
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 121, in create
> return TableEnvironment(j_tenv)
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 100, in __init__
> self._open()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1637, in _open
> startup_loopback_server()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1628, in startup_loopback_server
> from pyflink.fn_execution.beam.beam_worker_pool_service import \
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_worker_pool_service.py",
>  line 44, in 
> from pyflink.fn_execution.beam import beam_sdk_worker_main  # noqa # 
> pylint: disable=unused-import
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_sdk_worker_main.py",
>  line 21, in 
> import pyflink.fn_execution.beam.beam_operations # noqa # pylint: 
> disable=unused-import
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_operations.py",
>  line 27, in 
> from pyflink.fn_execution.state_impl import RemoteKeyedStateBackend, 
> RemoteOperatorStateBackend
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/state_impl.py",
>  line 33, in 
> from pyflink.fn_execution.beam.beam_coders import FlinkCoder
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_coders.py",
>  line 27, in 
> from pyflink.fn_execution.beam import beam_coder_impl_fast as 
> beam_coder_impl
>   File "pyflink/fn_execution/beam/beam_coder_impl_fast.pyx", line 1, in init 
> pyflink.fn_execution.beam.beam_coder_impl_fast
> KeyError: '__pyx_vtable__'
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Reopened] (FLINK-28786) Cannot run PyFlink 1.16 on MacOS with M1 chip

2023-02-07 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28786?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo reopened FLINK-28786:
--

> Cannot run PyFlink 1.16 on MacOS with M1 chip
> -
>
> Key: FLINK-28786
> URL: https://issues.apache.org/jira/browse/FLINK-28786
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
>Reporter: Ran Tao
>Assignee: Huang Xingbo
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0, 1.16.1
>
>
> I have tested it with 2 m1 machines. i will reproduce the bug 100%.
> 1.m1 machine
> macos bigsur 11.5.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> 1.m1 machine
> macos monterey 12.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> reproduce step:
> 1.python -m pip install -r flink-python/dev/dev-requirements.txt
> 2.cd flink-python; python setup.py sdist bdist_wheel; cd 
> apache-flink-libraries; python setup.py sdist; cd ..;
> 3.python -m pip install apache-flink-libraries/dist/*.tar.gz
> 4.python -m pip install dist/*.whl
> when run 
> [word_count.py|https://nightlies.apache.org/flink/flink-docs-master/docs/dev/python/table_api_tutorial/]
>  it will cause
> {code:java}
> :219: RuntimeWarning: 
> apache_beam.coders.coder_impl.StreamCoderImpl size changed, may indicate 
> binary incompatibility. Expected 24 from C header, got 32 from PyObject
> Traceback (most recent call last):
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 129, in 
> word_count(known_args.input, known_args.output)
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 49, in word_count
> t_env = TableEnvironment.create(EnvironmentSettings.in_streaming_mode())
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 121, in create
> return TableEnvironment(j_tenv)
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 100, in __init__
> self._open()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1637, in _open
> startup_loopback_server()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1628, in startup_loopback_server
> from pyflink.fn_execution.beam.beam_worker_pool_service import \
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_worker_pool_service.py",
>  line 44, in 
> from pyflink.fn_execution.beam import beam_sdk_worker_main  # noqa # 
> pylint: disable=unused-import
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_sdk_worker_main.py",
>  line 21, in 
> import pyflink.fn_execution.beam.beam_operations # noqa # pylint: 
> disable=unused-import
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_operations.py",
>  line 27, in 
> from pyflink.fn_execution.state_impl import RemoteKeyedStateBackend, 
> RemoteOperatorStateBackend
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/state_impl.py",
>  line 33, in 
> from pyflink.fn_execution.beam.beam_coders import FlinkCoder
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_coders.py",
>  line 27, in 
> from pyflink.fn_execution.beam import beam_coder_impl_fast as 
> beam_coder_impl
>   File "pyflink/fn_execution/beam/beam_coder_impl_fast.pyx", line 1, in init 
> pyflink.fn_execution.beam.beam_coder_impl_fast
> KeyError: '__pyx_vtable__'
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-28786) Cannot run PyFlink 1.16 on MacOS with M1 chip

2023-02-07 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-28786?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17685164#comment-17685164
 ] 

Huang Xingbo edited comment on FLINK-28786 at 2/7/23 9:12 AM:
--

If you encounter the following stack trace in m1

{code:java}
:219: RuntimeWarning: 
apache_beam.coders.coder_impl.StreamCoderImpl size changed, may indicate binary 
incompatibility. Expected 24 from C header, got 32 from PyObject
Traceback (most recent call last):
  File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
line 129, in 
word_count(known_args.input, known_args.output)
  File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
line 49, in word_count
t_env = TableEnvironment.create(EnvironmentSettings.in_streaming_mode())
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 121, in create
return TableEnvironment(j_tenv)
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 100, in __init__
self._open()
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 1637, in _open
startup_loopback_server()
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 1628, in startup_loopback_server
from pyflink.fn_execution.beam.beam_worker_pool_service import \
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_worker_pool_service.py",
 line 44, in 
from pyflink.fn_execution.beam import beam_sdk_worker_main  # noqa # 
pylint: disable=unused-import
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_sdk_worker_main.py",
 line 21, in 
import pyflink.fn_execution.beam.beam_operations # noqa # pylint: 
disable=unused-import
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_operations.py",
 line 27, in 
from pyflink.fn_execution.state_impl import RemoteKeyedStateBackend, 
RemoteOperatorStateBackend
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/state_impl.py",
 line 33, in 
from pyflink.fn_execution.beam.beam_coders import FlinkCoder
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_coders.py",
 line 27, in 
from pyflink.fn_execution.beam import beam_coder_impl_fast as 
beam_coder_impl
  File "pyflink/fn_execution/beam/beam_coder_impl_fast.pyx", line 1, in init 
pyflink.fn_execution.beam.beam_coder_impl_fast
KeyError: '__pyx_vtable__'
{code}

you can execute the following command to solve the problem

{code:java}
pip install cython==0.29.24
brew install gcc
pip uninstall apache-flink
pip uninstall apache-beam
pip install apache-flink
{code}




was (Author: hxbks2ks):
If you encounter the following stack trace in m1

{code:java}
:219: RuntimeWarning: 
apache_beam.coders.coder_impl.StreamCoderImpl size changed, may indicate binary 
incompatibility. Expected 24 from C header, got 32 from PyObject
Traceback (most recent call last):
  File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
line 129, in 
word_count(known_args.input, known_args.output)
  File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
line 49, in word_count
t_env = TableEnvironment.create(EnvironmentSettings.in_streaming_mode())
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 121, in create
return TableEnvironment(j_tenv)
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 100, in __init__
self._open()
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 1637, in _open
startup_loopback_server()
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 1628, in startup_loopback_server
from pyflink.fn_execution.beam.beam_worker_pool_service import \
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_worker_pool_service.py",
 line 44, in 
from pyflink.fn_execution.beam import beam_sdk_worker_main  # noqa # 
pylint: disable=unused-import
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_sdk_worker_main.py",
 line 21, in 
import pyflink.fn_execution.beam.beam_operations # noqa # pylint: 
disable=unused-import
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_operations.py",
 line 27, in 
from pyflink.fn_execution.state_impl import RemoteKeyedStateBackend, 
RemoteOperatorStateBackend
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/state_impl.py",
 line 33, in 
from pyflink.fn_execution.beam.beam_coders import 

[jira] [Comment Edited] (FLINK-28786) Cannot run PyFlink 1.16 on MacOS with M1 chip

2023-02-07 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-28786?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17685164#comment-17685164
 ] 

Huang Xingbo edited comment on FLINK-28786 at 2/7/23 9:11 AM:
--

If you encounter the following stack trace in m1

{code:java}
:219: RuntimeWarning: 
apache_beam.coders.coder_impl.StreamCoderImpl size changed, may indicate binary 
incompatibility. Expected 24 from C header, got 32 from PyObject
Traceback (most recent call last):
  File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
line 129, in 
word_count(known_args.input, known_args.output)
  File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
line 49, in word_count
t_env = TableEnvironment.create(EnvironmentSettings.in_streaming_mode())
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 121, in create
return TableEnvironment(j_tenv)
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 100, in __init__
self._open()
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 1637, in _open
startup_loopback_server()
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 1628, in startup_loopback_server
from pyflink.fn_execution.beam.beam_worker_pool_service import \
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_worker_pool_service.py",
 line 44, in 
from pyflink.fn_execution.beam import beam_sdk_worker_main  # noqa # 
pylint: disable=unused-import
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_sdk_worker_main.py",
 line 21, in 
import pyflink.fn_execution.beam.beam_operations # noqa # pylint: 
disable=unused-import
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_operations.py",
 line 27, in 
from pyflink.fn_execution.state_impl import RemoteKeyedStateBackend, 
RemoteOperatorStateBackend
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/state_impl.py",
 line 33, in 
from pyflink.fn_execution.beam.beam_coders import FlinkCoder
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_coders.py",
 line 27, in 
from pyflink.fn_execution.beam import beam_coder_impl_fast as 
beam_coder_impl
  File "pyflink/fn_execution/beam/beam_coder_impl_fast.pyx", line 1, in init 
pyflink.fn_execution.beam.beam_coder_impl_fast
KeyError: '__pyx_vtable__'
{code}

you can execute the following command to solve the problem

{code:java}
pip install cython==0.29.24
brew install gcc
pip uninstall apache-flink
pip install apache-flink
{code}




was (Author: hxbks2ks):
If you encounter the following stack trace in m1

{code:java}
:219: RuntimeWarning: 
apache_beam.coders.coder_impl.StreamCoderImpl size changed, may indicate binary 
incompatibility. Expected 24 from C header, got 32 from PyObject
Traceback (most recent call last):
  File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
line 129, in 
word_count(known_args.input, known_args.output)
  File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
line 49, in word_count
t_env = TableEnvironment.create(EnvironmentSettings.in_streaming_mode())
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 121, in create
return TableEnvironment(j_tenv)
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 100, in __init__
self._open()
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 1637, in _open
startup_loopback_server()
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 1628, in startup_loopback_server
from pyflink.fn_execution.beam.beam_worker_pool_service import \
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_worker_pool_service.py",
 line 44, in 
from pyflink.fn_execution.beam import beam_sdk_worker_main  # noqa # 
pylint: disable=unused-import
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_sdk_worker_main.py",
 line 21, in 
import pyflink.fn_execution.beam.beam_operations # noqa # pylint: 
disable=unused-import
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_operations.py",
 line 27, in 
from pyflink.fn_execution.state_impl import RemoteKeyedStateBackend, 
RemoteOperatorStateBackend
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/state_impl.py",
 line 33, in 
from pyflink.fn_execution.beam.beam_coders import FlinkCoder
  File 

[jira] [Commented] (FLINK-28786) Cannot run PyFlink 1.16 on MacOS with M1 chip

2023-02-07 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-28786?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17685164#comment-17685164
 ] 

Huang Xingbo commented on FLINK-28786:
--

If you encounter the following stack trace in m1

{code:java}
:219: RuntimeWarning: 
apache_beam.coders.coder_impl.StreamCoderImpl size changed, may indicate binary 
incompatibility. Expected 24 from C header, got 32 from PyObject
Traceback (most recent call last):
  File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
line 129, in 
word_count(known_args.input, known_args.output)
  File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
line 49, in word_count
t_env = TableEnvironment.create(EnvironmentSettings.in_streaming_mode())
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 121, in create
return TableEnvironment(j_tenv)
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 100, in __init__
self._open()
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 1637, in _open
startup_loopback_server()
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
 line 1628, in startup_loopback_server
from pyflink.fn_execution.beam.beam_worker_pool_service import \
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_worker_pool_service.py",
 line 44, in 
from pyflink.fn_execution.beam import beam_sdk_worker_main  # noqa # 
pylint: disable=unused-import
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_sdk_worker_main.py",
 line 21, in 
import pyflink.fn_execution.beam.beam_operations # noqa # pylint: 
disable=unused-import
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_operations.py",
 line 27, in 
from pyflink.fn_execution.state_impl import RemoteKeyedStateBackend, 
RemoteOperatorStateBackend
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/state_impl.py",
 line 33, in 
from pyflink.fn_execution.beam.beam_coders import FlinkCoder
  File 
"/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_coders.py",
 line 27, in 
from pyflink.fn_execution.beam import beam_coder_impl_fast as 
beam_coder_impl
  File "pyflink/fn_execution/beam/beam_coder_impl_fast.pyx", line 1, in init 
pyflink.fn_execution.beam.beam_coder_impl_fast
KeyError: '__pyx_vtable__'
{code}

you can execute the following command to solve the problem

{code:java}
cython==0.29.24
brew install gcc
pip uninstall apache-flink
pip install apache-flink
{code}



> Cannot run PyFlink 1.16 on MacOS with M1 chip
> -
>
> Key: FLINK-28786
> URL: https://issues.apache.org/jira/browse/FLINK-28786
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
>Reporter: Ran Tao
>Assignee: Huang Xingbo
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0, 1.16.1
>
>
> I have tested it with 2 m1 machines. i will reproduce the bug 100%.
> 1.m1 machine
> macos bigsur 11.5.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> 1.m1 machine
> macos monterey 12.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> reproduce step:
> 1.python -m pip install -r flink-python/dev/dev-requirements.txt
> 2.cd flink-python; python setup.py sdist bdist_wheel; cd 
> apache-flink-libraries; python setup.py sdist; cd ..;
> 3.python -m pip install apache-flink-libraries/dist/*.tar.gz
> 4.python -m pip install dist/*.whl
> when run 
> [word_count.py|https://nightlies.apache.org/flink/flink-docs-master/docs/dev/python/table_api_tutorial/]
>  it will cause
> {code:java}
> :219: RuntimeWarning: 
> apache_beam.coders.coder_impl.StreamCoderImpl size changed, may indicate 
> binary incompatibility. Expected 24 from C header, got 32 from PyObject
> Traceback (most recent call last):
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 129, in 
> word_count(known_args.input, known_args.output)
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 49, in word_count
> t_env = TableEnvironment.create(EnvironmentSettings.in_streaming_mode())
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 121, in create
> return TableEnvironment(j_tenv)
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 100, in __init__
> self._open()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1637, in _open
> 

[jira] [Closed] (FLINK-29421) Support python 3.10

2023-01-19 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-29421.

Resolution: Fixed

Merged into master via 838b79f5b9cc1a4cf253b2c17009f337bf569ecc

> Support python 3.10
> ---
>
> Key: FLINK-29421
> URL: https://issues.apache.org/jira/browse/FLINK-29421
> Project: Flink
>  Issue Type: New Feature
>  Components: API / Python
>Reporter: Eric Sirianni
>Assignee: Huang Xingbo
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>
> The {{apache-flink}} package fails to install on Python 3.10 due to inability 
> to compile {{numpy}}
> {noformat}
> numpy/core/src/multiarray/scalartypes.c.src:3242:12: error: too 
> few arguments to function ‘_Py_HashDouble’
>  3242 | return 
> _Py_HashDouble(npy_half_to_double(((PyHalfScalarObject *)obj)->obval));
>   |^~
> In file included from 
> /home/sirianni/.asdf/installs/python/3.10.6/include/python3.10/Python.h:77,
>  from 
> numpy/core/src/multiarray/scalartypes.c.src:3:
> 
> /home/sirianni/.asdf/installs/python/3.10.6/include/python3.10/pyhash.h:10:23:
>  note: declared here
>10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double);
> {noformat}
> Numpy issue https://github.com/numpy/numpy/issues/19033
> [Mailing list 
> thread|https://lists.apache.org/thread/f4r9hjt1l33xf5ngnswszhnls4cxkk52]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-30753) Py4J cannot acquire Table.explain() method

2023-01-19 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30753?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-30753:
-
Affects Version/s: 1.17.0
   (was: 1.16.0)

> Py4J cannot acquire Table.explain() method
> --
>
> Key: FLINK-30753
> URL: https://issues.apache.org/jira/browse/FLINK-30753
> Project: Flink
>  Issue Type: Technical Debt
>  Components: API / Python
>Affects Versions: 1.17.0
>Reporter: Yunfeng Zhou
>Assignee: Yunfeng Zhou
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=45044=logs=9cada3cb-c1d3-5621-16da-0f718fb86602=c67e71ed-6451-5d26-8920-5a8cf9651901



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-30753) Py4J cannot acquire Table.explain() method

2023-01-19 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30753?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-30753.

Fix Version/s: 1.17.0
 Assignee: Yunfeng Zhou
   Resolution: Fixed

Merged into master via 9ebb331be45dc2b094aca215e1062dad9800cd76

> Py4J cannot acquire Table.explain() method
> --
>
> Key: FLINK-30753
> URL: https://issues.apache.org/jira/browse/FLINK-30753
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Python
>Affects Versions: 1.16.0
>Reporter: Yunfeng Zhou
>Assignee: Yunfeng Zhou
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=45044=logs=9cada3cb-c1d3-5621-16da-0f718fb86602=c67e71ed-6451-5d26-8920-5a8cf9651901



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-30753) Py4J cannot acquire Table.explain() method

2023-01-19 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30753?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-30753:
-
Issue Type: Technical Debt  (was: Improvement)

> Py4J cannot acquire Table.explain() method
> --
>
> Key: FLINK-30753
> URL: https://issues.apache.org/jira/browse/FLINK-30753
> Project: Flink
>  Issue Type: Technical Debt
>  Components: API / Python
>Affects Versions: 1.16.0
>Reporter: Yunfeng Zhou
>Assignee: Yunfeng Zhou
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=45044=logs=9cada3cb-c1d3-5621-16da-0f718fb86602=c67e71ed-6451-5d26-8920-5a8cf9651901



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-30191) Update py4j from 0.10.9.3 to 0.10.9.7

2023-01-18 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30191?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-30191.

Fix Version/s: 1.17.0
   Resolution: Done

Merged into master via 0bbc7b1e9fed89b8c3e8ec67b7b0dad5999c2c01

> Update py4j from 0.10.9.3 to 0.10.9.7
> -
>
> Key: FLINK-30191
> URL: https://issues.apache.org/jira/browse/FLINK-30191
> Project: Flink
>  Issue Type: Technical Debt
>  Components: API / Python
>Reporter: Martijn Visser
>Assignee: Yunfeng Zhou
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29000) Support python UDF in the SQL Gateway

2023-01-18 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29000?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29000:
-
Fix Version/s: 1.17.0

> Support python UDF in the SQL Gateway
> -
>
> Key: FLINK-29000
> URL: https://issues.apache.org/jira/browse/FLINK-29000
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / Python, Table SQL / Gateway
>Affects Versions: 1.17.0
>Reporter: Shengkai Fang
>Assignee: Xingbo Huang
>Priority: Major
> Fix For: 1.17.0
>
>
> Currently Flink SQL Client supports python UDF, the Gateway should also 
> support this feature if the SQL Client is able to submit SQL to the Gateway.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-27716) Add Python API docs in ML

2023-01-18 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27716?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo reassigned FLINK-27716:


Assignee: Jiang Xin

> Add Python API docs in ML
> -
>
> Key: FLINK-27716
> URL: https://issues.apache.org/jira/browse/FLINK-27716
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Python, Documentation, Library / Machine Learning
>Reporter: Huang Xingbo
>Assignee: Jiang Xin
>Priority: Major
> Fix For: ml-2.2.0
>
>
> We can use sphinx same as pyflink or other tools to generate Python API docs 
> of ML



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-30619) AdaptiveSchedulerTest.testStatusMetrics is not stable

2023-01-17 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30619?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-30619:
-
Affects Version/s: 1.17.0

> AdaptiveSchedulerTest.testStatusMetrics is not stable
> -
>
> Key: FLINK-30619
> URL: https://issues.apache.org/jira/browse/FLINK-30619
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.16.0, 1.17.0
>Reporter: Matthias Pohl
>Priority: Critical
>  Labels: test-stability
>
> We experience a test instability in 
> {{AdaptiveSchedulerTest.testStatusMetrics}}.
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=44635=logs=0e7be18f-84f2-53f0-a32d-4a5e4a174679=7c1d86e3-35bd-5fd5-3b7c-30c126a78702=8475



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-30619) AdaptiveSchedulerTest.testStatusMetrics is not stable

2023-01-17 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-30619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17677690#comment-17677690
 ] 

Huang Xingbo commented on FLINK-30619:
--

same case in master: 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=44951=logs=0da23115-68bb-5dcd-192c-bd4c8adebde1=24c3384f-1bcb-57b3-224f-51bf973bbee8

> AdaptiveSchedulerTest.testStatusMetrics is not stable
> -
>
> Key: FLINK-30619
> URL: https://issues.apache.org/jira/browse/FLINK-30619
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.16.0
>Reporter: Matthias Pohl
>Priority: Critical
>  Labels: test-stability
>
> We experience a test instability in 
> {{AdaptiveSchedulerTest.testStatusMetrics}}.
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=44635=logs=0e7be18f-84f2-53f0-a32d-4a5e4a174679=7c1d86e3-35bd-5fd5-3b7c-30c126a78702=8475



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-30637) In linux-aarch64 environment, using “is” judgment to match the window type of overwindow have returned incorrect matching results

2023-01-16 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo resolved FLINK-30637.
--
Fix Version/s: 1.16.1
   1.15.4
 Assignee: Xin Chen
   Resolution: Fixed

Merged into master via d053867fb5c0fc647ea9266aab35598d7f3fc5c4
Merged into release-1.16 via eca940c5bf9e17c90dbb6f35e4ba370027137368
Merged into release-1.15 via 4035d61a2756ec16046fb687f533be0501fbbd35

> In linux-aarch64 environment, using “is” judgment to match the window type of 
> overwindow have returned incorrect matching results
> -
>
> Key: FLINK-30637
> URL: https://issues.apache.org/jira/browse/FLINK-30637
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.13.6
> Environment: Linux version 5.10.0-60.18.0.50.oe2203.aarch64
> (abuild@obs-worker-002) (gcc_old (GCC) 10.3.1, GNU ld (GNU Binutils) 2.37) #1 
> SMP Wed Mar 30 02:43:08 UTC 2022
>  
> pyflink-version:1.13.6
>Reporter: Xin Chen
>Assignee: Xin Chen
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.16.1, 1.15.4
>
>
> In  linux-arch64 environment, “window_type is 
> OverWindow.ROW_UNBOUNDED_FOLLOWING” in  in the 
> PandasBatchOverWindowAggregateFunctionOperation class of the pyflink source 
> code has returned the wrong result.
> For example, when window_type is 6, it represents the window type of 
> ‘ROW_UNBOUNDED_FOLLOWING’, but “window_type is 
> OverWindow.ROW_UNBOUNDED_FOLLOWING” return false because the memory address 
> of window_type has changed. It will lead to the wrong type of window, such as 
> row sliding window, so, the wrong input data of python udf have been 
> assembled and wrong results of that have appeared.
>  
> Specifically, the pyflink unit testcase is 
> ‘test_over_window_aggregate_function’ in 
> ‘pyflink\table\tests\test_pandas_udaf.py’. It performance incorrectly when I 
> execute it by pytest on linux-aarch64 system. I cut this unit use case to the 
> following code and executed it in the flink standalone mode of aarch64 
> system, and got the same error result:
>  
> {code:java}
> import unittest
> from pyflink.table import DataTypes, TableEnvironment, EnvironmentSettings
> from pyflink.table.udf import udaf, AggregateFunction
> class MaxAdd(AggregateFunction, unittest.TestCase):
> def open(self, function_context):
> mg = function_context.get_metric_group()
> self.counter = mg.add_group("key", "value").counter("my_counter")
> self.counter_sum = 0
> def get_value(self, accumulator):
> # counter
> self.counter.inc(10)
> self.counter_sum += 10
> return accumulator[0]
> def create_accumulator(self):
> return []
> def accumulate(self, accumulator, *args):
> result = 0
> for arg in args:
> result += arg.max()
> accumulator.append(result)
> @udaf(result_type=DataTypes.FLOAT(), func_type="pandas")
> def mean_udaf(v):
> import logging
> logging.error("debug")
> logging.error(v)
> return v.mean()
> t_env = TableEnvironment.create(
> 
> EnvironmentSettings.new_instance().in_batch_mode().use_blink_planner().build())
> t_env.get_config().get_configuration().set_string("parallelism.default", "2")
> t_env.get_config().get_configuration().set_string(
> "python.fn-execution.bundle.size", "1")
> import datetime
> t = t_env.from_elements(
> [
> (1, 2, 3, datetime.datetime(2018, 3, 11, 3, 10, 0, 0)),
> (1, 3, 1, datetime.datetime(2018, 3, 11, 3, 10, 0, 0)),
> (1, 8, 5, datetime.datetime(2018, 3, 11, 4, 20, 0, 0))
> ],
> DataTypes.ROW(
> [DataTypes.FIELD("a", DataTypes.TINYINT()),
>  DataTypes.FIELD("b", DataTypes.SMALLINT()),
>  DataTypes.FIELD("c", DataTypes.INT()),
>  DataTypes.FIELD("rowtime", DataTypes.TIMESTAMP(3))]))
> # sink
> t_env.execute_sql("""
> CREATE TABLE mySink (
>   c INT,
>   d FLOAT 
> ) WITH (
>   'connector' = 'print'
> )
> """)
> t_env.create_temporary_system_function("mean_udaf", mean_udaf)
> t_env.register_function("max_add", udaf(MaxAdd(),
> result_type=DataTypes.INT(),
> func_type="pandas"))
> t_env.register_table("T", t)
> t_env.execute_sql("""
> insert into mySink
> select
>  max_add(b, c)
>  over (PARTITION BY a ORDER BY rowtime
>  ROWS BETWEEN UNBOUNDED preceding AND 0 FOLLOWING),
>  mean_udaf(b)
>  over (PARTITION BY a ORDER BY rowtime
>  ROWS BETWEEN 1 PRECEDING AND UNBOUNDED FOLLOWING)
> from 

[jira] [Updated] (FLINK-29421) Support python 3.10

2023-01-16 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29421:
-
Fix Version/s: 1.17.0

> Support python 3.10
> ---
>
> Key: FLINK-29421
> URL: https://issues.apache.org/jira/browse/FLINK-29421
> Project: Flink
>  Issue Type: New Feature
>  Components: API / Python
>Reporter: Eric Sirianni
>Assignee: Huang Xingbo
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>
> The {{apache-flink}} package fails to install on Python 3.10 due to inability 
> to compile {{numpy}}
> {noformat}
> numpy/core/src/multiarray/scalartypes.c.src:3242:12: error: too 
> few arguments to function ‘_Py_HashDouble’
>  3242 | return 
> _Py_HashDouble(npy_half_to_double(((PyHalfScalarObject *)obj)->obval));
>   |^~
> In file included from 
> /home/sirianni/.asdf/installs/python/3.10.6/include/python3.10/Python.h:77,
>  from 
> numpy/core/src/multiarray/scalartypes.c.src:3:
> 
> /home/sirianni/.asdf/installs/python/3.10.6/include/python3.10/pyhash.h:10:23:
>  note: declared here
>10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double);
> {noformat}
> Numpy issue https://github.com/numpy/numpy/issues/19033
> [Mailing list 
> thread|https://lists.apache.org/thread/f4r9hjt1l33xf5ngnswszhnls4cxkk52]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-29421) Support python 3.10

2023-01-16 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo reassigned FLINK-29421:


Assignee: Huang Xingbo

> Support python 3.10
> ---
>
> Key: FLINK-29421
> URL: https://issues.apache.org/jira/browse/FLINK-29421
> Project: Flink
>  Issue Type: New Feature
>  Components: API / Python
>Reporter: Eric Sirianni
>Assignee: Huang Xingbo
>Priority: Minor
>  Labels: pull-request-available
>
> The {{apache-flink}} package fails to install on Python 3.10 due to inability 
> to compile {{numpy}}
> {noformat}
> numpy/core/src/multiarray/scalartypes.c.src:3242:12: error: too 
> few arguments to function ‘_Py_HashDouble’
>  3242 | return 
> _Py_HashDouble(npy_half_to_double(((PyHalfScalarObject *)obj)->obval));
>   |^~
> In file included from 
> /home/sirianni/.asdf/installs/python/3.10.6/include/python3.10/Python.h:77,
>  from 
> numpy/core/src/multiarray/scalartypes.c.src:3:
> 
> /home/sirianni/.asdf/installs/python/3.10.6/include/python3.10/pyhash.h:10:23:
>  note: declared here
>10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double);
> {noformat}
> Numpy issue https://github.com/numpy/numpy/issues/19033
> [Mailing list 
> thread|https://lists.apache.org/thread/f4r9hjt1l33xf5ngnswszhnls4cxkk52]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-26974) Python EmbeddedThreadDependencyTests.test_add_python_file failed on azure

2023-01-09 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-26974?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17656417#comment-17656417
 ] 

Huang Xingbo commented on FLINK-26974:
--

I'm very sorry, due to several other unexpected things in last week, it may 
take me a few more days to find the root cause of this unstable test. If this 
unstable case affects the development progress of other developers, please ping 
me at any time, and I will temporarily disable this test.

> Python EmbeddedThreadDependencyTests.test_add_python_file failed on azure
> -
>
> Key: FLINK-26974
> URL: https://issues.apache.org/jira/browse/FLINK-26974
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.15.0, 1.16.0, 1.17.0
>Reporter: Yun Gao
>Assignee: Huang Xingbo
>Priority: Critical
>  Labels: auto-deprioritized-major, test-stability
>
> {code:java}
> Mar 31 10:49:17 === FAILURES 
> ===
> Mar 31 10:49:17 __ 
> EmbeddedThreadDependencyTests.test_add_python_file __
> Mar 31 10:49:17 
> Mar 31 10:49:17 self = 
>  testMethod=test_add_python_file>
> Mar 31 10:49:17 
> Mar 31 10:49:17 def test_add_python_file(self):
> Mar 31 10:49:17 python_file_dir = os.path.join(self.tempdir, 
> "python_file_dir_" + str(uuid.uuid4()))
> Mar 31 10:49:17 os.mkdir(python_file_dir)
> Mar 31 10:49:17 python_file_path = os.path.join(python_file_dir, 
> "test_dependency_manage_lib.py")
> Mar 31 10:49:17 with open(python_file_path, 'w') as f:
> Mar 31 10:49:17 f.write("def add_two(a):\nraise 
> Exception('This function should not be called!')")
> Mar 31 10:49:17 self.t_env.add_python_file(python_file_path)
> Mar 31 10:49:17 
> Mar 31 10:49:17 python_file_dir_with_higher_priority = os.path.join(
> Mar 31 10:49:17 self.tempdir, "python_file_dir_" + 
> str(uuid.uuid4()))
> Mar 31 10:49:17 os.mkdir(python_file_dir_with_higher_priority)
> Mar 31 10:49:17 python_file_path_higher_priority = 
> os.path.join(python_file_dir_with_higher_priority,
> Mar 31 10:49:17 
> "test_dependency_manage_lib.py")
> Mar 31 10:49:17 with open(python_file_path_higher_priority, 'w') as f:
> Mar 31 10:49:17 f.write("def add_two(a):\nreturn a + 2")
> Mar 31 10:49:17 
> self.t_env.add_python_file(python_file_path_higher_priority)
> Mar 31 10:49:17 
> Mar 31 10:49:17 def plus_two(i):
> Mar 31 10:49:17 from test_dependency_manage_lib import add_two
> Mar 31 10:49:17 return add_two(i)
> Mar 31 10:49:17 
> Mar 31 10:49:17 self.t_env.create_temporary_system_function(
> Mar 31 10:49:17 "add_two", udf(plus_two, DataTypes.BIGINT(), 
> DataTypes.BIGINT()))
> Mar 31 10:49:17 table_sink = source_sink_utils.TestAppendSink(
> Mar 31 10:49:17 ['a', 'b'], [DataTypes.BIGINT(), 
> DataTypes.BIGINT()])
> Mar 31 10:49:17 self.t_env.register_table_sink("Results", table_sink)
> Mar 31 10:49:17 t = self.t_env.from_elements([(1, 2), (2, 5), (3, 
> 1)], ['a', 'b'])
> Mar 31 10:49:17 >   t.select(expr.call("add_two", t.a), 
> t.a).execute_insert("Results").wait()
> Mar 31 10:49:17 
> Mar 31 10:49:17 pyflink/table/tests/test_dependency.py:63: 
> Mar 31 10:49:17 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ 
> Mar 31 10:49:17 pyflink/table/table_result.py:76: in wait
> Mar 31 10:49:17 get_method(self._j_table_result, "await")()
> Mar 31 10:49:17 
> .tox/py38-cython/lib/python3.8/site-packages/py4j/java_gateway.py:1321: in 
> __call__
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=34001=logs=821b528f-1eed-5598-a3b4-7f748b13f261=6bb545dd-772d-5d8c-f258-f5085fba3295=27239



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-29796) pyflink protobuf requirement out of date

2023-01-04 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29796?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-29796.

Resolution: Fixed

> pyflink protobuf requirement out of date
> 
>
> Key: FLINK-29796
> URL: https://issues.apache.org/jira/browse/FLINK-29796
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
>Reporter: Jorge Villatoro
>Priority: Major
>
> The setup.py file for pyflink currently requires protobuf<3.18 but the 
> dev-requirements.txt file lists protubuf<=3.21 which seems to indicate that 
> the library works with newer version of protobuf. The latest version of 
> protobuf which satisfies the requirement was 3.17.3 which was released over a 
> year ago, and notably the various gcloud api packages all require much newer 
> versions (3.19+ I think). Obviously there are ways around this but the right 
> answer is likely to ease/change the requirement.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-28786) Cannot run PyFlink 1.16 on MacOS with M1 chip

2022-11-17 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28786?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-28786.

Resolution: Fixed

Merged into master via e5762a558f3697294cd73da4247a741fc6f73456
Merged into release-1.16 via 4f0ffa0ddfeffbe25435595d08825cada713ac44

> Cannot run PyFlink 1.16 on MacOS with M1 chip
> -
>
> Key: FLINK-28786
> URL: https://issues.apache.org/jira/browse/FLINK-28786
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
>Reporter: Ran Tao
>Assignee: Huang Xingbo
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0, 1.16.1
>
>
> I have tested it with 2 m1 machines. i will reproduce the bug 100%.
> 1.m1 machine
> macos bigsur 11.5.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> 1.m1 machine
> macos monterey 12.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> reproduce step:
> 1.python -m pip install -r flink-python/dev/dev-requirements.txt
> 2.cd flink-python; python setup.py sdist bdist_wheel; cd 
> apache-flink-libraries; python setup.py sdist; cd ..;
> 3.python -m pip install apache-flink-libraries/dist/*.tar.gz
> 4.python -m pip install dist/*.whl
> when run 
> [word_count.py|https://nightlies.apache.org/flink/flink-docs-master/docs/dev/python/table_api_tutorial/]
>  it will cause
> {code:java}
> :219: RuntimeWarning: 
> apache_beam.coders.coder_impl.StreamCoderImpl size changed, may indicate 
> binary incompatibility. Expected 24 from C header, got 32 from PyObject
> Traceback (most recent call last):
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 129, in 
> word_count(known_args.input, known_args.output)
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 49, in word_count
> t_env = TableEnvironment.create(EnvironmentSettings.in_streaming_mode())
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 121, in create
> return TableEnvironment(j_tenv)
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 100, in __init__
> self._open()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1637, in _open
> startup_loopback_server()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1628, in startup_loopback_server
> from pyflink.fn_execution.beam.beam_worker_pool_service import \
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_worker_pool_service.py",
>  line 44, in 
> from pyflink.fn_execution.beam import beam_sdk_worker_main  # noqa # 
> pylint: disable=unused-import
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_sdk_worker_main.py",
>  line 21, in 
> import pyflink.fn_execution.beam.beam_operations # noqa # pylint: 
> disable=unused-import
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_operations.py",
>  line 27, in 
> from pyflink.fn_execution.state_impl import RemoteKeyedStateBackend, 
> RemoteOperatorStateBackend
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/state_impl.py",
>  line 33, in 
> from pyflink.fn_execution.beam.beam_coders import FlinkCoder
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_coders.py",
>  line 27, in 
> from pyflink.fn_execution.beam import beam_coder_impl_fast as 
> beam_coder_impl
>   File "pyflink/fn_execution/beam/beam_coder_impl_fast.pyx", line 1, in init 
> pyflink.fn_execution.beam.beam_coder_impl_fast
> KeyError: '__pyx_vtable__'
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-29817) Published metadata for apache-flink in pypi are inconsistent and causes poetry to fail

2022-11-17 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29817?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-29817.

Resolution: Fixed

Merged into master via e5762a558f3697294cd73da4247a741fc6f73456
Merged into release-1.16 via d5b10d8ec9ca9fa03201ce57421bb0e714e224a7

> Published metadata for apache-flink in pypi are inconsistent and causes 
> poetry to fail
> --
>
> Key: FLINK-29817
> URL: https://issues.apache.org/jira/browse/FLINK-29817
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
> Environment: macos 12.6 (M1)
> Poetry Version: 1.2.2
> Python Version:  3.9.12
>Reporter: André Casimiro
>Assignee: Huang Xingbo
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.17.0, 1.16.1
>
>
> Hi, 
> Following the debug steps described in [this github 
> thread|https://github.com/python-poetry/poetry/issues/3011] I got to the 
> conclusion that the metadata of the apache-flink 1.16.0 package is wrong. And 
> because of that I cannot properly manage my dependencies using poetry.
> I can successfully install it with pip (runs with no errors), as stated in 
> the docs:
> {code:java}
> python -m pip install apache-flink {code}
> But when I try to include the dependency in my poetry project I got the 
> following error:
>  
> {code:java}
> ❯ poetry add apache-flink@1.16.0Updating dependencies
> Resolving dependencies... (2.0s)Because pemja (0.2.6) depends on numpy 
> (1.21.4)
>  and apache-flink (1.16.0) depends on numpy (>=1.14.3,<1.20), pemja (0.2.6) 
> is incompatible with apache-flink (1.16.0).
> So, because cv-features depends on apache-flink (1.16.0) which depends on 
> pemja (0.2.6), version solving failed. {code}
>  
> I've followed the same debug steps as in [this github 
> thread|https://github.com/python-poetry/poetry/issues/3011] and can confirm 
> that apache-flink has exactly the same problem as described in the thread: 
> the wheel package in pypi has correct dependency metadata but the pypi 
> published don't.
>  
> {code:java}
> ❯ pkginfo -f requires_dist 
> /Users/andre/Downloads/apache_flink-1.16.0-cp39-cp39-macosx_11_0_arm64.whl
> requires_dist: ['py4j (==0.10.9.3)', 'python-dateutil (==2.8.0)', 
> 'apache-beam (==2.38.0)', 'cloudpickle (==2.1.0)', 'avro-python3 
> (!=1.9.2,<1.10.0,>=1.8.1)', 'pytz (>=2018.3)', 'fastavro (<1.4.8,>=1.1.0)', 
> 'requests (>=2.26.0)', 'protobuf (<3.18)', 'httplib2 (<=0.20.4,>=0.19.0)', 
> 'apache-flink-libraries (<1.16.1,>=1.16.0)', 'numpy (<1.22.0,>=1.21.4)', 
> 'pandas (<1.4.0,>=1.3.0)', 'pyarrow (<9.0.0,>=5.0.0)', 'pemja (==0.2.6) ; 
> python_full_version >= "3.7" and platform_system != "Windows"'] {code}
> but the pipy json metadata is wrong:
>  
>  
> {code:java}
> ❯ curl -sL https://pypi.org/pypi/apache-flink/json | jq '.info.requires_dist'[
>   "py4j (==0.10.9.3)",
>   "python-dateutil (==2.8.0)",
>   "apache-beam (==2.38.0)",
>   "cloudpickle (==2.1.0)",
>   "avro-python3 (!=1.9.2,<1.10.0,>=1.8.1)",
>   "pytz (>=2018.3)",
>   "fastavro (<1.4.8,>=1.1.0)",
>   "requests (>=2.26.0)",
>   "protobuf (<3.18)",
>   "httplib2 (<=0.20.4,>=0.19.0)",
>   "apache-flink-libraries (<1.16.1,>=1.16.0)",
>   "numpy (<1.20,>=1.14.3)",
>   "pandas (<1.2.0,>=1.0)",
>   "pyarrow (<7.0.0,>=0.15.1)",
>   "pemja (==0.2.6) ; python_full_version >= \"3.7\" and platform_system != 
> \"Windows\""
> ]{code}
>  
> As per [this 
> comment|https://github.com/python-poetry/poetry/issues/3011#issuecomment-702826616],
>  could you please republish the package correcting this metadata information, 
> please? This [other 
> comment|https://github.com/apple/turicreate/issues/3342#issuecomment-702957550]
>  can help gain more context.
>  
> Thanks
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30040) Add more PyFlink API examples

2022-11-16 Thread Huang Xingbo (Jira)
Huang Xingbo created FLINK-30040:


 Summary: Add more PyFlink API examples
 Key: FLINK-30040
 URL: https://issues.apache.org/jira/browse/FLINK-30040
 Project: Flink
  Issue Type: Sub-task
  Components: API / Python, Documentation
Affects Versions: 1.16.0, 1.15.3
Reporter: Huang Xingbo


Many APIs of PyFlink currently do not provide corresponding examples



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-21223) Support to specify the input/output types of Python UDFs via string

2022-11-16 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo reassigned FLINK-21223:


Assignee: Huang Xingbo

> Support to specify the input/output types of Python UDFs via string
> ---
>
> Key: FLINK-21223
> URL: https://issues.apache.org/jira/browse/FLINK-21223
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Python
>Reporter: Dian Fu
>Assignee: Huang Xingbo
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> auto-unassigned
>
> Currently, users need to specify the input/output types as following:
> {code}
> {{@udf(result_type=DataTypes.BIGINT())
> def add(i, j):
>    return i + j
> }}{code}
> [FLIP-65|https://cwiki.apache.org/confluence/display/FLINK/FLIP-65%3A+New+type+inference+for+Table+API+UDFs]
>  makes it possible to support syntaxes as following:
> {code}
> {{@udf(result_type="BIGINT")
> def add(i, j):
>    return i + j
> }}{code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-29966) Replace and redesign the Python api documentation base

2022-11-15 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29966?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-29966.

Fix Version/s: 1.16.1
   1.15.4
   Resolution: Fixed

Merged into master via b838a4e76e226a0567d39db1f9c29305cb6d913f
Merged into release-1.16 via 7d5585dd9ea4f26bf28c43cbb5207f2a5b74420b
Merged into release-1.15 via d6bf30a013569c4a8e0cd546f3105296c7b32efe

> Replace and redesign the Python api documentation base
> --
>
> Key: FLINK-29966
> URL: https://issues.apache.org/jira/browse/FLINK-29966
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Python, Documentation
>Affects Versions: 1.16.0, 1.15.3
>Reporter: Huang Xingbo
>Assignee: Huang Xingbo
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.16.1, 1.15.4
>
>
> The doc of the existing python api is difficult to read and use. I have a 
> demo site for redesigning the Python api documentation base. See 
> https://pyflink-api-docs-test.readthedocs.io/en/latest/ as an example.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-28394) Python py36-cython: InvocationError for command install_command.sh fails with exit code 1

2022-11-14 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-28394?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17634177#comment-17634177
 ] 

Huang Xingbo commented on FLINK-28394:
--

`Python py36-cython` is not the cause of the actual error. Any test running 
under Python 3.6 fails, and then `Python py36-cython` will be displayed at the 
end. You need to scroll up in the log to find out the actual failed test. I 
looked at the links posted and they are all about 
https://issues.apache.org/jira/browse/FLINK-29461. 

> Python py36-cython: InvocationError for command install_command.sh fails with 
> exit code 1
> -
>
> Key: FLINK-28394
> URL: https://issues.apache.org/jira/browse/FLINK-28394
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0, 1.15.3
>Reporter: Martijn Visser
>Assignee: Huang Xingbo
>Priority: Major
>  Labels: stale-assigned, test-stability
>
> {code:java}
> Jul 05 03:47:22 Picked up JAVA_TOOL_OPTIONS: -XX:+HeapDumpOnOutOfMemoryError
> Jul 05 03:47:32 Using Python version 3.8.13 (default, Mar 28 2022 11:38:47)
> Jul 05 03:47:32 pip_test_code.py success!
> Jul 05 03:47:32 py38-cython finish: run-test  after 1658.14 seconds
> Jul 05 03:47:32 py38-cython start: run-test-post 
> Jul 05 03:47:32 py38-cython finish: run-test-post  after 0.00 seconds
> Jul 05 03:47:32 ___ summary 
> 
> Jul 05 03:47:32 ERROR:   py36-cython: InvocationError for command 
> /__w/3/s/flink-python/dev/install_command.sh --exists-action w 
> .tox/.tmp/package/1/apache-flink-1.15.dev0.zip (exited with code 1)
> Jul 05 03:47:32   py37-cython: commands succeeded
> Jul 05 03:47:32   py38-cython: commands succeeded
> Jul 05 03:47:32 cleanup 
> /__w/3/s/flink-python/.tox/.tmp/package/1/apache-flink-1.15.dev0.zip
> Jul 05 03:47:33 tox checks... [FAILED]
> Jul 05 03:47:33 Process exited with EXIT CODE: 1.
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=37604=logs=bf5e383b-9fd3-5f02-ca1c-8f788e2e76d3=85189c57-d8a0-5c9c-b61d-fc05cfac62cf=27789



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-28957) Optimize Python API doc

2022-11-10 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28957?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo reassigned FLINK-28957:


Assignee: Huang Xingbo

> Optimize Python API doc
> ---
>
> Key: FLINK-28957
> URL: https://issues.apache.org/jira/browse/FLINK-28957
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Python, Documentation
>Affects Versions: 1.14.5, 1.15.1, 1.16.0
>Reporter: Huang Xingbo
>Assignee: Huang Xingbo
>Priority: Major
>
> The overall style and structure of the current python api doc is not so 
> convenient to new users to learn and use, we can use numpydoc or other 
> document styles to optimize.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29966) Replace and redesign the Python api documentation base

2022-11-09 Thread Huang Xingbo (Jira)
Huang Xingbo created FLINK-29966:


 Summary: Replace and redesign the Python api documentation base
 Key: FLINK-29966
 URL: https://issues.apache.org/jira/browse/FLINK-29966
 Project: Flink
  Issue Type: Improvement
  Components: API / Python, Documentation
Affects Versions: 1.15.2, 1.16.0
Reporter: Huang Xingbo
Assignee: Huang Xingbo


The doc of the existing python api is difficult to read and use. I have a demo 
site for redesigning the Python api documentation base. See 
https://pyflink-api-docs-test.readthedocs.io/en/latest/ as an example.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-28786) Cannot run PyFlink 1.16 on MacOS with M1 chip

2022-11-07 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28786?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-28786:
-
Fix Version/s: 1.17.0
   1.16.1

> Cannot run PyFlink 1.16 on MacOS with M1 chip
> -
>
> Key: FLINK-28786
> URL: https://issues.apache.org/jira/browse/FLINK-28786
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
>Reporter: Ran Tao
>Assignee: Huang Xingbo
>Priority: Major
> Fix For: 1.17.0, 1.16.1
>
>
> I have tested it with 2 m1 machines. i will reproduce the bug 100%.
> 1.m1 machine
> macos bigsur 11.5.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> 1.m1 machine
> macos monterey 12.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> reproduce step:
> 1.python -m pip install -r flink-python/dev/dev-requirements.txt
> 2.cd flink-python; python setup.py sdist bdist_wheel; cd 
> apache-flink-libraries; python setup.py sdist; cd ..;
> 3.python -m pip install apache-flink-libraries/dist/*.tar.gz
> 4.python -m pip install dist/*.whl
> when run 
> [word_count.py|https://nightlies.apache.org/flink/flink-docs-master/docs/dev/python/table_api_tutorial/]
>  it will cause
> {code:java}
> :219: RuntimeWarning: 
> apache_beam.coders.coder_impl.StreamCoderImpl size changed, may indicate 
> binary incompatibility. Expected 24 from C header, got 32 from PyObject
> Traceback (most recent call last):
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 129, in 
> word_count(known_args.input, known_args.output)
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 49, in word_count
> t_env = TableEnvironment.create(EnvironmentSettings.in_streaming_mode())
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 121, in create
> return TableEnvironment(j_tenv)
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 100, in __init__
> self._open()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1637, in _open
> startup_loopback_server()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1628, in startup_loopback_server
> from pyflink.fn_execution.beam.beam_worker_pool_service import \
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_worker_pool_service.py",
>  line 44, in 
> from pyflink.fn_execution.beam import beam_sdk_worker_main  # noqa # 
> pylint: disable=unused-import
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_sdk_worker_main.py",
>  line 21, in 
> import pyflink.fn_execution.beam.beam_operations # noqa # pylint: 
> disable=unused-import
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_operations.py",
>  line 27, in 
> from pyflink.fn_execution.state_impl import RemoteKeyedStateBackend, 
> RemoteOperatorStateBackend
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/state_impl.py",
>  line 33, in 
> from pyflink.fn_execution.beam.beam_coders import FlinkCoder
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_coders.py",
>  line 27, in 
> from pyflink.fn_execution.beam import beam_coder_impl_fast as 
> beam_coder_impl
>   File "pyflink/fn_execution/beam/beam_coder_impl_fast.pyx", line 1, in init 
> pyflink.fn_execution.beam.beam_coder_impl_fast
> KeyError: '__pyx_vtable__'
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Reopened] (FLINK-28786) Cannot run PyFlink 1.16 on MacOS with M1 chip

2022-11-07 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28786?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo reopened FLINK-28786:
--
  Assignee: Huang Xingbo

> Cannot run PyFlink 1.16 on MacOS with M1 chip
> -
>
> Key: FLINK-28786
> URL: https://issues.apache.org/jira/browse/FLINK-28786
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
>Reporter: Ran Tao
>Assignee: Huang Xingbo
>Priority: Major
>
> I have tested it with 2 m1 machines. i will reproduce the bug 100%.
> 1.m1 machine
> macos bigsur 11.5.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> 1.m1 machine
> macos monterey 12.1 & jdk8 * & jdk11 & python 3.8 & python 3.9
> reproduce step:
> 1.python -m pip install -r flink-python/dev/dev-requirements.txt
> 2.cd flink-python; python setup.py sdist bdist_wheel; cd 
> apache-flink-libraries; python setup.py sdist; cd ..;
> 3.python -m pip install apache-flink-libraries/dist/*.tar.gz
> 4.python -m pip install dist/*.whl
> when run 
> [word_count.py|https://nightlies.apache.org/flink/flink-docs-master/docs/dev/python/table_api_tutorial/]
>  it will cause
> {code:java}
> :219: RuntimeWarning: 
> apache_beam.coders.coder_impl.StreamCoderImpl size changed, may indicate 
> binary incompatibility. Expected 24 from C header, got 32 from PyObject
> Traceback (most recent call last):
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 129, in 
> word_count(known_args.input, known_args.output)
>   File "/Users/chucheng/GitLab/pyflink-demo/table/streaming/word_count.py", 
> line 49, in word_count
> t_env = TableEnvironment.create(EnvironmentSettings.in_streaming_mode())
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 121, in create
> return TableEnvironment(j_tenv)
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 100, in __init__
> self._open()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1637, in _open
> startup_loopback_server()
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/table/table_environment.py",
>  line 1628, in startup_loopback_server
> from pyflink.fn_execution.beam.beam_worker_pool_service import \
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_worker_pool_service.py",
>  line 44, in 
> from pyflink.fn_execution.beam import beam_sdk_worker_main  # noqa # 
> pylint: disable=unused-import
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_sdk_worker_main.py",
>  line 21, in 
> import pyflink.fn_execution.beam.beam_operations # noqa # pylint: 
> disable=unused-import
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_operations.py",
>  line 27, in 
> from pyflink.fn_execution.state_impl import RemoteKeyedStateBackend, 
> RemoteOperatorStateBackend
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/state_impl.py",
>  line 33, in 
> from pyflink.fn_execution.beam.beam_coders import FlinkCoder
>   File 
> "/Users/chucheng/venv/lib/python3.8/site-packages/pyflink/fn_execution/beam/beam_coders.py",
>  line 27, in 
> from pyflink.fn_execution.beam import beam_coder_impl_fast as 
> beam_coder_impl
>   File "pyflink/fn_execution/beam/beam_coder_impl_fast.pyx", line 1, in init 
> pyflink.fn_execution.beam.beam_coder_impl_fast
> KeyError: '__pyx_vtable__'
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-26974) Python EmbeddedThreadDependencyTests.test_add_python_file failed on azure

2022-11-02 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-26974?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo reassigned FLINK-26974:


Assignee: Huang Xingbo

> Python EmbeddedThreadDependencyTests.test_add_python_file failed on azure
> -
>
> Key: FLINK-26974
> URL: https://issues.apache.org/jira/browse/FLINK-26974
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.15.0, 1.16.0, 1.17.0
>Reporter: Yun Gao
>Assignee: Huang Xingbo
>Priority: Minor
>  Labels: auto-deprioritized-major, test-stability
>
> {code:java}
> Mar 31 10:49:17 === FAILURES 
> ===
> Mar 31 10:49:17 __ 
> EmbeddedThreadDependencyTests.test_add_python_file __
> Mar 31 10:49:17 
> Mar 31 10:49:17 self = 
>  testMethod=test_add_python_file>
> Mar 31 10:49:17 
> Mar 31 10:49:17 def test_add_python_file(self):
> Mar 31 10:49:17 python_file_dir = os.path.join(self.tempdir, 
> "python_file_dir_" + str(uuid.uuid4()))
> Mar 31 10:49:17 os.mkdir(python_file_dir)
> Mar 31 10:49:17 python_file_path = os.path.join(python_file_dir, 
> "test_dependency_manage_lib.py")
> Mar 31 10:49:17 with open(python_file_path, 'w') as f:
> Mar 31 10:49:17 f.write("def add_two(a):\nraise 
> Exception('This function should not be called!')")
> Mar 31 10:49:17 self.t_env.add_python_file(python_file_path)
> Mar 31 10:49:17 
> Mar 31 10:49:17 python_file_dir_with_higher_priority = os.path.join(
> Mar 31 10:49:17 self.tempdir, "python_file_dir_" + 
> str(uuid.uuid4()))
> Mar 31 10:49:17 os.mkdir(python_file_dir_with_higher_priority)
> Mar 31 10:49:17 python_file_path_higher_priority = 
> os.path.join(python_file_dir_with_higher_priority,
> Mar 31 10:49:17 
> "test_dependency_manage_lib.py")
> Mar 31 10:49:17 with open(python_file_path_higher_priority, 'w') as f:
> Mar 31 10:49:17 f.write("def add_two(a):\nreturn a + 2")
> Mar 31 10:49:17 
> self.t_env.add_python_file(python_file_path_higher_priority)
> Mar 31 10:49:17 
> Mar 31 10:49:17 def plus_two(i):
> Mar 31 10:49:17 from test_dependency_manage_lib import add_two
> Mar 31 10:49:17 return add_two(i)
> Mar 31 10:49:17 
> Mar 31 10:49:17 self.t_env.create_temporary_system_function(
> Mar 31 10:49:17 "add_two", udf(plus_two, DataTypes.BIGINT(), 
> DataTypes.BIGINT()))
> Mar 31 10:49:17 table_sink = source_sink_utils.TestAppendSink(
> Mar 31 10:49:17 ['a', 'b'], [DataTypes.BIGINT(), 
> DataTypes.BIGINT()])
> Mar 31 10:49:17 self.t_env.register_table_sink("Results", table_sink)
> Mar 31 10:49:17 t = self.t_env.from_elements([(1, 2), (2, 5), (3, 
> 1)], ['a', 'b'])
> Mar 31 10:49:17 >   t.select(expr.call("add_two", t.a), 
> t.a).execute_insert("Results").wait()
> Mar 31 10:49:17 
> Mar 31 10:49:17 pyflink/table/tests/test_dependency.py:63: 
> Mar 31 10:49:17 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ 
> Mar 31 10:49:17 pyflink/table/table_result.py:76: in wait
> Mar 31 10:49:17 get_method(self._j_table_result, "await")()
> Mar 31 10:49:17 
> .tox/py38-cython/lib/python3.8/site-packages/py4j/java_gateway.py:1321: in 
> __call__
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=34001=logs=821b528f-1eed-5598-a3b4-7f748b13f261=6bb545dd-772d-5d8c-f258-f5085fba3295=27239



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-29461) ProcessDataStreamStreamingTests.test_process_function unstable

2022-11-01 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29461?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo reassigned FLINK-29461:


Assignee: Huang Xingbo

> ProcessDataStreamStreamingTests.test_process_function unstable
> --
>
> Key: FLINK-29461
> URL: https://issues.apache.org/jira/browse/FLINK-29461
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
>Reporter: Huang Xingbo
>Assignee: Huang Xingbo
>Priority: Critical
>  Labels: test-stability
>
> {code:java}
> 2022-09-29T02:10:45.3571648Z Sep 29 02:10:45 self = 
>  testMethod=test_process_function>
> 2022-09-29T02:10:45.3572279Z Sep 29 02:10:45 
> 2022-09-29T02:10:45.3572810Z Sep 29 02:10:45 def 
> test_process_function(self):
> 2022-09-29T02:10:45.3573495Z Sep 29 02:10:45 
> self.env.set_parallelism(1)
> 2022-09-29T02:10:45.3574148Z Sep 29 02:10:45 
> self.env.get_config().set_auto_watermark_interval(2000)
> 2022-09-29T02:10:45.3580634Z Sep 29 02:10:45 
> self.env.set_stream_time_characteristic(TimeCharacteristic.EventTime)
> 2022-09-29T02:10:45.3583194Z Sep 29 02:10:45 data_stream = 
> self.env.from_collection([(1, '1603708211000'),
> 2022-09-29T02:10:45.3584515Z Sep 29 02:10:45  
>(2, '1603708224000'),
> 2022-09-29T02:10:45.3585957Z Sep 29 02:10:45  
>(3, '1603708226000'),
> 2022-09-29T02:10:45.3587132Z Sep 29 02:10:45  
>(4, '1603708289000')],
> 2022-09-29T02:10:45.3588094Z Sep 29 02:10:45  
>   type_info=Types.ROW([Types.INT(), Types.STRING()]))
> 2022-09-29T02:10:45.3589090Z Sep 29 02:10:45 
> 2022-09-29T02:10:45.3589949Z Sep 29 02:10:45 class 
> MyProcessFunction(ProcessFunction):
> 2022-09-29T02:10:45.3590710Z Sep 29 02:10:45 
> 2022-09-29T02:10:45.3591856Z Sep 29 02:10:45 def 
> process_element(self, value, ctx):
> 2022-09-29T02:10:45.3592873Z Sep 29 02:10:45 
> current_timestamp = ctx.timestamp()
> 2022-09-29T02:10:45.3593862Z Sep 29 02:10:45 
> current_watermark = ctx.timer_service().current_watermark()
> 2022-09-29T02:10:45.3594915Z Sep 29 02:10:45 yield "current 
> timestamp: {}, current watermark: {}, current_value: {}"\
> 2022-09-29T02:10:45.3596201Z Sep 29 02:10:45 
> .format(str(current_timestamp), str(current_watermark), str(value))
> 2022-09-29T02:10:45.3597089Z Sep 29 02:10:45 
> 2022-09-29T02:10:45.3597942Z Sep 29 02:10:45 watermark_strategy = 
> WatermarkStrategy.for_monotonous_timestamps()\
> 2022-09-29T02:10:45.3599260Z Sep 29 02:10:45 
> .with_timestamp_assigner(SecondColumnTimestampAssigner())
> 2022-09-29T02:10:45.3600611Z Sep 29 02:10:45 
> data_stream.assign_timestamps_and_watermarks(watermark_strategy)\
> 2022-09-29T02:10:45.3601877Z Sep 29 02:10:45 
> .process(MyProcessFunction(), 
> output_type=Types.STRING()).add_sink(self.test_sink)
> 2022-09-29T02:10:45.3603527Z Sep 29 02:10:45 self.env.execute('test 
> process function')
> 2022-09-29T02:10:45.3604445Z Sep 29 02:10:45 results = 
> self.test_sink.get_results()
> 2022-09-29T02:10:45.3605684Z Sep 29 02:10:45 expected = ["current 
> timestamp: 1603708211000, current watermark: "
> 2022-09-29T02:10:45.3607157Z Sep 29 02:10:45 
> "-9223372036854775808, current_value: Row(f0=1, f1='1603708211000')",
> 2022-09-29T02:10:45.3608256Z Sep 29 02:10:45 "current 
> timestamp: 1603708224000, current watermark: "
> 2022-09-29T02:10:45.3609650Z Sep 29 02:10:45 
> "-9223372036854775808, current_value: Row(f0=2, f1='1603708224000')",
> 2022-09-29T02:10:45.3610854Z Sep 29 02:10:45 "current 
> timestamp: 1603708226000, current watermark: "
> 2022-09-29T02:10:45.3612279Z Sep 29 02:10:45 
> "-9223372036854775808, current_value: Row(f0=3, f1='1603708226000')",
> 2022-09-29T02:10:45.3613382Z Sep 29 02:10:45 "current 
> timestamp: 1603708289000, current watermark: "
> 2022-09-29T02:10:45.3615683Z Sep 29 02:10:45 
> "-9223372036854775808, current_value: Row(f0=4, f1='1603708289000')"]
> 2022-09-29T02:10:45.3617687Z Sep 29 02:10:45 >   
> self.assert_equals_sorted(expected, results)
> 2022-09-29T02:10:45.3618620Z Sep 29 02:10:45 
> 2022-09-29T02:10:45.3619425Z Sep 29 02:10:45 
> pyflink/datastream/tests/test_data_stream.py:986: 
> 2022-09-29T02:10:45.3620424Z Sep 29 02:10:45 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> 2022-09-29T02:10:45.3621886Z Sep 

[jira] [Assigned] (FLINK-29817) Published metadata for apache-flink in pypi are inconsistent and causes poetry to fail

2022-11-01 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29817?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo reassigned FLINK-29817:


Assignee: Huang Xingbo

> Published metadata for apache-flink in pypi are inconsistent and causes 
> poetry to fail
> --
>
> Key: FLINK-29817
> URL: https://issues.apache.org/jira/browse/FLINK-29817
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
> Environment: macos 12.6 (M1)
> Poetry Version: 1.2.2
> Python Version:  3.9.12
>Reporter: André Casimiro
>Assignee: Huang Xingbo
>Priority: Critical
>
> Hi, 
> Following the debug steps described in [this github 
> thread|https://github.com/python-poetry/poetry/issues/3011] I got to the 
> conclusion that the metadata of the apache-flink 1.16.0 package is wrong. And 
> because of that I cannot properly manage my dependencies using poetry.
> I can successfully install it with pip (runs with no errors), as stated in 
> the docs:
> {code:java}
> python -m pip install apache-flink {code}
> But when I try to include the dependency in my poetry project I got the 
> following error:
>  
> {code:java}
> ❯ poetry add apache-flink@1.16.0Updating dependencies
> Resolving dependencies... (2.0s)Because pemja (0.2.6) depends on numpy 
> (1.21.4)
>  and apache-flink (1.16.0) depends on numpy (>=1.14.3,<1.20), pemja (0.2.6) 
> is incompatible with apache-flink (1.16.0).
> So, because cv-features depends on apache-flink (1.16.0) which depends on 
> pemja (0.2.6), version solving failed. {code}
>  
> I've followed the same debug steps as in [this github 
> thread|https://github.com/python-poetry/poetry/issues/3011] and can confirm 
> that apache-flink has exactly the same problem as described in the thread: 
> the wheel package in pypi has correct dependency metadata but the pypi 
> published don't.
>  
> {code:java}
> ❯ pkginfo -f requires_dist 
> /Users/andre/Downloads/apache_flink-1.16.0-cp39-cp39-macosx_11_0_arm64.whl
> requires_dist: ['py4j (==0.10.9.3)', 'python-dateutil (==2.8.0)', 
> 'apache-beam (==2.38.0)', 'cloudpickle (==2.1.0)', 'avro-python3 
> (!=1.9.2,<1.10.0,>=1.8.1)', 'pytz (>=2018.3)', 'fastavro (<1.4.8,>=1.1.0)', 
> 'requests (>=2.26.0)', 'protobuf (<3.18)', 'httplib2 (<=0.20.4,>=0.19.0)', 
> 'apache-flink-libraries (<1.16.1,>=1.16.0)', 'numpy (<1.22.0,>=1.21.4)', 
> 'pandas (<1.4.0,>=1.3.0)', 'pyarrow (<9.0.0,>=5.0.0)', 'pemja (==0.2.6) ; 
> python_full_version >= "3.7" and platform_system != "Windows"'] {code}
> but the pipy json metadata is wrong:
>  
>  
> {code:java}
> ❯ curl -sL https://pypi.org/pypi/apache-flink/json | jq '.info.requires_dist'[
>   "py4j (==0.10.9.3)",
>   "python-dateutil (==2.8.0)",
>   "apache-beam (==2.38.0)",
>   "cloudpickle (==2.1.0)",
>   "avro-python3 (!=1.9.2,<1.10.0,>=1.8.1)",
>   "pytz (>=2018.3)",
>   "fastavro (<1.4.8,>=1.1.0)",
>   "requests (>=2.26.0)",
>   "protobuf (<3.18)",
>   "httplib2 (<=0.20.4,>=0.19.0)",
>   "apache-flink-libraries (<1.16.1,>=1.16.0)",
>   "numpy (<1.20,>=1.14.3)",
>   "pandas (<1.2.0,>=1.0)",
>   "pyarrow (<7.0.0,>=0.15.1)",
>   "pemja (==0.2.6) ; python_full_version >= \"3.7\" and platform_system != 
> \"Windows\""
> ]{code}
>  
> As per [this 
> comment|https://github.com/python-poetry/poetry/issues/3011#issuecomment-702826616],
>  could you please republish the package correcting this metadata information, 
> please? This [other 
> comment|https://github.com/apple/turicreate/issues/3342#issuecomment-702957550]
>  can help gain more context.
>  
> Thanks
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29817) Published metadata for apache-flink in pypi are inconsistent and causes poetry to fail

2022-11-01 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29817?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29817:
-
Fix Version/s: 1.17.0
   1.16.1

> Published metadata for apache-flink in pypi are inconsistent and causes 
> poetry to fail
> --
>
> Key: FLINK-29817
> URL: https://issues.apache.org/jira/browse/FLINK-29817
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
> Environment: macos 12.6 (M1)
> Poetry Version: 1.2.2
> Python Version:  3.9.12
>Reporter: André Casimiro
>Assignee: Huang Xingbo
>Priority: Critical
> Fix For: 1.17.0, 1.16.1
>
>
> Hi, 
> Following the debug steps described in [this github 
> thread|https://github.com/python-poetry/poetry/issues/3011] I got to the 
> conclusion that the metadata of the apache-flink 1.16.0 package is wrong. And 
> because of that I cannot properly manage my dependencies using poetry.
> I can successfully install it with pip (runs with no errors), as stated in 
> the docs:
> {code:java}
> python -m pip install apache-flink {code}
> But when I try to include the dependency in my poetry project I got the 
> following error:
>  
> {code:java}
> ❯ poetry add apache-flink@1.16.0Updating dependencies
> Resolving dependencies... (2.0s)Because pemja (0.2.6) depends on numpy 
> (1.21.4)
>  and apache-flink (1.16.0) depends on numpy (>=1.14.3,<1.20), pemja (0.2.6) 
> is incompatible with apache-flink (1.16.0).
> So, because cv-features depends on apache-flink (1.16.0) which depends on 
> pemja (0.2.6), version solving failed. {code}
>  
> I've followed the same debug steps as in [this github 
> thread|https://github.com/python-poetry/poetry/issues/3011] and can confirm 
> that apache-flink has exactly the same problem as described in the thread: 
> the wheel package in pypi has correct dependency metadata but the pypi 
> published don't.
>  
> {code:java}
> ❯ pkginfo -f requires_dist 
> /Users/andre/Downloads/apache_flink-1.16.0-cp39-cp39-macosx_11_0_arm64.whl
> requires_dist: ['py4j (==0.10.9.3)', 'python-dateutil (==2.8.0)', 
> 'apache-beam (==2.38.0)', 'cloudpickle (==2.1.0)', 'avro-python3 
> (!=1.9.2,<1.10.0,>=1.8.1)', 'pytz (>=2018.3)', 'fastavro (<1.4.8,>=1.1.0)', 
> 'requests (>=2.26.0)', 'protobuf (<3.18)', 'httplib2 (<=0.20.4,>=0.19.0)', 
> 'apache-flink-libraries (<1.16.1,>=1.16.0)', 'numpy (<1.22.0,>=1.21.4)', 
> 'pandas (<1.4.0,>=1.3.0)', 'pyarrow (<9.0.0,>=5.0.0)', 'pemja (==0.2.6) ; 
> python_full_version >= "3.7" and platform_system != "Windows"'] {code}
> but the pipy json metadata is wrong:
>  
>  
> {code:java}
> ❯ curl -sL https://pypi.org/pypi/apache-flink/json | jq '.info.requires_dist'[
>   "py4j (==0.10.9.3)",
>   "python-dateutil (==2.8.0)",
>   "apache-beam (==2.38.0)",
>   "cloudpickle (==2.1.0)",
>   "avro-python3 (!=1.9.2,<1.10.0,>=1.8.1)",
>   "pytz (>=2018.3)",
>   "fastavro (<1.4.8,>=1.1.0)",
>   "requests (>=2.26.0)",
>   "protobuf (<3.18)",
>   "httplib2 (<=0.20.4,>=0.19.0)",
>   "apache-flink-libraries (<1.16.1,>=1.16.0)",
>   "numpy (<1.20,>=1.14.3)",
>   "pandas (<1.2.0,>=1.0)",
>   "pyarrow (<7.0.0,>=0.15.1)",
>   "pemja (==0.2.6) ; python_full_version >= \"3.7\" and platform_system != 
> \"Windows\""
> ]{code}
>  
> As per [this 
> comment|https://github.com/python-poetry/poetry/issues/3011#issuecomment-702826616],
>  could you please republish the package correcting this metadata information, 
> please? This [other 
> comment|https://github.com/apple/turicreate/issues/3342#issuecomment-702957550]
>  can help gain more context.
>  
> Thanks
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29817) Published metadata for apache-flink in pypi are inconsistent and causes poetry to fail

2022-11-01 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29817?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17627413#comment-17627413
 ] 

Huang Xingbo commented on FLINK-29817:
--

I think we can optimize the `metadata` to solve this problem in the 1.16.1.

> Published metadata for apache-flink in pypi are inconsistent and causes 
> poetry to fail
> --
>
> Key: FLINK-29817
> URL: https://issues.apache.org/jira/browse/FLINK-29817
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
> Environment: macos 12.6 (M1)
> Poetry Version: 1.2.2
> Python Version:  3.9.12
>Reporter: André Casimiro
>Priority: Critical
>
> Hi, 
> Following the debug steps described in [this github 
> thread|https://github.com/python-poetry/poetry/issues/3011] I got to the 
> conclusion that the metadata of the apache-flink 1.16.0 package is wrong. And 
> because of that I cannot properly manage my dependencies using poetry.
> I can successfully install it with pip (runs with no errors), as stated in 
> the docs:
> {code:java}
> python -m pip install apache-flink {code}
> But when I try to include the dependency in my poetry project I got the 
> following error:
>  
> {code:java}
> ❯ poetry add apache-flink@1.16.0Updating dependencies
> Resolving dependencies... (2.0s)Because pemja (0.2.6) depends on numpy 
> (1.21.4)
>  and apache-flink (1.16.0) depends on numpy (>=1.14.3,<1.20), pemja (0.2.6) 
> is incompatible with apache-flink (1.16.0).
> So, because cv-features depends on apache-flink (1.16.0) which depends on 
> pemja (0.2.6), version solving failed. {code}
>  
> I've followed the same debug steps as in [this github 
> thread|https://github.com/python-poetry/poetry/issues/3011] and can confirm 
> that apache-flink has exactly the same problem as described in the thread: 
> the wheel package in pypi has correct dependency metadata but the pypi 
> published don't.
>  
> {code:java}
> ❯ pkginfo -f requires_dist 
> /Users/andre/Downloads/apache_flink-1.16.0-cp39-cp39-macosx_11_0_arm64.whl
> requires_dist: ['py4j (==0.10.9.3)', 'python-dateutil (==2.8.0)', 
> 'apache-beam (==2.38.0)', 'cloudpickle (==2.1.0)', 'avro-python3 
> (!=1.9.2,<1.10.0,>=1.8.1)', 'pytz (>=2018.3)', 'fastavro (<1.4.8,>=1.1.0)', 
> 'requests (>=2.26.0)', 'protobuf (<3.18)', 'httplib2 (<=0.20.4,>=0.19.0)', 
> 'apache-flink-libraries (<1.16.1,>=1.16.0)', 'numpy (<1.22.0,>=1.21.4)', 
> 'pandas (<1.4.0,>=1.3.0)', 'pyarrow (<9.0.0,>=5.0.0)', 'pemja (==0.2.6) ; 
> python_full_version >= "3.7" and platform_system != "Windows"'] {code}
> but the pipy json metadata is wrong:
>  
>  
> {code:java}
> ❯ curl -sL https://pypi.org/pypi/apache-flink/json | jq '.info.requires_dist'[
>   "py4j (==0.10.9.3)",
>   "python-dateutil (==2.8.0)",
>   "apache-beam (==2.38.0)",
>   "cloudpickle (==2.1.0)",
>   "avro-python3 (!=1.9.2,<1.10.0,>=1.8.1)",
>   "pytz (>=2018.3)",
>   "fastavro (<1.4.8,>=1.1.0)",
>   "requests (>=2.26.0)",
>   "protobuf (<3.18)",
>   "httplib2 (<=0.20.4,>=0.19.0)",
>   "apache-flink-libraries (<1.16.1,>=1.16.0)",
>   "numpy (<1.20,>=1.14.3)",
>   "pandas (<1.2.0,>=1.0)",
>   "pyarrow (<7.0.0,>=0.15.1)",
>   "pemja (==0.2.6) ; python_full_version >= \"3.7\" and platform_system != 
> \"Windows\""
> ]{code}
>  
> As per [this 
> comment|https://github.com/python-poetry/poetry/issues/3011#issuecomment-702826616],
>  could you please republish the package correcting this metadata information, 
> please? This [other 
> comment|https://github.com/apple/turicreate/issues/3342#issuecomment-702957550]
>  can help gain more context.
>  
> Thanks
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29833) Improve PyFlink support in Windows

2022-11-01 Thread Huang Xingbo (Jira)
Huang Xingbo created FLINK-29833:


 Summary: Improve PyFlink support in Windows
 Key: FLINK-29833
 URL: https://issues.apache.org/jira/browse/FLINK-29833
 Project: Flink
  Issue Type: Improvement
  Components: API / Python
Affects Versions: 1.15.2, 1.16.0
Reporter: Huang Xingbo
Assignee: Huang Xingbo


Many users are used to developing PyFlink jobs on Windows. It is necessary to 
improve the simplicity of PyFlink job development on Windows



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29567) Revert sink output metric names from numRecordsSend back to numRecordsOut

2022-10-31 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29567?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17626927#comment-17626927
 ] 

Huang Xingbo commented on FLINK-29567:
--

 Hi [~renqs], the fixed PR hasn't been backported to release-1.15?


> Revert sink output metric names from numRecordsSend back to numRecordsOut
> -
>
> Key: FLINK-29567
> URL: https://issues.apache.org/jira/browse/FLINK-29567
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Common
>Affects Versions: 1.16.0, 1.15.3
>Reporter: Qingsheng Ren
>Assignee: Qingsheng Ren
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.16.0, 1.15.3
>
>
> As discussed in [the mailing 
> list|https://lists.apache.org/thread/vxhty3q97s7pw2zn0jhkyd6sxwwodzbv], 
> all{color:#33} sink metrics with name “numXXXOut” defined in FLIP-33 are 
> replace by “numXXXSend” in FLINK-26126 and FLINK-26492. {color}
> {color:#33}Considering metric names are public APIs, this is a breaking 
> change to end users and not backward compatible. We need to revert these 
> metric names back. {color}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29817) Published metadata for apache-flink in pypi are inconsistent and causes poetry to fail

2022-10-31 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29817?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17626911#comment-17626911
 ] 

Huang Xingbo commented on FLINK-29817:
--

The `install_requirements` of PyFlink depends on your `python_version`.   
{code:python}
if sys.version_info < (3, 7):
# python 3.6 upper and lower limit
install_requires.append('numpy>=1.14.3,<1.20')
install_requires.append('pandas>=1.0,<1.2.0')
install_requires.append('pyarrow>=0.15.1,<7.0.0')
else:
# python 3.7, 3.8 and 3.9 upper limit and M1 chip lower limit,
install_requires.append('numpy>=1.21.4,<1.22.0')
install_requires.append('pandas>=1.3.0,<1.4.0')
install_requires.append('pyarrow>=5.0.0,<9.0.0')
{code}
 So I guess Pypi collects the `metadata` used in Python3.6.
 


> Published metadata for apache-flink in pypi are inconsistent and causes 
> poetry to fail
> --
>
> Key: FLINK-29817
> URL: https://issues.apache.org/jira/browse/FLINK-29817
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0
> Environment: macos 12.6 (M1)
> Poetry Version: 1.2.2
> Python Version:  3.9.12
>Reporter: André Casimiro
>Priority: Critical
>
> Hi, 
> Following the debug steps described in [this github 
> thread|https://github.com/python-poetry/poetry/issues/3011] I got to the 
> conclusion that the metadata of the apache-flink 1.16.0 package is wrong. And 
> because of that I cannot properly manage my dependencies using poetry.
> I can successfully install it with pip (runs with no errors), as stated in 
> the docs:
> {code:java}
> python -m pip install apache-flink {code}
> But when I try to include the dependency in my poetry project I got the 
> following error:
>  
> {code:java}
> ❯ poetry add apache-flink@1.16.0Updating dependencies
> Resolving dependencies... (2.0s)Because pemja (0.2.6) depends on numpy 
> (1.21.4)
>  and apache-flink (1.16.0) depends on numpy (>=1.14.3,<1.20), pemja (0.2.6) 
> is incompatible with apache-flink (1.16.0).
> So, because cv-features depends on apache-flink (1.16.0) which depends on 
> pemja (0.2.6), version solving failed. {code}
>  
> I've followed the same debug steps as in [this github 
> thread|https://github.com/python-poetry/poetry/issues/3011] and can confirm 
> that apache-flink has exactly the same problem as described in the thread: 
> the wheel package in pypi has correct dependency metadata but the pypi 
> published don't.
>  
> {code:java}
> ❯ pkginfo -f requires_dist 
> /Users/andre/Downloads/apache_flink-1.16.0-cp39-cp39-macosx_11_0_arm64.whl
> requires_dist: ['py4j (==0.10.9.3)', 'python-dateutil (==2.8.0)', 
> 'apache-beam (==2.38.0)', 'cloudpickle (==2.1.0)', 'avro-python3 
> (!=1.9.2,<1.10.0,>=1.8.1)', 'pytz (>=2018.3)', 'fastavro (<1.4.8,>=1.1.0)', 
> 'requests (>=2.26.0)', 'protobuf (<3.18)', 'httplib2 (<=0.20.4,>=0.19.0)', 
> 'apache-flink-libraries (<1.16.1,>=1.16.0)', 'numpy (<1.22.0,>=1.21.4)', 
> 'pandas (<1.4.0,>=1.3.0)', 'pyarrow (<9.0.0,>=5.0.0)', 'pemja (==0.2.6) ; 
> python_full_version >= "3.7" and platform_system != "Windows"'] {code}
> but the pipy json metadata is wrong:
>  
>  
> {code:java}
> ❯ curl -sL https://pypi.org/pypi/apache-flink/json | jq '.info.requires_dist'[
>   "py4j (==0.10.9.3)",
>   "python-dateutil (==2.8.0)",
>   "apache-beam (==2.38.0)",
>   "cloudpickle (==2.1.0)",
>   "avro-python3 (!=1.9.2,<1.10.0,>=1.8.1)",
>   "pytz (>=2018.3)",
>   "fastavro (<1.4.8,>=1.1.0)",
>   "requests (>=2.26.0)",
>   "protobuf (<3.18)",
>   "httplib2 (<=0.20.4,>=0.19.0)",
>   "apache-flink-libraries (<1.16.1,>=1.16.0)",
>   "numpy (<1.20,>=1.14.3)",
>   "pandas (<1.2.0,>=1.0)",
>   "pyarrow (<7.0.0,>=0.15.1)",
>   "pemja (==0.2.6) ; python_full_version >= \"3.7\" and platform_system != 
> \"Windows\""
> ]{code}
>  
> As per [this 
> comment|https://github.com/python-poetry/poetry/issues/3011#issuecomment-702826616],
>  could you please republish the package correcting this metadata information, 
> please? This [other 
> comment|https://github.com/apple/turicreate/issues/3342#issuecomment-702957550]
>  can help gain more context.
>  
> Thanks
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-29733) Error Flink connector hive Test failing

2022-10-31 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29733?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-29733.

Fix Version/s: 1.17.0
 Assignee: Samrat Deb
   Resolution: Fixed

Merged into master via 759d954960cbec28437955e5f482788902342035

> Error Flink connector hive Test failing
> ---
>
> Key: FLINK-29733
> URL: https://issues.apache.org/jira/browse/FLINK-29733
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.17.0
>Reporter: Samrat Deb
>Assignee: Samrat Deb
>Priority: Major
>  Labels: pull-request-available, test-stability
> Fix For: 1.17.0
>
>
> [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=42328=logs=245e1f2e-ba5b-5570-d689-25ae21e5302f=d04c9862-880c-52f5-574b-a7a79fef8e0f]
> This is caused by FLINK-29478
> reported by [~hxbks2ks] 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-29788) StatefulJobWBroadcastStateMigrationITCase failed in native savepoints

2022-10-31 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29788?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-29788.

Fix Version/s: 1.17.0
   1.16.1
 Assignee: Hangxiang Yu
   Resolution: Fixed

Merged into master via c89e400ae379c8b7490d9af20f82f49319895dce
Merged into release-1.16 via 87d86c05c7a91d7546935ec1ceee3c0c55e2e191

> StatefulJobWBroadcastStateMigrationITCase failed in native savepoints
> -
>
> Key: FLINK-29788
> URL: https://issues.apache.org/jira/browse/FLINK-29788
> Project: Flink
>  Issue Type: Bug
>  Components: Release System, Runtime / State Backends
>Affects Versions: 1.16.0
>Reporter: Huang Xingbo
>Assignee: Hangxiang Yu
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.17.0, 1.16.1
>
> Attachments: image-2022-10-28-11-18-45-471.png
>
>
>  !image-2022-10-28-11-18-45-471.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29788) StatefulJobWBroadcastStateMigrationITCase failed in native savepoints

2022-10-31 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29788?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17626434#comment-17626434
 ] 

Huang Xingbo commented on FLINK-29788:
--

[~masteryhx] Thanks a lot for the fix. Before the release, the code may change, 
so I think preparing the migrated test data in advance may result in some 
problems.


> StatefulJobWBroadcastStateMigrationITCase failed in native savepoints
> -
>
> Key: FLINK-29788
> URL: https://issues.apache.org/jira/browse/FLINK-29788
> Project: Flink
>  Issue Type: Bug
>  Components: Release System, Runtime / State Backends
>Affects Versions: 1.16.0
>Reporter: Huang Xingbo
>Priority: Critical
>  Labels: pull-request-available
> Attachments: image-2022-10-28-11-18-45-471.png
>
>
>  !image-2022-10-28-11-18-45-471.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29788) StatefulJobWBroadcastStateMigrationITCase failed in native savepoints

2022-10-28 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29788?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625491#comment-17625491
 ] 

Huang Xingbo commented on FLINK-29788:
--

It is not the blocker of release 1.16.0,but it belongs to the finalized step of 
a release.

> StatefulJobWBroadcastStateMigrationITCase failed in native savepoints
> -
>
> Key: FLINK-29788
> URL: https://issues.apache.org/jira/browse/FLINK-29788
> Project: Flink
>  Issue Type: Bug
>  Components: Release System, Runtime / State Backends
>Affects Versions: 1.16.0
>Reporter: Huang Xingbo
>Priority: Critical
> Attachments: image-2022-10-28-11-18-45-471.png
>
>
>  !image-2022-10-28-11-18-45-471.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29788) StatefulJobWBroadcastStateMigrationITCase failed in native savepoints

2022-10-27 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29788?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29788:
-
Priority: Critical  (was: Major)

> StatefulJobWBroadcastStateMigrationITCase failed in native savepoints
> -
>
> Key: FLINK-29788
> URL: https://issues.apache.org/jira/browse/FLINK-29788
> Project: Flink
>  Issue Type: Bug
>  Components: Release System, Runtime / State Backends
>Affects Versions: 1.16.0
>Reporter: Huang Xingbo
>Priority: Critical
> Attachments: image-2022-10-28-11-18-45-471.png
>
>
>  !image-2022-10-28-11-18-45-471.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29788) StatefulJobWBroadcastStateMigrationITCase failed in native savepoints

2022-10-27 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29788?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17625404#comment-17625404
 ] 

Huang Xingbo commented on FLINK-29788:
--

[~masteryhx] Could you help take a look? Thx.

> StatefulJobWBroadcastStateMigrationITCase failed in native savepoints
> -
>
> Key: FLINK-29788
> URL: https://issues.apache.org/jira/browse/FLINK-29788
> Project: Flink
>  Issue Type: Bug
>  Components: Release System, Runtime / State Backends
>Affects Versions: 1.16.0
>Reporter: Huang Xingbo
>Priority: Major
> Attachments: image-2022-10-28-11-18-45-471.png
>
>
>  !image-2022-10-28-11-18-45-471.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29788) StatefulJobWBroadcastStateMigrationITCase failed in native savepoints

2022-10-27 Thread Huang Xingbo (Jira)
Huang Xingbo created FLINK-29788:


 Summary: StatefulJobWBroadcastStateMigrationITCase failed in 
native savepoints
 Key: FLINK-29788
 URL: https://issues.apache.org/jira/browse/FLINK-29788
 Project: Flink
  Issue Type: Bug
  Components: Release System, Runtime / State Backends
Affects Versions: 1.16.0
Reporter: Huang Xingbo
 Attachments: image-2022-10-28-11-18-45-471.png

 !image-2022-10-28-11-18-45-471.png! 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29427) LookupJoinITCase failed with classloader problem

2022-09-30 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29427?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17611629#comment-17611629
 ] 

Huang Xingbo commented on FLINK-29427:
--

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=41488=logs=0c940707-2659-5648-cbe6-a1ad63045f0a=075c2716-8010-5565-fe08-3c4bb45824a4=21930

> LookupJoinITCase failed with classloader problem
> 
>
> Key: FLINK-29427
> URL: https://issues.apache.org/jira/browse/FLINK-29427
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.16.0
>Reporter: Huang Xingbo
>Priority: Critical
>  Labels: test-stability
>
> {code:java}
> 2022-09-27T02:49:20.9501313Z Sep 27 02:49:20 Caused by: 
> org.codehaus.janino.InternalCompilerException: Compiling 
> "KeyProjection$108341": Trying to access closed classloader. Please check if 
> you store classloaders directly or indirectly in static fields. If the 
> stacktrace suggests that the leak occurs in a third party library and cannot 
> be fixed immediately, you can disable this check with the configuration 
> 'classloader.check-leaked-classloader'.
> 2022-09-27T02:49:20.9502654Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler.compileUnit(UnitCompiler.java:382)
> 2022-09-27T02:49:20.9503366Z Sep 27 02:49:20  at 
> org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:237)
> 2022-09-27T02:49:20.9504044Z Sep 27 02:49:20  at 
> org.codehaus.janino.SimpleCompiler.compileToClassLoader(SimpleCompiler.java:465)
> 2022-09-27T02:49:20.9504704Z Sep 27 02:49:20  at 
> org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:216)
> 2022-09-27T02:49:20.9505341Z Sep 27 02:49:20  at 
> org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:207)
> 2022-09-27T02:49:20.9505965Z Sep 27 02:49:20  at 
> org.codehaus.commons.compiler.Cookable.cook(Cookable.java:80)
> 2022-09-27T02:49:20.9506584Z Sep 27 02:49:20  at 
> org.codehaus.commons.compiler.Cookable.cook(Cookable.java:75)
> 2022-09-27T02:49:20.9507261Z Sep 27 02:49:20  at 
> org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:104)
> 2022-09-27T02:49:20.9507883Z Sep 27 02:49:20  ... 30 more
> 2022-09-27T02:49:20.9509266Z Sep 27 02:49:20 Caused by: 
> java.lang.IllegalStateException: Trying to access closed classloader. Please 
> check if you store classloaders directly or indirectly in static fields. If 
> the stacktrace suggests that the leak occurs in a third party library and 
> cannot be fixed immediately, you can disable this check with the 
> configuration 'classloader.check-leaked-classloader'.
> 2022-09-27T02:49:20.9510835Z Sep 27 02:49:20  at 
> org.apache.flink.util.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.ensureInner(FlinkUserCodeClassLoaders.java:184)
> 2022-09-27T02:49:20.9511760Z Sep 27 02:49:20  at 
> org.apache.flink.util.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.loadClass(FlinkUserCodeClassLoaders.java:192)
> 2022-09-27T02:49:20.9512456Z Sep 27 02:49:20  at 
> java.lang.Class.forName0(Native Method)
> 2022-09-27T02:49:20.9513014Z Sep 27 02:49:20  at 
> java.lang.Class.forName(Class.java:348)
> 2022-09-27T02:49:20.9513649Z Sep 27 02:49:20  at 
> org.codehaus.janino.ClassLoaderIClassLoader.findIClass(ClassLoaderIClassLoader.java:89)
> 2022-09-27T02:49:20.9514339Z Sep 27 02:49:20  at 
> org.codehaus.janino.IClassLoader.loadIClass(IClassLoader.java:312)
> 2022-09-27T02:49:20.9514990Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler.findTypeByName(UnitCompiler.java:8556)
> 2022-09-27T02:49:20.9515659Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler.getReferenceType(UnitCompiler.java:6749)
> 2022-09-27T02:49:20.9516337Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler.getReferenceType(UnitCompiler.java:6594)
> 2022-09-27T02:49:20.9516989Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler.getType2(UnitCompiler.java:6573)
> 2022-09-27T02:49:20.9517632Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler.access$13900(UnitCompiler.java:215)
> 2022-09-27T02:49:20.9518319Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler$22$1.visitReferenceType(UnitCompiler.java:6481)
> 2022-09-27T02:49:20.9519018Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler$22$1.visitReferenceType(UnitCompiler.java:6476)
> 2022-09-27T02:49:20.9519680Z Sep 27 02:49:20  at 
> org.codehaus.janino.Java$ReferenceType.accept(Java.java:3928)
> 2022-09-27T02:49:20.9520386Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler$22.visitType(UnitCompiler.java:6476)
> 2022-09-27T02:49:20.9521042Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler$22.visitType(UnitCompiler.java:6469)
> 2022-09-27T02:49:20.9521677Z Sep 27 02:49:20  at 
> org.codehaus.janino.Java$ReferenceType.accept(Java.java:3927)
> 

[jira] [Commented] (FLINK-29427) LookupJoinITCase failed with classloader problem

2022-09-30 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29427?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17611628#comment-17611628
 ] 

Huang Xingbo commented on FLINK-29427:
--

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=41493=logs=0c940707-2659-5648-cbe6-a1ad63045f0a=075c2716-8010-5565-fe08-3c4bb45824a4=21545

> LookupJoinITCase failed with classloader problem
> 
>
> Key: FLINK-29427
> URL: https://issues.apache.org/jira/browse/FLINK-29427
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.16.0
>Reporter: Huang Xingbo
>Priority: Critical
>  Labels: test-stability
>
> {code:java}
> 2022-09-27T02:49:20.9501313Z Sep 27 02:49:20 Caused by: 
> org.codehaus.janino.InternalCompilerException: Compiling 
> "KeyProjection$108341": Trying to access closed classloader. Please check if 
> you store classloaders directly or indirectly in static fields. If the 
> stacktrace suggests that the leak occurs in a third party library and cannot 
> be fixed immediately, you can disable this check with the configuration 
> 'classloader.check-leaked-classloader'.
> 2022-09-27T02:49:20.9502654Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler.compileUnit(UnitCompiler.java:382)
> 2022-09-27T02:49:20.9503366Z Sep 27 02:49:20  at 
> org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:237)
> 2022-09-27T02:49:20.9504044Z Sep 27 02:49:20  at 
> org.codehaus.janino.SimpleCompiler.compileToClassLoader(SimpleCompiler.java:465)
> 2022-09-27T02:49:20.9504704Z Sep 27 02:49:20  at 
> org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:216)
> 2022-09-27T02:49:20.9505341Z Sep 27 02:49:20  at 
> org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:207)
> 2022-09-27T02:49:20.9505965Z Sep 27 02:49:20  at 
> org.codehaus.commons.compiler.Cookable.cook(Cookable.java:80)
> 2022-09-27T02:49:20.9506584Z Sep 27 02:49:20  at 
> org.codehaus.commons.compiler.Cookable.cook(Cookable.java:75)
> 2022-09-27T02:49:20.9507261Z Sep 27 02:49:20  at 
> org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:104)
> 2022-09-27T02:49:20.9507883Z Sep 27 02:49:20  ... 30 more
> 2022-09-27T02:49:20.9509266Z Sep 27 02:49:20 Caused by: 
> java.lang.IllegalStateException: Trying to access closed classloader. Please 
> check if you store classloaders directly or indirectly in static fields. If 
> the stacktrace suggests that the leak occurs in a third party library and 
> cannot be fixed immediately, you can disable this check with the 
> configuration 'classloader.check-leaked-classloader'.
> 2022-09-27T02:49:20.9510835Z Sep 27 02:49:20  at 
> org.apache.flink.util.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.ensureInner(FlinkUserCodeClassLoaders.java:184)
> 2022-09-27T02:49:20.9511760Z Sep 27 02:49:20  at 
> org.apache.flink.util.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.loadClass(FlinkUserCodeClassLoaders.java:192)
> 2022-09-27T02:49:20.9512456Z Sep 27 02:49:20  at 
> java.lang.Class.forName0(Native Method)
> 2022-09-27T02:49:20.9513014Z Sep 27 02:49:20  at 
> java.lang.Class.forName(Class.java:348)
> 2022-09-27T02:49:20.9513649Z Sep 27 02:49:20  at 
> org.codehaus.janino.ClassLoaderIClassLoader.findIClass(ClassLoaderIClassLoader.java:89)
> 2022-09-27T02:49:20.9514339Z Sep 27 02:49:20  at 
> org.codehaus.janino.IClassLoader.loadIClass(IClassLoader.java:312)
> 2022-09-27T02:49:20.9514990Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler.findTypeByName(UnitCompiler.java:8556)
> 2022-09-27T02:49:20.9515659Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler.getReferenceType(UnitCompiler.java:6749)
> 2022-09-27T02:49:20.9516337Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler.getReferenceType(UnitCompiler.java:6594)
> 2022-09-27T02:49:20.9516989Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler.getType2(UnitCompiler.java:6573)
> 2022-09-27T02:49:20.9517632Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler.access$13900(UnitCompiler.java:215)
> 2022-09-27T02:49:20.9518319Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler$22$1.visitReferenceType(UnitCompiler.java:6481)
> 2022-09-27T02:49:20.9519018Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler$22$1.visitReferenceType(UnitCompiler.java:6476)
> 2022-09-27T02:49:20.9519680Z Sep 27 02:49:20  at 
> org.codehaus.janino.Java$ReferenceType.accept(Java.java:3928)
> 2022-09-27T02:49:20.9520386Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler$22.visitType(UnitCompiler.java:6476)
> 2022-09-27T02:49:20.9521042Z Sep 27 02:49:20  at 
> org.codehaus.janino.UnitCompiler$22.visitType(UnitCompiler.java:6469)
> 2022-09-27T02:49:20.9521677Z Sep 27 02:49:20  at 
> org.codehaus.janino.Java$ReferenceType.accept(Java.java:3927)
> 

[jira] [Commented] (FLINK-29419) HybridShuffle.testHybridFullExchangesRestart hangs

2022-09-30 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29419?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17611627#comment-17611627
 ] 

Huang Xingbo commented on FLINK-29419:
--

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=41494=logs=5c8e7682-d68f-54d1-16a2-a09310218a49=86f654fa-ab48-5c1a-25f4-7e7f6afb9bba=11782

> HybridShuffle.testHybridFullExchangesRestart hangs
> --
>
> Key: FLINK-29419
> URL: https://issues.apache.org/jira/browse/FLINK-29419
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.16.0, 1.17.0
>Reporter: Huang Xingbo
>Assignee: Weijie Guo
>Priority: Critical
>  Labels: pull-request-available, test-stability
>
> {code:java}
> 2022-09-26T10:56:44.0766792Z Sep 26 10:56:44 "ForkJoinPool-1-worker-25" #27 
> daemon prio=5 os_prio=0 tid=0x7f41a4efa000 nid=0x6d76 waiting on 
> condition [0x7f40ac135000]
> 2022-09-26T10:56:44.0767432Z Sep 26 10:56:44java.lang.Thread.State: 
> WAITING (parking)
> 2022-09-26T10:56:44.0767892Z Sep 26 10:56:44  at sun.misc.Unsafe.park(Native 
> Method)
> 2022-09-26T10:56:44.0768644Z Sep 26 10:56:44  - parking to wait for  
> <0xa0704e18> (a java.util.concurrent.CompletableFuture$Signaller)
> 2022-09-26T10:56:44.0769287Z Sep 26 10:56:44  at 
> java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> 2022-09-26T10:56:44.0769949Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
> 2022-09-26T10:56:44.0770623Z Sep 26 10:56:44  at 
> java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3313)
> 2022-09-26T10:56:44.0771349Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
> 2022-09-26T10:56:44.0772092Z Sep 26 10:56:44  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
> 2022-09-26T10:56:44.0772777Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.JobGraphRunningUtil.execute(JobGraphRunningUtil.java:57)
> 2022-09-26T10:56:44.0773534Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.BatchShuffleITCaseBase.executeJob(BatchShuffleITCaseBase.java:115)
> 2022-09-26T10:56:44.0774333Z Sep 26 10:56:44  at 
> org.apache.flink.test.runtime.HybridShuffleITCase.testHybridFullExchangesRestart(HybridShuffleITCase.java:59)
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=41343=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29483) flink python udf arrow in thread model bug

2022-09-30 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17611617#comment-17611617
 ] 

Huang Xingbo commented on FLINK-29483:
--

[~jackylau] You means your job runs in `thread mode` will raise this exception? 

> flink python udf arrow in thread model bug
> --
>
> Key: FLINK-29483
> URL: https://issues.apache.org/jira/browse/FLINK-29483
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0, 1.15.2
>Reporter: jackylau
>Priority: Major
> Fix For: 1.17.0, 1.15.3, 1.16.1
>
> Attachments: image-2022-09-30-17-03-05-005.png
>
>
> !image-2022-09-30-17-03-05-005.png!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-29479) support whether using system env for pythonpath

2022-09-30 Thread Huang Xingbo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-29479?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17611469#comment-17611469
 ] 

Huang Xingbo commented on FLINK-29479:
--

[~jackylau] From my side, the users dependency will have a higher priority, so 
it shouldn't be affected by the PYTHONPATH of the system env  under normal 
circumstances. Could you give a more specific example of how the system's 
PYTHONPATH is having a bad effect in some circumstance.

> support whether using system env for pythonpath
> ---
>
> Key: FLINK-29479
> URL: https://issues.apache.org/jira/browse/FLINK-29479
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.17.0
>Reporter: jackylau
>Priority: Major
> Fix For: 1.17.0
>
>
> It exists PYTHONPATH env in system,like yarn/k8s images, it will cause 
> conflict with users python depdendency sometimes. so i suggest add a config 
> to do whether using system env of PYTHONPATH



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29350) Add a section for moving planner jar in Hive dependencies page

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29350?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29350:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Add a section for moving planner jar in Hive dependencies page
> --
>
> Key: FLINK-29350
> URL: https://issues.apache.org/jira/browse/FLINK-29350
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hive, Documentation
>Affects Versions: 1.16.0
>Reporter: luoyuxia
>Assignee: luoyuxia
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-28071) Support missing built-in functions in Table API

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28071?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-28071:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Support missing built-in functions in Table API
> ---
>
> Key: FLINK-28071
> URL: https://issues.apache.org/jira/browse/FLINK-28071
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Reporter: Dian Fu
>Assignee: Luning Wang
>Priority: Major
> Fix For: 1.17.0
>
>
> There are many built-in functions are not supported. See 
> https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/functions/systemfunctions/
>  for more details. There are two columns for each built-in function: *SQL 
> Function* and *Table Function*, if a function is not supported in *Table 
> API*, the *Table Function* column is documented as *N/A*. We need to evaluate 
> each of these functions to ensure that they could be used in both SQL and 
> Table API.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-27467) Remove CliFrontendTestBase

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-27467:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Remove CliFrontendTestBase
> --
>
> Key: FLINK-27467
> URL: https://issues.apache.org/jira/browse/FLINK-27467
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Command Line Client, Tests
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Fix For: 1.17.0
>
>
> The class isn't really providing any value.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29372) Add a suffix to keys that violate YAML spec

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29372?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29372:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Add a suffix to keys that violate YAML spec
> ---
>
> Key: FLINK-29372
> URL: https://issues.apache.org/jira/browse/FLINK-29372
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Configuration
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>
> We have a few options where the key is a prefix of other options (e.g., 
> {{high-availability}} and {{high-availability.cluster-id}}.
> Add a suffix to these options and keep the old key as deprecated.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29430) Sanity check in InternalKeyContextImpl#setCurrentKeyGroupIndex

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29430?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29430:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Sanity check in InternalKeyContextImpl#setCurrentKeyGroupIndex
> --
>
> Key: FLINK-29430
> URL: https://issues.apache.org/jira/browse/FLINK-29430
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / State Backends
>Affects Versions: 1.15.2
>Reporter: Zakelly Lan
>Assignee: Zakelly Lan
>Priority: Major
> Fix For: 1.17.0
>
>
> Currently the HeapStateBackend check whether the current key group index is a 
> valid one while the RocksDBStateBackend will not. When using 
> HeapStateBackend, if the user uses a non-deterministic shuffle key, an 
> exception is thrown as follows:
>  
> {code:java}
> java.lang.IllegalArgumentException: Key group 84 is not in 
> KeyGroupRange{startKeyGroup=32, endKeyGroup=63}. Unless you're directly using 
> low level state access APIs, this is most likely caused by non-deterministic 
> shuffle key (hashCode and equals implementation).
>     at 
> org.apache.flink.runtime.state.KeyGroupRangeOffsets.newIllegalKeyGroupException(KeyGroupRangeOffsets.java:37)
>     at 
> org.apache.flink.runtime.state.heap.StateTable.getMapForKeyGroup(StateTable.java:305)
>     at org.apache.flink.runtime.state.heap.StateTable.get(StateTable.java:261)
>     at org.apache.flink.runtime.state.heap.StateTable.get(StateTable.java:143)
>     at 
> org.apache.flink.runtime.state.heap.HeapValueState.value(HeapValueState.java:72)
>     at 
> com.alibaba.ververica.flink.state.benchmark.wordcount.WordCount$MixedFlatMapper.flatMap(WordCount.java:169)
>     at 
> com.alibaba.ververica.flink.state.benchmark.wordcount.WordCount$MixedFlatMapper.flatMap(WordCount.java:160)
>     at 
> org.apache.flink.streaming.api.operators.StreamFlatMap.processElement(StreamFlatMap.java:47)
>     at 
> org.apache.flink.streaming.runtime.tasks.OneInputStreamTask$StreamTaskNetworkOutput.emitRecord(OneInputStreamTask.java:233)
>     at 
> org.apache.flink.streaming.runtime.io.AbstractStreamTaskNetworkInput.processElement(AbstractStreamTaskNetworkInput.java:135)
>     at 
> org.apache.flink.streaming.runtime.io.AbstractStreamTaskNetworkInput.emitNext(AbstractStreamTaskNetworkInput.java:106)
>     at 
> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:65)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:526)
>     at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:203)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:811)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:760)
>     at 
> org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:954)
>     at 
> org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:933)
>     at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:746)
>     at org.apache.flink.runtime.taskmanager.Task.run(Task.java:568)
>     at java.lang.Thread.run(Thread.java:750)
>  {code}
> However, the RocksDBStateBackend will run without an exception. The wrong key 
> group index will cause a state correctness problem, so it is better to do a 
> check in {_}InternalKeyContextImpl#{_}{_}setCurrentKeyGroupIndex{_}, and 
> throw an exception immediately.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-27432) Replace Time with Duration in TaskSlotTable

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27432?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-27432:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Replace Time with Duration in TaskSlotTable
> ---
>
> Key: FLINK-27432
> URL: https://issues.apache.org/jira/browse/FLINK-27432
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Coordination
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
> Fix For: 1.17.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-27075) Migrate CI from Azure to Github Actions

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27075?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-27075:
-
Fix Version/s: 1.17.0
   1.16.1
   (was: 1.16.0)
   (was: 1.14.7)

> Migrate CI from Azure to Github Actions
> ---
>
> Key: FLINK-27075
> URL: https://issues.apache.org/jira/browse/FLINK-27075
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Build System / CI
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
> Fix For: 1.17.0, 1.15.3, 1.16.1
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-26360) [Umbrella] Improvement for Hive Query Syntax Compatibility

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-26360?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-26360:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> [Umbrella] Improvement for Hive Query Syntax Compatibility
> --
>
> Key: FLINK-26360
> URL: https://issues.apache.org/jira/browse/FLINK-26360
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Hive
>Reporter: luoyuxia
>Assignee: luoyuxia
>Priority: Major
> Fix For: 1.17.0
>
>
> Currently, we have a support for hive synatax compatibility in flink as 
> described in [FLINK-21529|https://issues.apache.org/jira/browse/FLINK-21529].
> But there're still some features we don't support or some other issues when 
> using hive synatax.
> In here, we want to make a improvement to solve the issues encountered when 
> using Hive dialect to make it be more smoothly when you mrigate your hive job 
> to flink or enable you write flink job using hive synatax with less knowledge 
> about flink sql.
> Feel free to leave your comment.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29402) Add USE_DIRECT_READ configuration parameter for RocksDB

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29402:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Add USE_DIRECT_READ configuration parameter for RocksDB
> ---
>
> Key: FLINK-29402
> URL: https://issues.apache.org/jira/browse/FLINK-29402
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.16.0
>Reporter: Donatien
>Priority: Not a Priority
>  Labels: Enhancement, pull-request-available, rocksdb
> Fix For: 1.17.0
>
> Attachments: directIO-performance-comparison.png
>
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> RocksDB allows the use of DirectIO for read operations to bypass the Linux 
> Page Cache. To understand the impact of Linux Page Cache on performance, one 
> can run a heavy workload on a single-tasked Task Manager with a container 
> memory limit identical to the TM process memory. Running this same workload 
> on a TM with no container memory limit will result in better performances but 
> with the host memory exceeding the TM requirement.
> Linux Page Cache are of course useful but can give false results when 
> benchmarking the Managed Memory used by RocksDB. DirectIO is typically 
> enabled for benchmarks on working set estimation [Zwaenepoel et 
> al.|[https://arxiv.org/abs/1702.04323].]
> I propose to add a configuration key allowing users to enable the use of 
> DirectIO for reads thanks to the RocksDB API. This configuration would be 
> disabled by default.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29157) Clarify the contract between CompletedCheckpointStore and SharedStateRegistry

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29157?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29157:
-
Fix Version/s: 1.17.0
   1.16.1
   (was: 1.16.0)

> Clarify the contract between CompletedCheckpointStore and SharedStateRegistry
> -
>
> Key: FLINK-29157
> URL: https://issues.apache.org/jira/browse/FLINK-29157
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Runtime / Checkpointing
>Affects Versions: 1.16.0, 1.15.2
>Reporter: Roman Khachatryan
>Assignee: Roman Khachatryan
>Priority: Minor
> Fix For: 1.17.0, 1.15.3, 1.16.1
>
>
> After FLINK-24611, CompletedCheckpointStore is required to call 
> SharedStateRegistry.unregisterUnusedState() on checkpoint subsumption and 
> shutdown.
> Although it's not clear whether CompletedCheckpointStore is internal there 
> are in fact external implementations (which weren't updated accordingly).
>  
> After FLINK-25872, CompletedCheckpointStore also must call 
> checkpointsCleaner.cleanSubsumedCheckpoints.
>  
> Another issue with a custom implementation was using different java objects 
> for state for CheckpointStore and SharedStateRegistry (after FLINK-24086). 
>  
> So it makes sense to:
>  * clarify the contract (different in 1.15 and 1.16)
>  * require using the same checkpoint objects by SharedStateRegistryFactory 
> and CompletedCheckpointStore
>  * mark the interface(s) as PublicEvolving



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-27077) Setup per-commit builds

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27077?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-27077:
-
Fix Version/s: 1.17.0
   1.16.1
   (was: 1.16.0)
   (was: 1.14.7)

> Setup per-commit builds
> ---
>
> Key: FLINK-27077
> URL: https://issues.apache.org/jira/browse/FLINK-27077
> Project: Flink
>  Issue Type: Sub-task
>  Components: Build System / CI
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
> Fix For: 1.17.0, 1.15.3, 1.16.1
>
>
> For initial validation, setup the per-commit builds for pushes to the master 
> branch.
> Once validation is complete, backport to 1.14/1.15.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29298) LocalBufferPool request buffer from NetworkBufferPool hanging

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29298?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29298:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> LocalBufferPool request buffer from NetworkBufferPool hanging
> -
>
> Key: FLINK-29298
> URL: https://issues.apache.org/jira/browse/FLINK-29298
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.16.0
>Reporter: Weijie Guo
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.17.0
>
> Attachments: image-2022-09-14-10-52-15-259.png, 
> image-2022-09-14-10-58-45-987.png, image-2022-09-14-11-00-47-309.png
>
>
> In the scenario where the buffer contention is fierce, sometimes the task 
> hang can be observed. Through the thread dump information, we can found that 
> the task thread is blocked by requestMemorySegmentBlocking forever. After 
> investigating the dumped heap information, I found that the NetworkBufferPool 
> actually has many buffers, but the LocalBufferPool is still unavailable and 
> no buffer has been obtained.
> By looking at the code, I am sure that this is a bug in thread race: when the 
> task thread polled out the last buffer in LocalBufferPool and triggered the 
> onGlobalPoolAvailable callback itself, it will skip this notification  (as 
> currently the LocalBufferPool is available), which will cause the BufferPool 
> to eventually become unavailable and will never register a callback to the 
> NetworkBufferPool.
> The conditions for triggering the problem are relatively strict, but I have 
> found a stable way to reproduce it, I will try to fix and verify this problem.
> !image-2022-09-14-10-52-15-259.png|width=1021,height=219!
> !image-2022-09-14-10-58-45-987.png|width=997,height=315!
> !image-2022-09-14-11-00-47-309.png|width=453,height=121!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-27753) Scala suffix checks should check for maven errors

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27753?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-27753:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Scala suffix checks should check for maven errors
> -
>
> Key: FLINK-27753
> URL: https://issues.apache.org/jira/browse/FLINK-27753
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Build System, Build System / CI
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
> Fix For: 1.17.0
>
>
> As shown in FLINK-27751 errors in maven when retrieving the dependency tree 
> are currently ignored, causing garbage output being passed to the suffix 
> checker.
> This results in highly misleading error messages about incorrect suffixes.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-27455) [JUnit5 Migration] SnapshotMigrationTestBase

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27455?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-27455:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> [JUnit5 Migration] SnapshotMigrationTestBase
> 
>
> Key: FLINK-27455
> URL: https://issues.apache.org/jira/browse/FLINK-27455
> Project: Flink
>  Issue Type: Sub-task
>  Components: Tests
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
> Fix For: 1.17.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-27299) flink parsing parameter bug

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27299?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-27299:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> flink parsing parameter bug
> ---
>
> Key: FLINK-27299
> URL: https://issues.apache.org/jira/browse/FLINK-27299
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Configuration
>Affects Versions: 1.11.6, 1.13.5, 1.14.2, 1.13.6, 1.14.3, 1.14.4
>Reporter: Huajie Wang
>Assignee: Huajie Wang
>Priority: Minor
>  Labels: easyfix, pull-request-available, stale-assigned
> Fix For: 1.17.0
>
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> When I am running a flink job, I specify a running parameter with a "#" sign 
> in it. The parsing fails.
> e.g: flink run com.myJob --sink.password db@123#123 
> only parse the content in front of "#", after reading the source code It is 
> found that the parameters are intercepted according to "#" in the 
> loadYAMLResource method of GlobalConfiguration. This part needs to be improved



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-28960) Pulsar throws java.lang.NoClassDefFoundError: javax/xml/bind/annotation/XmlElement

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28960?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-28960:
-
Fix Version/s: 1.17.0
   1.16.1
   (was: 1.16.0)

> Pulsar throws java.lang.NoClassDefFoundError: 
> javax/xml/bind/annotation/XmlElement
> --
>
> Key: FLINK-28960
> URL: https://issues.apache.org/jira/browse/FLINK-28960
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Pulsar
>Affects Versions: 1.15.1, 1.14.6
>Reporter: Yufan Sheng
>Priority: Minor
> Fix For: 1.17.0, 1.15.3, 1.14.7, 1.16.1
>
>
> {code:java}
> Unknown HK2 failure detected:
> MultiException stack 1 of 2
> java.lang.NoClassDefFoundError: javax/xml/bind/annotation/XmlElement
>   at 
> org.apache.pulsar.shade.com.fasterxml.jackson.module.jaxb.JaxbAnnotationIntrospector.(JaxbAnnotationIntrospector.java:137)
>   at 
> org.apache.pulsar.shade.com.fasterxml.jackson.module.jaxb.JaxbAnnotationIntrospector.(JaxbAnnotationIntrospector.java:124)
>   at 
> org.apache.pulsar.shade.com.fasterxml.jackson.module.jaxb.JaxbAnnotationIntrospector.(JaxbAnnotationIntrospector.java:116)
>   at 
> java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>  Method)
>   at 
> java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>   at 
> java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>   at 
> java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)
>   at java.base/java.lang.Class.newInstance(Class.java:584)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-25612) Update the outdated illustration of ExecutionState in the documentation

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-25612?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-25612:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Update the outdated illustration of ExecutionState in the documentation
> ---
>
> Key: FLINK-25612
> URL: https://issues.apache.org/jira/browse/FLINK-25612
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Documentation
>Affects Versions: 1.13.5, 1.14.2, 1.15.0
>Reporter: Zhilong Hong
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.17.0
>
> Attachments: current-illustration-2.jpg, new-illustration-2.jpg
>
>
> Currently, the illustration of {{ExecutionState}} located in the page "Jobs 
> and Scheduling" 
> ([https://nightlies.apache.org/flink/flink-docs-master/docs/internals/job_scheduling/])
>  is outdated. It doesn't involve the INITIALIZING state, which is introduced 
> in FLINK-17102.
>  
> Current illustration:
> !current-illustration-2.jpg|width=400!
> New illustration:
>  !new-illustration-2.jpg|width=400!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-24943) SequenceNumber class is not POJO type

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-24943?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-24943:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> SequenceNumber class is not POJO type
> -
>
> Key: FLINK-24943
> URL: https://issues.apache.org/jira/browse/FLINK-24943
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kinesis
>Affects Versions: 1.12.5, 1.13.3, 1.14.4, 1.15.0
>Reporter: Alexander Egorov
>Assignee: Alexander Egorov
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>
> SequenceNumber class is currently part of the "Kinesis-Stream-Shard-State", 
> but it does not follow requirements of POJO compatible type. Because of that 
> we are getting warning like this:
> {{TypeExtractor - class 
> software.amazon.kinesis.connectors.flink.model.SequenceNumber does not 
> contain a setter for field sequenceNumber}}
> While the warning itself or inability to use optimal sterilizer for such a 
> small state is not the problem, this warning prevents us to disable Generic 
> Types via {{disableGenericTypes()}}
> So, the problem is similar to 
> https://issues.apache.org/jira/browse/FLINK-15904



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-28553) The serializer in StateMap has not been updated when metaInfo of StateTable updated

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28553?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-28553:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> The serializer in StateMap has not been updated when metaInfo of StateTable 
> updated
> ---
>
> Key: FLINK-28553
> URL: https://issues.apache.org/jira/browse/FLINK-28553
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / State Backends
>Affects Versions: 1.14.5, 1.15.1
>Reporter: Hangxiang Yu
>Assignee: Hangxiang Yu
>Priority: Minor
> Fix For: 1.17.0
>
>
> When the meta info in StateTable updated, the serializer in StateMap has not 
> been updated.
> (See StateTable#setMetaInfo)
> The value may be serialized/deserialized/copied incorrectly after triggering 
> state migration of HashMapStateBackend.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-25842) [v2] FLIP-158: Generalized incremental checkpoints

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-25842?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-25842:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> [v2] FLIP-158: Generalized incremental checkpoints
> --
>
> Key: FLINK-25842
> URL: https://issues.apache.org/jira/browse/FLINK-25842
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Reporter: Roman Khachatryan
>Assignee: Yuan Mei
>Priority: Major
> Fix For: 1.17.0
>
>
> Umbrella ticket for the 2nd iteration of [FLIP-158: Generalized incremental 
> checkpoints|https://cwiki.apache.org/confluence/display/FLINK/FLIP-158%3A+Generalized+incremental+checkpoints]
>  
> 1st iteration: FLINK-21352



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-28766) UnalignedCheckpointStressITCase.runStressTest failed with NoSuchFileException

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-28766?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-28766:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> UnalignedCheckpointStressITCase.runStressTest failed with NoSuchFileException
> -
>
> Key: FLINK-28766
> URL: https://issues.apache.org/jira/browse/FLINK-28766
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.16.0
>Reporter: Huang Xingbo
>Assignee: Anton Kalashnikov
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.17.0
>
>
> {code:java}
> 2022-08-01T01:36:16.0563880Z Aug 01 01:36:16 [ERROR] 
> org.apache.flink.test.checkpointing.UnalignedCheckpointStressITCase.runStressTest
>   Time elapsed: 12.579 s  <<< ERROR!
> 2022-08-01T01:36:16.0565407Z Aug 01 01:36:16 java.io.UncheckedIOException: 
> java.nio.file.NoSuchFileException: 
> /tmp/junit1058240190382532303/f0f99754a53d2c4633fed75011da58dd/chk-7/61092e4a-5b9a-4f56-83f7-d9960c53ed3e
> 2022-08-01T01:36:16.0566296Z Aug 01 01:36:16  at 
> java.nio.file.FileTreeIterator.fetchNextIfNeeded(FileTreeIterator.java:88)
> 2022-08-01T01:36:16.0566972Z Aug 01 01:36:16  at 
> java.nio.file.FileTreeIterator.hasNext(FileTreeIterator.java:104)
> 2022-08-01T01:36:16.0567600Z Aug 01 01:36:16  at 
> java.util.Iterator.forEachRemaining(Iterator.java:115)
> 2022-08-01T01:36:16.0568290Z Aug 01 01:36:16  at 
> java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
> 2022-08-01T01:36:16.0569172Z Aug 01 01:36:16  at 
> java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:482)
> 2022-08-01T01:36:16.0569877Z Aug 01 01:36:16  at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:472)
> 2022-08-01T01:36:16.0570554Z Aug 01 01:36:16  at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
> 2022-08-01T01:36:16.0571371Z Aug 01 01:36:16  at 
> java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
> 2022-08-01T01:36:16.0572417Z Aug 01 01:36:16  at 
> java.util.stream.ReferencePipeline.reduce(ReferencePipeline.java:546)
> 2022-08-01T01:36:16.0573618Z Aug 01 01:36:16  at 
> org.apache.flink.test.checkpointing.UnalignedCheckpointStressITCase.discoverRetainedCheckpoint(UnalignedCheckpointStressITCase.java:289)
> 2022-08-01T01:36:16.0575187Z Aug 01 01:36:16  at 
> org.apache.flink.test.checkpointing.UnalignedCheckpointStressITCase.runAndTakeExternalCheckpoint(UnalignedCheckpointStressITCase.java:262)
> 2022-08-01T01:36:16.0576540Z Aug 01 01:36:16  at 
> org.apache.flink.test.checkpointing.UnalignedCheckpointStressITCase.runStressTest(UnalignedCheckpointStressITCase.java:158)
> 2022-08-01T01:36:16.0577684Z Aug 01 01:36:16  at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 2022-08-01T01:36:16.0578546Z Aug 01 01:36:16  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 2022-08-01T01:36:16.0579374Z Aug 01 01:36:16  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 2022-08-01T01:36:16.0580298Z Aug 01 01:36:16  at 
> java.lang.reflect.Method.invoke(Method.java:498)
> 2022-08-01T01:36:16.0581243Z Aug 01 01:36:16  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> 2022-08-01T01:36:16.0582029Z Aug 01 01:36:16  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2022-08-01T01:36:16.0582766Z Aug 01 01:36:16  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> 2022-08-01T01:36:16.0583488Z Aug 01 01:36:16  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 2022-08-01T01:36:16.0584203Z Aug 01 01:36:16  at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> 2022-08-01T01:36:16.0585087Z Aug 01 01:36:16  at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> 2022-08-01T01:36:16.0585778Z Aug 01 01:36:16  at 
> org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
> 2022-08-01T01:36:16.0586482Z Aug 01 01:36:16  at 
> org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
> 2022-08-01T01:36:16.0587155Z Aug 01 01:36:16  at 
> org.apache.flink.util.TestNameProvider$1.evaluate(TestNameProvider.java:45)
> 2022-08-01T01:36:16.0587809Z Aug 01 01:36:16  at 
> org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
> 2022-08-01T01:36:16.0588434Z Aug 01 01:36:16  at 
> org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
> 2022-08-01T01:36:16.0589203Z Aug 01 01:36:16  at 
> 

[jira] [Updated] (FLINK-29179) Translate "Temporal Table Function" page of "Streaming Concepts" into Chinese

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29179?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29179:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Translate "Temporal Table Function" page of "Streaming Concepts" into Chinese
> -
>
> Key: FLINK-29179
> URL: https://issues.apache.org/jira/browse/FLINK-29179
> Project: Flink
>  Issue Type: Improvement
>  Components: chinese-translation
>Affects Versions: 1.15.2
>Reporter: ChunJi
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>
> The page url is 
> [https://nightlies.apache.org/flink/flink-docs-release-1.15/docs/dev/table/concepts/temporal_table_function/]
> The markdown file is located in 
> docs/content.zh/docs/dev/table/concepts/temporal_table_function.md



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-25862) Refactor SharedStateRegistry to not limit StreamStateHandle to register/unregister

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-25862?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-25862:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Refactor SharedStateRegistry to not limit StreamStateHandle to 
> register/unregister
> --
>
> Key: FLINK-25862
> URL: https://issues.apache.org/jira/browse/FLINK-25862
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Runtime / Checkpointing, Runtime / State Backends
>Reporter: Yun Tang
>Priority: Minor
> Fix For: 1.17.0
>
>
> Current implementation of SharedStateRegistry would use `StreamStateHandle` 
> to register and unregister. This would limit the usage for other componments, 
> such as change-log state backend handle usage.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-26369) Translate the part zh-page mixed with not be translated.

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-26369?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-26369:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Translate the part zh-page mixed with not be translated. 
> -
>
> Key: FLINK-26369
> URL: https://issues.apache.org/jira/browse/FLINK-26369
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.15.0
>Reporter: Aiden Gong
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>
> These file should be translated.
> Files:
> docs/content.zh/docs/deployment/ha/overview.md
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-24932) Frocksdb cannot run on Apple M1

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-24932?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-24932:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Frocksdb cannot run on Apple M1
> ---
>
> Key: FLINK-24932
> URL: https://issues.apache.org/jira/browse/FLINK-24932
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / State Backends
>Reporter: Yun Tang
>Assignee: Yun Tang
>Priority: Minor
> Fix For: 1.17.0
>
>
> After we bump up RocksDB version to 6.20.3, we support to run RocksDB on 
> linux arm cluster. However, according to the feedback from Robert, Apple M1 
> machines cannot run FRocksDB yet:
> {code:java}
> java.lang.Exception: Exception while creating StreamOperatorStateContext.
>   at 
> org.apache.flink.streaming.api.operators.StreamTaskStateInitializerImpl.streamOperatorStateContext(StreamTaskStateInitializerImpl.java:255)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:268)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.initializeStateAndOpenOperators(RegularOperatorChain.java:109)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.restoreGates(StreamTask.java:711)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$1.call(StreamTaskActionExecutor.java:55)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.restoreInternal(StreamTask.java:687)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:654)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:958)
>  ~[flink-runtime-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:927) 
> ~[flink-runtime-1.14.0.jar:1.14.0]
>   at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:766) 
> ~[flink-runtime-1.14.0.jar:1.14.0]
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:575) 
> ~[flink-runtime-1.14.0.jar:1.14.0]
>   at java.lang.Thread.run(Thread.java:748) ~[?:1.8.0_312]
> Caused by: org.apache.flink.util.FlinkException: Could not restore keyed 
> state backend for StreamFlatMap_c21234bcbf1e8eb4c61f1927190efebd_(1/1) from 
> any of the 1 provided restore options.
>   at 
> org.apache.flink.streaming.api.operators.BackendRestorerProcedure.createAndRestore(BackendRestorerProcedure.java:160)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.streaming.api.operators.StreamTaskStateInitializerImpl.keyedStatedBackend(StreamTaskStateInitializerImpl.java:346)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.streaming.api.operators.StreamTaskStateInitializerImpl.streamOperatorStateContext(StreamTaskStateInitializerImpl.java:164)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   ... 11 more
> Caused by: java.io.IOException: Could not load the native RocksDB library
>   at 
> org.apache.flink.contrib.streaming.state.EmbeddedRocksDBStateBackend.ensureRocksDBIsLoaded(EmbeddedRocksDBStateBackend.java:882)
>  ~[flink-statebackend-rocksdb_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.contrib.streaming.state.EmbeddedRocksDBStateBackend.createKeyedStateBackend(EmbeddedRocksDBStateBackend.java:402)
>  ~[flink-statebackend-rocksdb_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.contrib.streaming.state.RocksDBStateBackend.createKeyedStateBackend(RocksDBStateBackend.java:345)
>  ~[flink-statebackend-rocksdb_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.contrib.streaming.state.RocksDBStateBackend.createKeyedStateBackend(RocksDBStateBackend.java:87)
>  ~[flink-statebackend-rocksdb_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.streaming.api.operators.StreamTaskStateInitializerImpl.lambda$keyedStatedBackend$1(StreamTaskStateInitializerImpl.java:329)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.streaming.api.operators.BackendRestorerProcedure.attemptCreateAndRestore(BackendRestorerProcedure.java:168)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   at 
> org.apache.flink.streaming.api.operators.BackendRestorerProcedure.createAndRestore(BackendRestorerProcedure.java:135)
>  ~[flink-streaming-java_2.11-1.14.0.jar:1.14.0]
>   at 
> 

[jira] [Updated] (FLINK-23035) Add explicit method to StateChangelogWriter to write metadata

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-23035?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-23035:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Add explicit method to StateChangelogWriter to write metadata
> -
>
> Key: FLINK-23035
> URL: https://issues.apache.org/jira/browse/FLINK-23035
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Runtime / State Backends
>Reporter: Roman Khachatryan
>Assignee: Xinbin Huang
>Priority: Minor
> Fix For: 1.17.0
>
>
> Currently, metadata is written to the state changelog using the same 
> StateChangelogWriter.append() method as data.
> However, it doesn't belong to a specific group, and should be read first on 
> recovery. Because of that, -1 is used.
> An explicit append() without keygroup would be less fragile (probably still 
> using -1 under the hood).



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-24688) yarn.application-attempt-failures-validity-interval link url is invalid

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-24688?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-24688:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> yarn.application-attempt-failures-validity-interval link url is invalid
> ---
>
> Key: FLINK-24688
> URL: https://issues.apache.org/jira/browse/FLINK-24688
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN
>Affects Versions: 1.9.0, 1.10.0, 1.11.0, 1.12.0, 1.13.0, 1.14.0
>Reporter: baizhendong
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>
> yarn.application-attempt-failures-validity-interval property was added in 
> YarnConfigOptions since FLINK-12472[yarn], but now the link 
> [https://hortonworks.com/blog/apache-hadoop-yarn-hdp-2-2-fault-tolerance-features-long-running-services/]
>  is not available. So we should update this link.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-24497) Kafka metrics fetching throws IllegalStateException

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-24497?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-24497:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Kafka metrics fetching throws IllegalStateException 
> 
>
> Key: FLINK-24497
> URL: https://issues.apache.org/jira/browse/FLINK-24497
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Affects Versions: 1.14.0, 1.15.0
>Reporter: Juha
>Priority: Minor
> Fix For: 1.17.0
>
>
> I have a simple job that just consumes from a single Kafka topic, performs 
> some filtering and produces to another topic.
> The TaskManager log has these periodically. This is a new problem in 1.14.0, 
> the same setup didn't have the issue when using 1.13.0 or 1.13.2.
>  {code}
> 2021-10-05T15:22:31.928316  [2021-10-05 15:22:31,927] WARN Error when getting 
> Kafka consumer metric "records-lag" for partition "cpu.kafka-1". Metric 
> "pendingBytes" may not be reported correctly.  
> (org.apache.flink.connector.kafka.source.metrics.KafkaSourceReaderMetrics:306)
> 2021-10-05T15:22:31.928316  java.lang.IllegalStateException: Cannot find 
> Kafka metric matching current filter.
> 2021-10-05T15:22:31.928316at 
> org.apache.flink.connector.kafka.MetricUtil.lambda$getKafkaMetric$1(MetricUtil.java:63)
>  ~[flink-connector-kafka_2.12-1.14.0.jar:1.14.0]
> 2021-10-05T15:22:31.928316at 
> java.util.Optional.orElseThrow(Optional.java:408) ~[?:?]
> 2021-10-05T15:22:31.928316at 
> org.apache.flink.connector.kafka.MetricUtil.getKafkaMetric(MetricUtil.java:61)
>  ~[flink-connector-kafka_2.12-1.14.0.jar:1.14.0]
> 2021-10-05T15:22:31.928316at 
> org.apache.flink.connector.kafka.source.metrics.KafkaSourceReaderMetrics.getRecordsLagMetric(KafkaSourceReaderMetrics.java:304)
>  ~[flink-connector-kafka_2.12-1.14.0.jar:1.14.0]
> 2021-10-05T15:22:31.928316at 
> org.apache.flink.connector.kafka.source.metrics.KafkaSourceReaderMetrics.lambda$maybeAddRecordsLagMetric$4(KafkaSourceReaderMetrics.java:229)
>  ~[flink-connector-kafka_2.12-1.14.0.jar:1.14.0]
> 2021-10-05T15:22:31.928316at 
> java.util.concurrent.ConcurrentHashMap.computeIfAbsent(ConcurrentHashMap.java:1705)
>  [?:?]
> 2021-10-05T15:22:31.928316at 
> org.apache.flink.connector.kafka.source.metrics.KafkaSourceReaderMetrics.maybeAddRecordsLagMetric(KafkaSourceReaderMetrics.java:228)
>  [flink-connector-kafka_2.12-1.14.0.jar:1.14.0]
> 2021-10-05T15:22:31.928316at 
> org.apache.flink.connector.kafka.source.reader.KafkaPartitionSplitReader.fetch(KafkaPartitionSplitReader.java:187)
>  [flink-connector-kafka_2.12-1.14.0.jar:1.14.0]
> 2021-10-05T15:22:31.928316at 
> org.apache.flink.connector.base.source.reader.fetcher.FetchTask.run(FetchTask.java:58)
>  [flink-table_2.12-1.14.0.jar:1.14.0]
> 2021-10-05T15:22:31.928316at 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:142)
>  [flink-table_2.12-1.14.0.jar:1.14.0]
> 2021-10-05T15:22:31.928316at 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:105)
>  [flink-table_2.12-1.14.0.jar:1.14.0]
> 2021-10-05T15:22:31.928316at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) [?:?]
> 2021-10-05T15:22:31.928316at 
> java.util.concurrent.FutureTask.run(FutureTask.java:264) [?:?]
> 2021-10-05T15:22:31.928316at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
>  [?:?]
> 2021-10-05T15:22:31.928316at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
>  [?:?]
> 2021-10-05T15:22:31.928316at java.lang.Thread.run(Thread.java:829) [?:?]
> {code}
> Regards,
> Juha



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-22842) Streaming File Sink end-to-end test fail due to not finish after 900s

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22842?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-22842.

Resolution: Cannot Reproduce

> Streaming File Sink end-to-end test fail due to not finish after 900s
> -
>
> Key: FLINK-22842
> URL: https://issues.apache.org/jira/browse/FLINK-22842
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem
>Affects Versions: 1.14.0
>Reporter: Xintong Song
>Priority: Minor
>  Labels: auto-deprioritized-major, test-stability
> Fix For: 1.16.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=18525=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=ff888d9b-cd34-53cc-d90f-3e446d355529=13152



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-22334) Fail to translate the hive-sql in STREAMING mode

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22334?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-22334:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Fail to translate the hive-sql in STREAMING mode
> 
>
> Key: FLINK-22334
> URL: https://issues.apache.org/jira/browse/FLINK-22334
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.13.0
>Reporter: Shengkai Fang
>Priority: Minor
>  Labels: auto-deprioritized-major
> Fix For: 1.17.0
>
>
> Please run in the streaming mode.
> The failed statement 
> {code:java}
> // Some comments here
> insert into dest(y,x) select x,y from foo cluster by x
> {code}
> Exception stack:
> {code:java}
> org.apache.calcite.plan.RelOptPlanner$CannotPlanException: There are not 
> enough rules to produce a node with desired properties: convention=LOGICAL, 
> FlinkRelDistributionTraitDef=any, MiniBatchIntervalTraitDef=None: 0, 
> ModifyKindSetTraitDef=[NONE], UpdateKindTraitDef=[NONE].
> Missing conversion is LogicalDistribution[convention: NONE -> LOGICAL]
> There is 1 empty subset: rel#5176:RelSubset#43.LOGICAL.any.None: 
> 0.[NONE].[NONE], the relevant part of the original plan is as follows
> 5174:LogicalDistribution(collation=[[0 ASC-nulls-first]], dist=[[]])
>   5172:LogicalProject(subset=[rel#5173:RelSubset#42.NONE.any.None: 
> 0.[NONE].[NONE]], x=[$0])
> 5106:LogicalTableScan(subset=[rel#5171:RelSubset#41.NONE.any.None: 
> 0.[NONE].[NONE]], table=[[myhive, default, foo]])
> Root: rel#5176:RelSubset#43.LOGICAL.any.None: 0.[NONE].[NONE]
> Original rel:
> FlinkLogicalLegacySink(subset=[rel#4254:RelSubset#8.LOGICAL.any.None: 
> 0.[NONE].[NONE]], name=[collect], fields=[_o__c0]): rowcount = 1.0E8, 
> cumulative cost = {1.0E8 rows, 1.0E8 cpu, 0.0 io, 0.0 network, 0.0 memory}, 
> id = 4276
>   FlinkLogicalCalc(subset=[rel#4275:RelSubset#7.LOGICAL.any.None: 
> 0.[NONE].[NONE]], select=[CASE(IS NULL($f1), 0:BIGINT, $f1) AS _o__c0]): 
> rowcount = 1.0E8, cumulative cost = {1.0E8 rows, 1.0E8 cpu, 0.0 io, 0.0 
> network, 0.0 memory}, id = 4288
> FlinkLogicalJoin(subset=[rel#4272:RelSubset#6.LOGICAL.any.None: 
> 0.[NONE].[NONE]], condition=[=($0, $1)], joinType=[left]): rowcount = 1.0E8, 
> cumulative cost = {1.0E8 rows, 1.0856463237676364E8 cpu, 4.0856463237676364E8 
> io, 0.0 network, 0.0 memory}, id = 4271
>   
> FlinkLogicalTableSourceScan(subset=[rel#4270:RelSubset#1.LOGICAL.any.None: 
> 0.[NONE].[NONE]], table=[[myhive, default, bar, project=[i]]], fields=[i]): 
> rowcount = 1.0E8, cumulative cost = {1.0E8 rows, 1.0E8 cpu, 4.0E8 io, 0.0 
> network, 0.0 memory}, id = 4279
>   FlinkLogicalAggregate(subset=[rel#4268:RelSubset#5.LOGICAL.any.None: 
> 0.[NONE].[NONE]], group=[{1}], agg#0=[COUNT($0)]): rowcount = 
> 8564632.376763644, cumulative cost = {9.0E7 rows, 1.89E8 cpu, 7.2E8 io, 0.0 
> network, 0.0 memory}, id = 4286
> FlinkLogicalCalc(subset=[rel#4283:RelSubset#3.LOGICAL.any.None: 
> 0.[NONE].[NONE]], select=[x, y], where=[IS NOT NULL(y)]): rowcount = 9.0E7, 
> cumulative cost = {9.0E7 rows, 0.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id 
> = 4282
>   
> FlinkLogicalTableSourceScan(subset=[rel#4262:RelSubset#2.LOGICAL.any.None: 
> 0.[NONE].[NONE]], table=[[myhive, default, foo]], fields=[x, y]): rowcount = 
> 1.0E8, cumulative cost = {1.0E8 rows, 1.0E8 cpu, 8.0E8 io, 0.0 network, 0.0 
> memory}, id = 4261
> Sets:
> Set#41, type: RecordType(INTEGER x, INTEGER y)
>   rel#5171:RelSubset#41.NONE.any.None: 0.[NONE].[NONE], best=null
>   rel#5106:LogicalTableScan.NONE.any.None: 
> 0.[NONE].[NONE](table=[myhive, default, foo]), rowcount=1.0E8, cumulative 
> cost={inf}
>   rel#5179:RelSubset#41.LOGICAL.any.None: 0.[NONE].[NONE], best=rel#5178
>   rel#5178:FlinkLogicalTableSourceScan.LOGICAL.any.None: 
> 0.[NONE].[NONE](table=[myhive, default, foo],fields=x, y), rowcount=1.0E8, 
> cumulative cost={1.0E8 rows, 1.0E8 cpu, 8.0E8 io, 0.0 network, 0.0 memory}
> Set#42, type: RecordType(INTEGER x)
>   rel#5173:RelSubset#42.NONE.any.None: 0.[NONE].[NONE], best=null
>   rel#5172:LogicalProject.NONE.any.None: 
> 0.[NONE].[NONE](input=RelSubset#5171,inputs=0), rowcount=1.0E8, cumulative 
> cost={inf}
>   rel#5180:LogicalTableScan.NONE.any.None: 
> 0.[NONE].[NONE](table=[myhive, default, foo, project=[x]]), rowcount=1.0E8, 
> cumulative cost={inf}
>   rel#5182:LogicalCalc.NONE.any.None: 
> 0.[NONE].[NONE](input=RelSubset#5171,expr#0..1={inputs},0=$t0), 
> rowcount=1.0E8, cumulative cost={inf}
>   rel#5184:RelSubset#42.LOGICAL.any.None: 0.[NONE].[NONE], best=rel#5183
>   rel#5183:FlinkLogicalTableSourceScan.LOGICAL.any.None: 
> 0.[NONE].[NONE](table=[myhive, default, foo, 

[jira] [Updated] (FLINK-21125) Sum or Sum0 overflow quietly

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21125?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-21125:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Sum or Sum0 overflow quietly
> 
>
> Key: FLINK-21125
> URL: https://issues.apache.org/jira/browse/FLINK-21125
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Reporter: Sebastian Liu
>Priority: Minor
>  Labels: auto-deprioritized-major
> Fix For: 1.17.0
>
>
> Overflow is not calculated correctly in the build-in sum function of Blink 
> planner.
> For a aggregate calculation such as:
> {code:java}
> CREATE TABLE TestTable (
>   amount INT
> );
> insert into TestTable (2147483647);
> insert into TestTable (1);
> SELECT sum(amount) FROM TestTable;
> The result will be: -2147483648, which is an overflowed value and no 
> exception was thrown. {code}
> The overflow occurs quietly and is difficult to detect. 
> Compare the processing semantics of other systems:
>  * *mysql*: provide two SQL mode for handling overflow. If strict SQL mode is 
> enabled, MySQL rejects the out-of-range value with an error, and the insert 
> fails, in accordance with the SQL standard. If no restrictive modes are 
> enabled, MySQL clips the value to the appropriate endpoint of the column data 
> type range and stores the resulting value instead. FYI: 
> [https://dev.mysql.com/doc/refman/8.0/en/out-of-range-and-overflow.html]
>  * *presto*: all numeric types are automatically cast to Long type, and If 
> the long is out of range, an exception is thrown to prompt user.
> IMO, exception hint is necessary, instead of quietly overflow. 
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-21765) Remove implementation-specific MetricGroup parents

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21765?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-21765:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Remove implementation-specific MetricGroup parents
> --
>
> Key: FLINK-21765
> URL: https://issues.apache.org/jira/browse/FLINK-21765
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Runtime / Metrics
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Minor
>  Labels: auto-deprioritized-major, auto-unassigned, 
> stale-assigned, starter
> Fix For: 1.17.0
>
>
> MetricGroups currently form a bi-directly graph, usually with explicit 
> requirements that type the parent must have. For example, an OperatorMG has a 
> hard requirement that the parent is a TaskMG.
> As a result they are quite inflexible, which particular shows in tests, as 
> you can't just create one metric group, but have to build an entire tree.
> The end goal of this ticket is to remove AbstractMetricGroup#parent, and 
> along the way we'll decouple the various MG implementations from each other.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-20818) End to end test produce excessive amount of logs

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20818?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-20818:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> End to end test produce excessive amount of logs
> 
>
> Key: FLINK-20818
> URL: https://issues.apache.org/jira/browse/FLINK-20818
> Project: Flink
>  Issue Type: Bug
>  Components: Test Infrastructure
>Affects Versions: 1.13.0
>Reporter: Till Rohrmann
>Priority: Minor
>  Labels: auto-deprioritized-major
> Fix For: 1.17.0
>
>
> The end to end test produce an excessive amount of logs. For example in this 
> run [1] the log file is roughly 57 MB and it is no longer possible to 
> properly scroll in this file when using the web interface. I think there 
> should not be a reason for producing almost 60 MB of log output.
> [1] 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=11467=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=ff888d9b-cd34-53cc-d90f-3e446d355529



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-20155) java.lang.OutOfMemoryError: Direct buffer memory

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20155?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-20155:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> java.lang.OutOfMemoryError: Direct buffer memory
> 
>
> Key: FLINK-20155
> URL: https://issues.apache.org/jira/browse/FLINK-20155
> Project: Flink
>  Issue Type: Bug
>  Components: Client / Job Submission
>Affects Versions: 1.11.1
>Reporter: roee hershko
>Priority: Minor
>  Labels: auto-deprioritized-major
> Fix For: 1.17.0
>
> Attachments: image-2020-11-13-17-52-54-217.png
>
>
> update:
> this issue occur every time after a job fails the only way to fix it is to 
> manually re-create the task managers pods (i am using flink operator)
>  
> after submitting a job, it runs for few hours and then the job manager is 
> crushing, when trying to re-create the job i am getting the following error:
>  
> {code:java}
> 2020-11-13 17:44:58org.apache.pulsar.client.admin.PulsarAdminException: 
> org.apache.pulsar.shade.io.netty.handler.codec.EncoderException: 
> java.lang.OutOfMemoryError: Direct buffer memoryat 
> org.apache.pulsar.client.admin.internal.BaseResource.getApiException(BaseResource.java:228)
> at 
> org.apache.pulsar.client.admin.internal.TopicsImpl$7.failed(TopicsImpl.java:324)
> at 
> org.apache.pulsar.shade.org.glassfish.jersey.client.JerseyInvocation$4.failed(JerseyInvocation.java:1030)
> at 
> org.apache.pulsar.shade.org.glassfish.jersey.client.ClientRuntime.processFailure(ClientRuntime.java:231)
> at 
> org.apache.pulsar.shade.org.glassfish.jersey.client.ClientRuntime.access$100(ClientRuntime.java:85)
> at 
> org.apache.pulsar.shade.org.glassfish.jersey.client.ClientRuntime$2.lambda$failure$1(ClientRuntime.java:183)
> at 
> org.apache.pulsar.shade.org.glassfish.jersey.internal.Errors$1.call(Errors.java:272)
> at 
> org.apache.pulsar.shade.org.glassfish.jersey.internal.Errors$1.call(Errors.java:268)
> at 
> org.apache.pulsar.shade.org.glassfish.jersey.internal.Errors.process(Errors.java:316)
> at 
> org.apache.pulsar.shade.org.glassfish.jersey.internal.Errors.process(Errors.java:298)
> at 
> org.apache.pulsar.shade.org.glassfish.jersey.internal.Errors.process(Errors.java:268)
> at 
> org.apache.pulsar.shade.org.glassfish.jersey.process.internal.RequestScope.runInScope(RequestScope.java:312)
> at 
> org.apache.pulsar.shade.org.glassfish.jersey.client.ClientRuntime$2.failure(ClientRuntime.java:183)
> at 
> org.apache.pulsar.client.admin.internal.http.AsyncHttpConnector$3.onThrowable(AsyncHttpConnector.java:279)
> at 
> org.apache.pulsar.shade.org.asynchttpclient.netty.NettyResponseFuture.abort(NettyResponseFuture.java:277)
> at 
> org.apache.pulsar.shade.org.asynchttpclient.netty.request.WriteListener.abortOnThrowable(WriteListener.java:50)
> at 
> org.apache.pulsar.shade.org.asynchttpclient.netty.request.WriteListener.operationComplete(WriteListener.java:61)
> at 
> org.apache.pulsar.shade.org.asynchttpclient.netty.request.WriteCompleteListener.operationComplete(WriteCompleteListener.java:28)
> at 
> org.apache.pulsar.shade.org.asynchttpclient.netty.request.WriteCompleteListener.operationComplete(WriteCompleteListener.java:20)
> at 
> org.apache.pulsar.shade.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:577)
> at 
> org.apache.pulsar.shade.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:551)
> at 
> org.apache.pulsar.shade.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:490)
> at 
> org.apache.pulsar.shade.io.netty.util.concurrent.DefaultPromise.addListener(DefaultPromise.java:183)
> at 
> org.apache.pulsar.shade.io.netty.channel.DefaultChannelPromise.addListener(DefaultChannelPromise.java:95)
> at 
> org.apache.pulsar.shade.io.netty.channel.DefaultChannelPromise.addListener(DefaultChannelPromise.java:30)
> at 
> org.apache.pulsar.shade.org.asynchttpclient.netty.request.NettyRequestSender.writeRequest(NettyRequestSender.java:421)
> at 
> org.apache.pulsar.shade.org.asynchttpclient.netty.channel.NettyConnectListener.writeRequest(NettyConnectListener.java:80)
> at 
> org.apache.pulsar.shade.org.asynchttpclient.netty.channel.NettyConnectListener.onSuccess(NettyConnectListener.java:156)
> at 
> org.apache.pulsar.shade.org.asynchttpclient.netty.channel.NettyChannelConnector$1.onSuccess(NettyChannelConnector.java:92)
> at 
> org.apache.pulsar.shade.org.asynchttpclient.netty.SimpleChannelFutureListener.operationComplete(SimpleChannelFutureListener.java:26)
> at 
> 

[jira] [Updated] (FLINK-19562) the /flink/docs/docker/run.sh cause an error with numeric username

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19562?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-19562:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> the /flink/docs/docker/run.sh cause an error with numeric username
> --
>
> Key: FLINK-19562
> URL: https://issues.apache.org/jira/browse/FLINK-19562
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.11.2
>Reporter: shizhengchao
>Priority: Minor
> Fix For: 1.17.0
>
>
> if the user of the OS is a number, run the script and report an error: 
> {code:java}
> groupadd: '002117' is not a valid group name
> {code}
> if the user of the OS is english, will be ok, like this:
> {code:java}
> if [ "$(uname -s)" == "Linux" ]; then
>   USER_NAME=${SUDO_USER:=$USER}
>   USER_ID=$(id -u "${USER_NAME}")
>   GROUP_ID=$(id -g "${USER_NAME}")
> else # boot2docker uid and gid
>   USER_NAME='mwopr'
>   USER_ID=1000
>   GROUP_ID=50
> fi
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-20305) Iterative job with maxWaitTimeMillis does not terminate

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20305?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-20305:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Iterative job with maxWaitTimeMillis does not terminate
> ---
>
> Key: FLINK-20305
> URL: https://issues.apache.org/jira/browse/FLINK-20305
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream
>Affects Versions: 1.12.0
>Reporter: Till Rohrmann
>Priority: Minor
>  Labels: auto-deprioritized-major, usability
> Fix For: 1.17.0
>
>
> While testing the {{DataStream}} API, I noticed that an iterative job with 
> {{maxWaitTimeMillis}} set, sometimes did not terminate depending on the used 
> parallelism. Since I used a bounded source with few events, it looked a bit 
> as if the iteration tasks couldn't terminate if they didn't receive an event. 
> The corresponding testing job including test data can be found 
> [here|https://github.com/tillrohrmann/flink-streaming-batch-execution/tree/iteration-problem].



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-23869) Correct javadoc of ExecutionConfig#registerKryoType

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-23869?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-23869:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Correct javadoc of ExecutionConfig#registerKryoType
> ---
>
> Key: FLINK-23869
> URL: https://issues.apache.org/jira/browse/FLINK-23869
> Project: Flink
>  Issue Type: Bug
>  Components: API / Core, Documentation
>Reporter: Yun Tang
>Assignee: Yun Tang
>Priority: Minor
> Fix For: 1.17.0
>
>
> Current javadoc of ExecutionConfig#registerKryoType is mistakely copied from 
> ExecutionConfig#registerPojoType, we could fix this error by copying from 
> documentation,



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-22826) flink sql1.13.1 causes data loss based on change log stream data join

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22826?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-22826:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> flink sql1.13.1 causes data loss based on change log stream data join
> -
>
> Key: FLINK-22826
> URL: https://issues.apache.org/jira/browse/FLINK-22826
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.12.0, 1.13.1
>Reporter: 徐州州
>Priority: Minor
>  Labels: auto-deprioritized-major, stale-blocker
> Fix For: 1.17.0
>
>
> {code:java}
> insert into dwd_order_detail
> select
>ord.Id,
>ord.Code,
>Status
>  concat(cast(ord.Id as String),if(oed.Id is null,'oed_null',cast(oed.Id  
> as STRING)),DATE_FORMAT(LOCALTIMESTAMP,'-MM-dd'))  as uuids,
>  TO_DATE(DATE_FORMAT(LOCALTIMESTAMP,'-MM-dd')) as As_Of_Date
> from
> orders ord
> left join order_extend oed on  ord.Id=oed.OrderId and oed.IsDeleted=0 and 
> oed.CreateTime>CAST(DATE_FORMAT(LOCALTIMESTAMP,'-MM-dd') AS TIMESTAMP)
> where ( ord.OrderTime>CAST(DATE_FORMAT(LOCALTIMESTAMP,'-MM-dd') AS 
> TIMESTAMP)
> or ord.ReviewTime>CAST(DATE_FORMAT(LOCALTIMESTAMP,'-MM-dd') AS TIMESTAMP)
> or ord.RejectTime>CAST(DATE_FORMAT(LOCALTIMESTAMP,'-MM-dd') AS TIMESTAMP)
> ) and ord.IsDeleted=0;
> {code}
> My upsert-kafka table for PRIMARY KEY for uuids.
> This is the logic of my kafka based canal-json stream data join and write to 
> Upsert-kafka tables I confirm that version 1.12 also has this problem I just 
> upgraded from 1.12 to 1.13.
> I look up a user s order data and order number XJ0120210531004794 in 
> canal-json original table as U which is normal.
> {code:java}
> | +U | XJ0120210531004794 |  50 |
> | +U | XJ0120210531004672 |  50 |
> {code}
> But written to upsert-kakfa via join, the data consumed from upsert kafka is,
> {code:java}
> | +I | XJ0120210531004794 |  50 |
> | -U | XJ0120210531004794 |  50 |
> {code}
> The order is two records this sheet in orders and order_extend tables has not 
> changed since created -U status caused my data loss not computed and the 
> final result was wrong.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-22567) OrcFsStreamingSinkITCase crashes JVM

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22567?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-22567.

Resolution: Cannot Reproduce

> OrcFsStreamingSinkITCase crashes JVM
> 
>
> Key: FLINK-22567
> URL: https://issues.apache.org/jira/browse/FLINK-22567
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / ORC, Formats (JSON, Avro, Parquet, ORC, 
> SequenceFile)
>Affects Versions: 1.14.0
>Reporter: Dawid Wysakowicz
>Priority: Minor
>  Labels: auto-deprioritized-major, test-stability
> Fix For: 1.16.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=17552=logs=e1276d0f-df12-55ec-86b5-c0ad597d83c9=906e9244-f3be-5604-1979-e767c8a6f6d9=12508
> {code}
> May 04 23:51:48 [ERROR] Process Exit Code: 239
> May 04 23:51:48 [ERROR] Crashed tests:
> May 04 23:51:48 [ERROR] org.apache.flink.orc.OrcFsStreamingSinkITCase
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.plugin.surefire.booterclient.ForkStarter.awaitResultsDone(ForkStarter.java:510)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.plugin.surefire.booterclient.ForkStarter.runSuitesForkPerTestSet(ForkStarter.java:457)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:298)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:246)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1183)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1011)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:857)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:132)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:208)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:153)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:145)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:116)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:80)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build(SingleThreadedBuilder.java:51)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:120)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:355)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.DefaultMaven.execute(DefaultMaven.java:155)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.cli.MavenCli.execute(MavenCli.java:584)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.cli.MavenCli.doMain(MavenCli.java:216)
> May 04 23:51:48 [ERROR] at 
> org.apache.maven.cli.MavenCli.main(MavenCli.java:160)
> May 04 23:51:48 [ERROR] at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> May 04 23:51:48 [ERROR] at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> May 04 23:51:48 [ERROR] at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> May 04 23:51:48 [ERROR] at java.lang.reflect.Method.invoke(Method.java:498)
> May 04 23:51:48 [ERROR] at 
> org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced(Launcher.java:289)
> May 04 23:51:48 [ERROR] at 
> org.codehaus.plexus.classworlds.launcher.Launcher.launch(Launcher.java:229)
> May 04 23:51:48 [ERROR] at 
> org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode(Launcher.java:415)
> May 04 23:51:48 [ERROR] at 
> org.codehaus.plexus.classworlds.launcher.Launcher.main(Launcher.java:356)
> May 04 23:51:48 [ERROR] Caused by: 
> org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM 
> terminated without properly saying goodbye. VM crash or System.exit called?
> May 04 23:51:48 [ERROR] Command was /bin/sh -c cd 
> /__w/1/s/flink-formats/flink-orc/target && 
> /usr/lib/jvm/java-8-openjdk-amd64/jre/bin/java -Xms256m -Xmx2048m 
> -Dmvn.forkNumber=2 -XX:+UseG1GC -jar 
> /__w/1/s/flink-formats/flink-orc/target/surefire/surefirebooter2066197408916384359.jar
>  /__w/1/s/flink-formats/flink-orc/target/surefire 
> 2021-05-04T23-33-56_338-jvmRun2 surefire8400240753057854698tmp 
> surefire_512690828581116685255tmp
> May 04 23:51:48 [ERROR] Error 

[jira] [Updated] (FLINK-21307) Revisit activation model of FlinkSecurityManager

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-21307:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Revisit activation model of FlinkSecurityManager
> 
>
> Key: FLINK-21307
> URL: https://issues.apache.org/jira/browse/FLINK-21307
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Task
>Affects Versions: 1.13.0
>Reporter: Robert Metzger
>Priority: Minor
>  Labels: auto-deprioritized-critical, auto-deprioritized-major
> Fix For: 1.17.0
>
>
> In FLINK-15156, we introduced a feature that allows users to log or 
> completely disable calls to System.exit(). This feature is enabled for 
> certain threads / code sections intended to execute user-code.
> The activation of the security manager (for monitoring user calls to 
> System.exit() is currently not well-defined, and only implemented on a 
> best-effort basis.
> This ticket is to revisit the activation.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-21898) JobRetrievalITCase crash

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21898?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-21898.

Resolution: Cannot Reproduce

> JobRetrievalITCase crash
> 
>
> Key: FLINK-21898
> URL: https://issues.apache.org/jira/browse/FLINK-21898
> Project: Flink
>  Issue Type: Bug
>  Components: Client / Job Submission
>Affects Versions: 1.13.0
>Reporter: Guowei Ma
>Priority: Minor
>  Labels: auto-deprioritized-major, test-stability
> Fix For: 1.16.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=15083=logs=8fd9202e-fd17-5b26-353c-ac1ff76c8f28=a0a633b8-47ef-5c5a-2806-3c13b9e48228=4383



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-21613) Parse Compute Column with `IN` expression throws NPE

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21613?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-21613:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Parse Compute Column with `IN` expression throws NPE
> 
>
> Key: FLINK-21613
> URL: https://issues.apache.org/jira/browse/FLINK-21613
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.13.0
>Reporter: Shuo Cheng
>Priority: Minor
>  Labels: auto-deprioritized-major, auto-unassigned
> Fix For: 1.17.0
>
>
> Considering the following given sql:
> {code:sql}
> CREATE TABLE MyInputFormatTable (
>   `a` INT,
>   `b` BIGINT,
>   `c` STRING,
>   `d` as `c` IN ('Hi', 'Hello')
> ) WITH (
>   'connector' = 'values',
>   'data-id' = '$dataId',
>   'runtime-source' = 'InputFormat'
> )
> {code}
> NPE will be thrown during parsing the sql: 
> `select * from MyInputFormatTable`
> It seems it's the commit "[hotfix][table-planner-blink] Simplify SQL 
> expression to RexNode conversion" which introduces this problem. This hotfix 
> uses a method `SqlToRelConverter#convertExpression` and this method does not 
> has any tests and is not used in Calcite anywhere, which is unsafe. Maybe 
> reverting the hotfix is a good choice.
> CC [~twalthr]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-29483) flink python udf arrow in thread model bug

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-29483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-29483:
-
Fix Version/s: 1.16.1
   (was: 1.16.0)

> flink python udf arrow in thread model bug
> --
>
> Key: FLINK-29483
> URL: https://issues.apache.org/jira/browse/FLINK-29483
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.16.0, 1.15.2
>Reporter: jackylau
>Priority: Major
> Fix For: 1.17.0, 1.15.3, 1.16.1
>
> Attachments: image-2022-09-30-17-03-05-005.png
>
>
> !image-2022-09-30-17-03-05-005.png!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-19254) Invalid UTF-8 start byte exception

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19254?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-19254:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Invalid UTF-8 start byte exception 
> ---
>
> Key: FLINK-19254
> URL: https://issues.apache.org/jira/browse/FLINK-19254
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.11.0
>Reporter: Jun Zhang
>Priority: Minor
>  Labels: auto-deprioritized-major
> Fix For: 1.17.0
>
>
> when read  no utf8 data ,JsonRowDeserializationSchema throw a exception.
> {code:java}
> Caused by: 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.core.JsonParseException:
>  Invalid UTF-8 start byte xxx 
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-19482) OrcRowInputFormat does not define serialVersionUID

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-19482:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> OrcRowInputFormat does not define serialVersionUID
> --
>
> Key: FLINK-19482
> URL: https://issues.apache.org/jira/browse/FLINK-19482
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.11.0
>Reporter: hailong wang
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
> Fix For: 1.17.0
>
>
> The org.apache.flink.orc function does not define a {{serialVersionUID}}.
> We should define a {{serialVersionUID when object is serialized to avoid 
> [java.io|http://java.io/].InvalidClassExceptio exception.}}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-18777) Supports schema registry catalog

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo updated FLINK-18777:
-
Fix Version/s: 1.17.0
   (was: 1.16.0)

> Supports schema registry catalog
> 
>
> Key: FLINK-18777
> URL: https://issues.apache.org/jira/browse/FLINK-18777
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.11.0
>Reporter: Danny Chen
>Priority: Minor
>  Labels: auto-deprioritized-major, pull-request-available
> Fix For: 1.17.0
>
>
> Design doc: 
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-125%3A+Confluent+Schema+Registry+Catalog



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-17482) KafkaITCase.testMultipleSourcesOnePartition unstable

2022-09-30 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo closed FLINK-17482.

Resolution: Cannot Reproduce

> KafkaITCase.testMultipleSourcesOnePartition unstable
> 
>
> Key: FLINK-17482
> URL: https://issues.apache.org/jira/browse/FLINK-17482
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka, Tests
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Robert Metzger
>Assignee: Fabian Paul
>Priority: Minor
>  Labels: auto-deprioritized-major, stale-assigned, test-stability
> Fix For: 1.16.0
>
>
> CI run: 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=454=logs=c5f0071e-1851-543e-9a45-9ac140befc32=684b1416-4c17-504e-d5ab-97ee44e08a20
> {code}
> 07:29:40,472 [main] INFO  
> org.apache.flink.streaming.connectors.kafka.KafkaTestBase[] - 
> -
> [ERROR] Tests run: 22, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 
> 152.018 s <<< FAILURE! - in 
> org.apache.flink.streaming.connectors.kafka.KafkaITCase
> [ERROR] 
> testMultipleSourcesOnePartition(org.apache.flink.streaming.connectors.kafka.KafkaITCase)
>   Time elapsed: 4.257 s  <<< FAILURE!
> java.lang.AssertionError: Test failed: Job execution failed.
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.apache.flink.test.util.TestUtils.tryExecute(TestUtils.java:45)
>   at 
> org.apache.flink.streaming.connectors.kafka.KafkaConsumerTestBase.runMultipleSourcesOnePartitionExactlyOnceTest(KafkaConsumerTestBase.java:963)
>   at 
> org.apache.flink.streaming.connectors.kafka.KafkaITCase.testMultipleSourcesOnePartition(KafkaITCase.java:107)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
>   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
>   at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>   at java.lang.Thread.run(Thread.java:748)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


  1   2   3   4   5   6   7   8   9   10   >