退订

2021-06-18 Thread 伟波
退订


| |
伟波
|
|
weib...@126.com
|
签名由网易邮箱大师定制

Re: Re: [ANNOUNCE] New PMC member: Xintong Song

2021-06-18 Thread Roman Khachatryan
Congratulations!

Regards,
Roman

On Fri, Jun 18, 2021 at 4:40 AM Yu Li  wrote:
>
> Congratulations, Xintong!
>
> Best Regards,
> Yu
>
>
> On Thu, 17 Jun 2021 at 15:23, Yuan Mei  wrote:
>
> > Congratulations, Xintong :-)
> >
> > On Thu, Jun 17, 2021 at 11:57 AM Xingbo Huang  wrote:
> >
> > > Congratulations, Xintong!
> > >
> > > Best,
> > > Xingbo
> > >
> > > Yun Gao  于2021年6月17日周四 上午10:46写道:
> > >
> > > > Congratulations, Xintong!
> > > >
> > > > Best,
> > > > Yun
> > > >
> > > >
> > > > --
> > > > Sender:Jingsong Li
> > > > Date:2021/06/17 10:41:22
> > > > Recipient:dev
> > > > Theme:Re: [ANNOUNCE] New PMC member: Xintong Song
> > > >
> > > > Congratulations, Xintong!
> > > >
> > > > Best,
> > > > Jingsong
> > > >
> > > > On Thu, Jun 17, 2021 at 10:26 AM Yun Tang  wrote:
> > > >
> > > > > Congratulations, Xintong!
> > > > >
> > > > > Best
> > > > > Yun Tang
> > > > > 
> > > > > From: Leonard Xu 
> > > > > Sent: Wednesday, June 16, 2021 21:05
> > > > > To: dev (dev@flink.apache.org) 
> > > > > Subject: Re: [ANNOUNCE] New PMC member: Xintong Song
> > > > >
> > > > >
> > > > > Congratulations, Xintong!
> > > > >
> > > > >
> > > > > Best,
> > > > > Leonard
> > > > > > 在 2021年6月16日,20:07,Till Rohrmann  写道:
> > > > > >
> > > > > > Congratulations, Xintong!
> > > > > >
> > > > > > Cheers,
> > > > > > Till
> > > > > >
> > > > > > On Wed, Jun 16, 2021 at 1:47 PM JING ZHANG 
> > > > wrote:
> > > > > >
> > > > > >> Congratulations, Xintong!
> > > > > >>
> > > > > >>
> > > > > >> Jiayi Liao  于2021年6月16日周三 下午7:30写道:
> > > > > >>
> > > > > 
> > > > >  <
> > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/>
> > > > >  Congratulations Xintong!
> > > > > 
> > > > >  On Wed, Jun 16, 2021 at 7:24 PM Nicholas Jiang <
> > > programg...@163.com
> > > > >
> > > > >  wrote:
> > > > > 
> > > > > > Congratulations, Xintong!
> > > > > >
> > > > > >
> > > > > >
> > > > > > --
> > > > > > Sent from:
> > > > > >
> > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/
> > > > > 
> > > > > 
> > > > > >>>
> > > > > >>
> > > > >
> > > > >
> > > >
> > > > --
> > > > Best, Jingsong Lee
> > > >
> > > >
> > >
> >


Re: [DISCUSS] Releasing Flink 1.11.4, 1.12.5, 1.13.2

2021-06-18 Thread Piotr Nowojski
Hi,

Thanks for bringing this up. I think before releasing 1.12.x/1.13.x/1.14.x,
it would be good to decide what to do with FLINK-23011 [1] and if there is
a relatively easy fix, I would wait for it before releasing.

Best,
Piotrek

[1] with https://issues.apache.org/jira/browse/FLINK-23011

pt., 18 cze 2021 o 16:35 Konstantin Knauf  napisał(a):

> Hi Dawid,
>
> Thank you for starting the discussion. I'd like to add
> https://issues.apache.org/jira/browse/FLINK-23025 to the list for Flink
> 1.13.2.
>
> Cheers,
>
> Konstantin
>
> On Fri, Jun 18, 2021 at 3:26 PM Dawid Wysakowicz 
> wrote:
>
> > Hi devs,
> >
> > Quite recently we pushed, in our opinion, quite an important fix[1] for
> > unaligned checkpoints which disables UC for broadcast partitioning.
> > Without the fix there might be some broadcast state corruption.
> > Therefore we think it would be beneficial to release it soonish. What do
> > you think? Do you have other issues in mind you'd like to have included
> > in these versions.
> >
> > Would someone be willing to volunteer to help with the releases as a
> > release manager? I guess there is a couple of spots to fill in here ;)
> >
> > Best,
> >
> > Dawid
> >
> > [1] https://issues.apache.org/jira/browse/FLINK-22815
> >
> >
> >
>
> --
>
> Konstantin Knauf
>
> https://twitter.com/snntrable
>
> https://github.com/knaufk
>


[jira] [Created] (FLINK-23041) Change local alignment timeout back to the global time out

2021-06-18 Thread Anton Kalashnikov (Jira)
Anton Kalashnikov created FLINK-23041:
-

 Summary: Change local alignment timeout back to the global time out
 Key: FLINK-23041
 URL: https://issues.apache.org/jira/browse/FLINK-23041
 Project: Flink
  Issue Type: Bug
Reporter: Anton Kalashnikov
Assignee: Anton Kalashnikov


Local alignment timeouts are very confusing and especially without timeout on 
the outputs, they can significantly delay timeouting to UC.

Problematic case is when all CBs are received with long delay because of the 
back pressure, but they arrive at the same time. Alignment time can be low 
(milliseconds), while start delay is ~1 minute. In that case checkpoint doesn't 
timeout to UC and is passing the responsibility to timeout down the stream.

 

So it is not so transparant for the user why and when AC switches to UC. As 
mentioned before, the start delay is not correlated with the alignment timeout 
because it doesn't take into account time in output buffer. the alignment time 
is not fully correlated with the alignment timeout because the alignment time 
doesn't take into account the barrier announcement.

 

Based on this, there is the proposal to change the semantic of alignmentTimeout 
configuration to such meaning:

*The time between the starting of checkpoint(on the checkpont coordinator) and 
the time when the checkpoint barrier will be received by task.*

By this definition, we will have kind of global timeout which says that if the 
AC isn't finished for alignmentTimeout time it will be switched to UC.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [DISCUSS] Releasing Flink 1.11.4, 1.12.5, 1.13.2

2021-06-18 Thread Konstantin Knauf
Hi Dawid,

Thank you for starting the discussion. I'd like to add
https://issues.apache.org/jira/browse/FLINK-23025 to the list for Flink
1.13.2.

Cheers,

Konstantin

On Fri, Jun 18, 2021 at 3:26 PM Dawid Wysakowicz 
wrote:

> Hi devs,
>
> Quite recently we pushed, in our opinion, quite an important fix[1] for
> unaligned checkpoints which disables UC for broadcast partitioning.
> Without the fix there might be some broadcast state corruption.
> Therefore we think it would be beneficial to release it soonish. What do
> you think? Do you have other issues in mind you'd like to have included
> in these versions.
>
> Would someone be willing to volunteer to help with the releases as a
> release manager? I guess there is a couple of spots to fill in here ;)
>
> Best,
>
> Dawid
>
> [1] https://issues.apache.org/jira/browse/FLINK-22815
>
>
>

-- 

Konstantin Knauf

https://twitter.com/snntrable

https://github.com/knaufk


[jira] [Created] (FLINK-23040) Consider ConfigOption fallback keys in FactoryUtil

2021-06-18 Thread Timo Walther (Jira)
Timo Walther created FLINK-23040:


 Summary: Consider ConfigOption fallback keys in FactoryUtil
 Key: FLINK-23040
 URL: https://issues.apache.org/jira/browse/FLINK-23040
 Project: Flink
  Issue Type: Sub-task
  Components: Table SQL / API
Reporter: Timo Walther
Assignee: Timo Walther


We are currently not taking fallback keys into consideration when performing 
validation in FactoryUtil.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[DISCUSS] Releasing Flink 1.11.4, 1.12.5, 1.13.2

2021-06-18 Thread Dawid Wysakowicz
Hi devs,

Quite recently we pushed, in our opinion, quite an important fix[1] for
unaligned checkpoints which disables UC for broadcast partitioning.
Without the fix there might be some broadcast state corruption.
Therefore we think it would be beneficial to release it soonish. What do
you think? Do you have other issues in mind you'd like to have included
in these versions.

Would someone be willing to volunteer to help with the releases as a
release manager? I guess there is a couple of spots to fill in here ;)

Best,

Dawid

[1] https://issues.apache.org/jira/browse/FLINK-22815




OpenPGP_signature
Description: OpenPGP digital signature


[jira] [Created] (FLINK-23039) Support pluggable transports for HTTP endpoints

2021-06-18 Thread Igal Shilman (Jira)
Igal Shilman created FLINK-23039:


 Summary: Support pluggable transports for HTTP endpoints
 Key: FLINK-23039
 URL: https://issues.apache.org/jira/browse/FLINK-23039
 Project: Flink
  Issue Type: Improvement
  Components: Stateful Functions
Reporter: Igal Shilman
 Fix For: statefun-3.1.0


We've recently learned about a use case that requires using a custom client 
that dispatches the HTTP requests (due to some internal reasons).

This can be a useful addition to further customizing the exact client code, 
that suites the user's need. (for example adding company specific tracing 
information)

This is technically feasible as well, as all it takes is:

1) provide an implementation of this 
[interface|https://github.com/apache/flink-statefun/blob/master/statefun-flink/statefun-flink-core/src/main/java/org/apache/flink/statefun/flink/core/reqreply/RequestReplyClient.java]

2) extend the ability to configure it from information present at the 
module.yaml.

 

The current proposal is to add an optional "transport" section to endpoint 
definition:
{code:java}
- endpoint:
meta:
  kind: http
spec:
  functions: com.foo.bar/*
  transport:
provider_class: com.foo.bar.ClientProvider
some:
  internal:
property: 123
  urlPathTemplate: http://bar.foo.com:8080/functions/{function.name}
  maxNumBatchRequests: 1 {code}
If the transport is not present we assume that the StateFun's pre-bundled 
transport is present.

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-23038) Broken link in Documentation Style Guide

2021-06-18 Thread Daisy Tsang (Jira)
Daisy Tsang created FLINK-23038:
---

 Summary: Broken link in Documentation Style Guide
 Key: FLINK-23038
 URL: https://issues.apache.org/jira/browse/FLINK-23038
 Project: Flink
  Issue Type: Bug
Reporter: Daisy Tsang


The link to the Flink Glossary is broken here: 
https://flink.apache.org/contributing/docs-style.html



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-23037) Document limitations of UC

2021-06-18 Thread Dawid Wysakowicz (Jira)
Dawid Wysakowicz created FLINK-23037:


 Summary: Document limitations of UC
 Key: FLINK-23037
 URL: https://issues.apache.org/jira/browse/FLINK-23037
 Project: Flink
  Issue Type: Improvement
  Components: Documentation, Runtime / Checkpointing
Reporter: Dawid Wysakowicz
Assignee: Dawid Wysakowicz
 Fix For: 1.14.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-23036) org.apache.flink.util.FlinkRuntimeException: Error while adding data to RocksDB

2021-06-18 Thread guxiang (Jira)
guxiang created FLINK-23036:
---

 Summary: org.apache.flink.util.FlinkRuntimeException: Error while 
adding data to RocksDB
 Key: FLINK-23036
 URL: https://issues.apache.org/jira/browse/FLINK-23036
 Project: Flink
  Issue Type: Bug
  Components: API / Core
Affects Versions: 1.13.1, 1.12.2, 1.9.4
Reporter: guxiang


I had a similar problem to this one
{quote}[http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Checkpoint-failed-because-of-TimeWindow-cannot-be-cast-to-VoidNamespace-td36310.html]

https://issues.apache.org/jira/browse/FLINK-18464
{quote}
I tested three versions of Flink, and all of them experienced this problem 

1.9.0 / 1.12.2 / 1.13.1

This is because I defined a state (VoidNamespace) on a trigger and I used the 
state  in TimeWindow (WindowNamespace).

Runs locally and does not use backendState.

When you run locally and do not use backendstate, the program runs normally.
However, when backendState is used, an error will be reported

 
{code:java}
//代码占位符

org.apache.flink.util.FlinkRuntimeException: Error while adding data to RocksDB
at 
org.apache.flink.contrib.streaming.state.RocksDBValueState.update(RocksDBValueState.java:109)
at 
com.sankuai.grocery.crm.data.mallorg.trigger.MessageProcessOnTimeStateTrigger.onEventTime(MessageProcessOnTimeStateTrigger.java:116)
at 
com.sankuai.grocery.crm.data.mallorg.trigger.MessageProcessOnTimeStateTrigger.onEventTime(MessageProcessOnTimeStateTrigger.java:23)
at 
org.apache.flink.streaming.runtime.operators.windowing.WindowOperator$Context.onEventTime(WindowOperator.java:944)
at 
org.apache.flink.streaming.runtime.operators.windowing.WindowOperator.onEventTime(WindowOperator.java:481)
at 
org.apache.flink.streaming.api.operators.InternalTimerServiceImpl.advanceWatermark(InternalTimerServiceImpl.java:302)
at 
org.apache.flink.streaming.api.operators.InternalTimeServiceManagerImpl.advanceWatermark(InternalTimeServiceManagerImpl.java:194)
at 
org.apache.flink.streaming.api.operators.AbstractStreamOperator.processWatermark(AbstractStreamOperator.java:626)
at 
org.apache.flink.streaming.runtime.tasks.OneInputStreamTask$StreamTaskNetworkOutput.emitWatermark(OneInputStreamTask.java:197)
at 
org.apache.flink.streaming.runtime.streamstatus.StatusWatermarkValve.findAndOutputNewMinWatermarkAcrossAlignedChannels(StatusWatermarkValve.java:196)
at 
org.apache.flink.streaming.runtime.streamstatus.StatusWatermarkValve.inputWatermark(StatusWatermarkValve.java:105)
at 
org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.processElement(StreamTaskNetworkInput.java:206)
at 
org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.emitNext(StreamTaskNetworkInput.java:174)
at 
org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:65)
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:396)
at 
org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:191)
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:617)
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:581)
at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:755)
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:570)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.ClassCastException: 
org.apache.flink.streaming.api.windowing.windows.TimeWindow cannot be cast to 
org.apache.flink.runtime.state.VoidNamespace
at 
org.apache.flink.runtime.state.VoidNamespaceSerializer.serialize(VoidNamespaceSerializer.java:30)
at 
org.apache.flink.contrib.streaming.state.RocksDBKeySerializationUtils.writeNameSpace(RocksDBKeySerializationUtils.java:78)
at 
org.apache.flink.contrib.streaming.state.RocksDBSerializedCompositeKeyBuilder.serializeNamespace(RocksDBSerializedCompositeKeyBuilder.java:175)
at 
org.apache.flink.contrib.streaming.state.RocksDBSerializedCompositeKeyBuilder.buildCompositeKeyNamespace(RocksDBSerializedCompositeKeyBuilder.java:112)
at 
org.apache.flink.contrib.streaming.state.AbstractRocksDBState.serializeCurrentKeyWithGroupAndNamespace(AbstractRocksDBState.java:163)
at 
org.apache.flink.contrib.streaming.state.RocksDBValueState.update(RocksDBValueState.java:106)
... 20 more
{code}
 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Change in accumutors semantics with jobClient

2021-06-18 Thread Etienne Chauchot

Hi all,

I did a fix some time ago regarding accumulators: 
the/JobClient.getAccumulators()/ was infinitely  blocking in local 
environment for a streaming job (1). The change (2) consisted of giving 
the current accumulators value for the running job. And when fixing this 
in the PR, it appeared that I had to change the accumulators semantics 
with /JobClient/ and I just realized that I forgot to bring this back to 
the ML:


Previously /JobClient/ assumed that getAccumulator() was called on a 
bounded pipeline and that the user wanted to acquire the *final 
accumulator values* after the job is finished.


But now it returns the *current value of accumulators* immediately to be 
compatible with unbounded pipelines.


If it is run on a bounded pipeline, then to get the final accumulator 
values after the job is finished, one needs to call 
/getJobExecutionResult().thenApply(JobExecutionResult::getAllAccumulatorResults)/


(1): https://issues.apache.org/jira/browse/FLINK-18685

(2): https://github.com/apache/flink/pull/14558#


Cheers,

Etienne



[jira] [Created] (FLINK-23035) Add explicit method to StateChangelogWriter to write metadata

2021-06-18 Thread Roman Khachatryan (Jira)
Roman Khachatryan created FLINK-23035:
-

 Summary: Add explicit method to StateChangelogWriter to write 
metadata
 Key: FLINK-23035
 URL: https://issues.apache.org/jira/browse/FLINK-23035
 Project: Flink
  Issue Type: Sub-task
  Components: Runtime / State Backends
Reporter: Roman Khachatryan
 Fix For: 1.14.0


Currently, metadata is written to the state changelog using the same 
StateChangelogWriter.append() method as data.

However, it doesn't belong to a specific group, and should be read first on 
recovery. Because of that, -1 is used.

An explicit append() without keygroup would be less fragile (probably still 
using -1 under the hood).



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-23034) NPE in JobDetailsDeserializer during the reading old version of ExecutionState

2021-06-18 Thread Anton Kalashnikov (Jira)
Anton Kalashnikov created FLINK-23034:
-

 Summary: NPE in JobDetailsDeserializer during the reading old 
version of ExecutionState
 Key: FLINK-23034
 URL: https://issues.apache.org/jira/browse/FLINK-23034
 Project: Flink
  Issue Type: Bug
Reporter: Anton Kalashnikov
Assignee: Anton Kalashnikov


There is no compatibility for ExecutionState:

{noformat}

java.lang.NullPointerExceptionjava.lang.NullPointerException at 
org.apache.flink.runtime.messages.webmonitor.JobDetails$JobDetailsDeserializer.deserialize(JobDetails.java:308)
 at 
org.apache.flink.runtime.messages.webmonitor.JobDetails$JobDetailsDeserializer.deserialize(JobDetails.java:278)
 at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.deser.DefaultDeserializationContext.readRootValue(DefaultDeserializationContext.java:322)
 at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper._readMapAndClose(ObjectMapper.java:4593)
 at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3479)
 at 
org.apache.flink.runtime.messages.webmonitor.JobDetailsTest.testJobDetailsCompatibleUnmarshalling(JobDetailsTest.java:82)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:498) at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
 at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
 at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
 at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
 at org.apache.flink.util.TestNameProvider$1.evaluate(TestNameProvider.java:45) 
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at 
org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at 
org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
 at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
 at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
 at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at 
org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at 
org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at 
org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at 
org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at 
org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at 
org.junit.runners.ParentRunner.run(ParentRunner.java:413) at 
org.junit.runner.JUnitCore.run(JUnitCore.java:137) at 
com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:69)
 at 
com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:33)
 at 
com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:221)
 at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:54)

{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-23033) Support object array in Python DataStream API

2021-06-18 Thread Dian Fu (Jira)
Dian Fu created FLINK-23033:
---

 Summary: Support object array in Python DataStream API
 Key: FLINK-23033
 URL: https://issues.apache.org/jira/browse/FLINK-23033
 Project: Flink
  Issue Type: Improvement
  Components: API / Python
Reporter: Dian Fu
 Fix For: 1.14.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [DISCUSS] FLIP-129 (Update): Registering sources/sinks on Table API without SQL

2021-06-18 Thread Ingo Bürk
Thanks everyone for the feedback! Since there haven't been any
questions/discussions, I'll be moving this to a vote thread next week.


Regards
Ingo

On Tue, Jun 15, 2021 at 5:36 PM JING ZHANG  wrote:

> Thanks Ingo for starting this discussion.
>
>
>
> Big +1 for this feature. Looking forward to the feature.
>
>
> Best regards,
> JING ZHANG
>
> Jark Wu  于2021年6月15日周二 下午3:48写道:
>
> > Thanks Ingo for picking up this FLIP.
> >
> > FLIP-129 is an important piece to have a complete Table SQL story,
> > and users have been waiting for a long time. Let's finish it in this
> > release!
> > Your proposed changes look good to me.
> >
> > I also cc'd people who voted in previous FLIP-129.
> >
> > Best,
> > Jark
> >
> > On Thu, 10 Jun 2021 at 19:46, Ingo Bürk  wrote:
> >
> > > Hello everyone,
> > >
> > > we would like to pick up work on FLIP-129 which aims to improve the
> Table
> > > API by supporting the creation of sources / sinks without having to go
> > > through SQL/DDL. This FLIP was approved a while ago, and some things
> have
> > > changed since then. We'd like to propose a few changes, see [1], before
> > > starting work on it.
> > > Our proposal is mainly motivated by reducing the scope in some parts to
> > > improve maintainability and relying more on ConfigOptions being the
> > single
> > > source of truth. We also want to expose this functionality for
> > > non-temporary tables.
> > >
> > > We'd like to open this for discussion to collect any feedback. Once the
> > > discussion has stabilized I'll update the FLIP itself and start a new
> > vote.
> > >
> > > [1]
> > >
> > >
> >
> https://docs.google.com/document/d/1tpirvF0u723QF005UrgdbvF-Tp0Jbg_qhlbda4mk7Ck/edit?usp=sharing
> > >
> > >
> > > Regards
> > > Ingo
> > >
> >
>


Re: [DISCUSS] FLIP-171: Async Sink

2021-06-18 Thread Arvid Heise
Hi Danny,

to add I'd propose to use the flink-connector-base package which has the
rough equivalent on source-side SourceReaderBase [1]. Since it's such a
handy base implementation, I'd like to see it directly in the main flink
repository.

For the actual connectors, I'm currently working on a proposal for a common
connector repository under Flink umbrella.

[1]
https://github.com/AHeise/flink/blob/master/flink-connectors/flink-connector-base/src/main/java/org/apache/flink/connector/base/source/reader/SourceReaderBase.java#L58-L58

On Wed, Jun 16, 2021 at 6:06 PM Hausmann, Steffen 
wrote:

> Hi Danny,
>
> Right now, I'd expect the core of the Async Sink (without third party
> dependencies) to live in its own submodule. For instance
> `flink-connector-async` as part of `flink-connectors`.
>
> I'm currently planning to implement three different sinks to verify that
> the design of the sink if flexible enough to support different services:
> Amazon Kinesis Data Streams, Amazon Kinesis Data Firehose, and Amazon
> DynamoDB. But I'm not sure where to actually put them. To keep is simple,
> I'd start with a module that contains all AWS specific connectors. However,
> it has the obvious disadvantage that if someone wants to use a single sink,
> they would need to pull in all dependencies for all supported services that
> are included in this module (mainly the AWS SDK for these services). But I
> don't know how much of a problem that's going to be in practice. If the
> respective jar grows too big because all the included dependencies, that's
> certainly not going to work. But for now I'd just give it a try and then
> start a discussion once I have more data to share.
>
> What's more interesting is whether that module should be part of the Flink
> code base or live somewhere else. I'd be great to get some feedback from
> the community on this.
>
> Regarding the Kinesis Data Streams sink, I fully agree that it would be
> nice to remove the dependency to the KPL. So it seems to be desirable to
> keep the existing and the new FLIP-171  based implementation in separate
> modules. Otherwise people would be forced to pull in the KPL dependencies,
> even if they are only using the new implementation. In addition, the new
> implementation will not support the exact same functionality as the
> existing one: the KPL implements a very optimized form of aggregation on a
> shard level [1] by maintaining a mapping of shards and their respective key
> spaces. The new implementation can in principle support aggregation as
> well, but only on a partition key level, which may lead to less efficient
> aggregation and higher latencies.
>
> Cheers, Steffen
>
> [1]
> https://docs.aws.amazon.com/streams/latest/dev/kinesis-kpl-concepts.html#kinesis-kpl-concepts-aggretation
>
>
>
> On 15.06.21, 19:52, "Cranmer, Danny" 
> wrote:
>
> CAUTION: This email originated from outside of the organization. Do
> not click links or open attachments unless you can confirm the sender and
> know the content is safe.
>
>
>
> Hey Steffen,
>
> I have a few questions regarding the FLIP:
> 1. Where do you expect the core code to live, would it be in an
> existing module (say flink-clients) or would you introduce a new module?
> 2. Which destination implementations do you intend to ship with this
> FLIP? I see an example with Kinesis but you also list a bunch of other
> candidates.
> 3. For the Kinesis implementation, would you add the Sink to the
> existing flink-connector-kinesis repo, or create a new module? Reason I ask
> is that the existing Kinesis Sink depends on KPL and has a heavy transitive
> dependency chain, removing this would substantially reduce application size
> and clean the dependency chain
>
> Thanks,
>
> On 10/06/2021, 09:09, "Hausmann, Steffen" 
> wrote:
>
> CAUTION: This email originated from outside of the organization.
> Do not click links or open attachments unless you can confirm the sender
> and know the content is safe.
>
>
>
> Hey Piotrek,
>
> Thanks for your comments on the FLIP. I'll address your second
> question first, as I think it's more central to this FLIP. Just looking at
> the AWS ecosystem, there are several sinks with overlapping functionality.
> I've chosen AWS sinks here because I'm most familiar with those, but a
> similar argument applies more generically for destination that support
> async ingest.
>
> There is, for instance, a sink for Amazon Kinesis Data Streams
> that is part of Apache Flink [1], a sink for Amazon Kinesis Data Firehose
> [2], a sink for Amazon DynamoDB [3], and a sink for Amazon Timestream [4].
> All these sinks have implemented their own mechanisms for batching,
> persisting, and retrying events. And I'm not sure if all of them properly
> participate in checkpointing. [3] even seems to closely mirror [1] as it
> contains references to the Kinesis Producer Library, which is unrelated to
> Amazon DynamoDB.
>
> These sinks 

[jira] [Created] (FLINK-23032) Refactor HiveSource to make it usable in data stream job

2021-06-18 Thread Rui Li (Jira)
Rui Li created FLINK-23032:
--

 Summary: Refactor HiveSource to make it usable in data stream job
 Key: FLINK-23032
 URL: https://issues.apache.org/jira/browse/FLINK-23032
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / Hive
Reporter: Rui Li
Assignee: Rui Li
 Fix For: 1.14.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-23031) Support to emit window result with periodic or non_periodic

2021-06-18 Thread Aitozi (Jira)
Aitozi created FLINK-23031:
--

 Summary: Support to emit window result with periodic or 
non_periodic
 Key: FLINK-23031
 URL: https://issues.apache.org/jira/browse/FLINK-23031
 Project: Flink
  Issue Type: Improvement
  Components: Table SQL / Planner
Reporter: Aitozi






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-23030) PartitionRequestClientFactory#createPartitionRequestClient should throw when network failure

2021-06-18 Thread Jin Xing (Jira)
Jin Xing created FLINK-23030:


 Summary: 
PartitionRequestClientFactory#createPartitionRequestClient should throw when 
network failure
 Key: FLINK-23030
 URL: https://issues.apache.org/jira/browse/FLINK-23030
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Network
Reporter: Jin Xing


In current _PartitionRequestClientFactory#createPartitionRequestClient_, 
_ChannelFuture#await()_ is invoked, thus to build a connection to remote 
synchronously.

But with the doc of _io.netty.util.concurrent.Future_ [1] and its 
implementation _io.netty.channel.DefaultChannelPromise_ [2], 
_ChannelFuture#await()_ never throws when completed with failure. I guess what 
Flink needs is _ChannelFuture#sync()._

 

[1][https://netty.io/4.1/api/io/netty/util/concurrent/class-use/Future.html]

[2] 
https://github.com/netty/netty/blob/4.1/transport/src/main/java/io/netty/channel/DefaultChannelPromise.java



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-23029) Extend PluginLoader to allow closing of backing ClassLoader

2021-06-18 Thread Chesnay Schepler (Jira)
Chesnay Schepler created FLINK-23029:


 Summary: Extend PluginLoader to allow closing of backing 
ClassLoader
 Key: FLINK-23029
 URL: https://issues.apache.org/jira/browse/FLINK-23029
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / Coordination
Reporter: Chesnay Schepler
Assignee: Chesnay Schepler
 Fix For: 1.14.0


The PluginLoader is currently never closing the backing classloader, which 
means that plugins remain loaded until the application shuts down.

This for example makes it impossible to use the plugins in the mini cluster, 
because they would be loaded multiple times.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-23028) Improve documentation for pages of SQL.

2021-06-18 Thread Roc Marshal (Jira)
Roc Marshal created FLINK-23028:
---

 Summary: Improve documentation for pages of SQL.
 Key: FLINK-23028
 URL: https://issues.apache.org/jira/browse/FLINK-23028
 Project: Flink
  Issue Type: Improvement
  Components: Documentation
Reporter: Roc Marshal


Wrong style in 
[https://ci.apache.org/projects/flink/flink-docs-master/zh/docs/dev/table/sql/create/#create-function]
   section.

Wrong style in 
[https://ci.apache.org/projects/flink/flink-docs-master/zh/docs/dev/table/sql/drop/#drop-function]
 section.

Wrong style in 
[https://ci.apache.org/projects/flink/flink-docs-master/zh/docs/dev/table/sql/alter/#alter-function]
 section.

Add the description about drop catalog in the page of  
[https://ci.apache.org/projects/flink/flink-docs-master/docs/dev/table/sql/drop/]
 .

'catloag' of  
[https://ci.apache.org/projects/flink/flink-docs-master/zh/docs/dev/table/sql/use/#use-catloag]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-23027) Move chill dependency to flink-scala

2021-06-18 Thread Chesnay Schepler (Jira)
Chesnay Schepler created FLINK-23027:


 Summary: Move chill dependency to flink-scala
 Key: FLINK-23027
 URL: https://issues.apache.org/jira/browse/FLINK-23027
 Project: Flink
  Issue Type: Sub-task
  Components: API / Type Serialization System
Reporter: Chesnay Schepler
Assignee: Chesnay Schepler
 Fix For: 1.14.0


flink-runtime contains a few serializer classes that rely on scala. We could 
move these to flink-scala, which already depends on scala anyway, and is also 
bundled in flink-dist.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)