[jira] [Created] (FLINK-17935) Logs could not show up when deploying Flink on Yarn via "--executor"

2020-05-25 Thread Yang Wang (Jira)
Yang Wang created FLINK-17935:
-

 Summary: Logs could not show up when deploying Flink on Yarn via 
"--executor"
 Key: FLINK-17935
 URL: https://issues.apache.org/jira/browse/FLINK-17935
 Project: Flink
  Issue Type: Bug
Affects Versions: 1.11.0, 1.12.0
Reporter: Yang Wang
 Fix For: 1.11.0


{code:java}
./bin/flink run -d -p 5 -e yarn-per-job examples/streaming/WindowJoin.jar{code}
When we use the {{-e/--executor}} to specify the deploy target to Yarn per-job, 
the logs could not show up. The root cause is we do not set the logging files 
in {{ExecutorCLI}}. We only do it in the {{FlinkYarnSessionCli}}.

If we use {{-m yarn-cluster}}, everything works well.

 

Maybe we should move the {{setLogConfigFileInConfig}} to 
{{YarnClusterDescriptor}} to avoid this problem. cc [~kkl0u]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-17934) StreamingFileWriter should set chainingStrategy

2020-05-25 Thread Jingsong Lee (Jira)
Jingsong Lee created FLINK-17934:


 Summary: StreamingFileWriter should set chainingStrategy
 Key: FLINK-17934
 URL: https://issues.apache.org/jira/browse/FLINK-17934
 Project: Flink
  Issue Type: Bug
  Components: Connectors / FileSystem
Reporter: Jingsong Lee
 Fix For: 1.11.0


We should let {{StreamingFileWriter}} should be eagerly chained whenever 
possible.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [DISCUSS] Releasing Stateful Functions 2.1.0 soon?

2020-05-25 Thread Stephan Ewen
Nice work, thanks for pushing this, Gordon!

+1 also from my side for a quick release.

I think it already warrants a release to have the 1.10.1 upgrade and the
fix to not fail on savepoints that are triggered concurrently to a
checkpoint.
Even nicer that there are two cool new features included.

On Fri, May 22, 2020 at 7:13 AM Tzu-Li (Gordon) Tai 
wrote:

> Thanks for the positive feedback so far.
>
> Lets then set the feature freeze date for Stateful Functions 2.1.0 to be
> next Wednesday (May 27th).
>
> We've made good progress over the past days, all mentioned features merged
> besides the following:
> - https://issues.apache.org/jira/browse/FLINK-17875 State TTL support for
> remote functions
>
> Will keep track of that and hopefully cut the feature branch as scheduled.
>
> Cheers,
> Gordon
>
> On Thu, May 21, 2020 at 7:22 PM Yuan Mei  wrote:
>
> > faster iteration definitely helps early-stage projects.
> >
> > +1
> >
> > Best,
> > Yuan
> >
> >
> > On Thu, May 21, 2020 at 4:14 PM Congxian Qiu 
> > wrote:
> >
> > > +1 from my side to have smaller and more frequent feature releases for
> > the
> > > project in its early phases.
> > >
> > > Best,
> > > Congxian
> > >
> > >
> > > Marta Paes Moreira  于2020年5月21日周四 下午12:49写道:
> > >
> > > > +1 for more frequent releases with a shorter (but feedback-informed)
> > > > feature set.
> > > >
> > > > Thanks, Gordon (and Igal)!
> > > >
> > > > Marta
> > > >
> > > > On Thu, 21 May 2020 at 03:44, Yu Li  wrote:
> > > >
> > > > > +1, it makes a lot of sense for stateful functions to evolve
> faster.
> > > > >
> > > > > Best Regards,
> > > > > Yu
> > > > >
> > > > >
> > > > > On Wed, 20 May 2020 at 23:36, Zhijiang  > > > > .invalid>
> > > > > wrote:
> > > > >
> > > > > > I also like this idea, considering stateful functions flexible
> > enough
> > > > to
> > > > > > have a faster release cycle. +1 from my side.
> > > > > >
> > > > > > Best,
> > > > > > Zhijiang
> > > > > >
> > > > > >
> > > > > >
> --
> > > > > > From:Seth Wiesman 
> > > > > > Send Time:2020年5月20日(星期三) 21:45
> > > > > > To:dev 
> > > > > > Subject:Re: [DISCUSS] Releasing Stateful Functions 2.1.0 soon?
> > > > > >
> > > > > > +1 for a fast release cycle
> > > > > >
> > > > > > Seth
> > > > > >
> > > > > > On Wed, May 20, 2020 at 8:43 AM Robert Metzger <
> > rmetz...@apache.org>
> > > > > > wrote:
> > > > > >
> > > > > > > I like the idea of releasing Statefun more frequently to have
> > > faster
> > > > > > > feedback cycles!
> > > > > > >
> > > > > > > No objections for releasing 2.1.0 from my side.
> > > > > > >
> > > > > > > On Wed, May 20, 2020 at 2:22 PM Tzu-Li (Gordon) Tai <
> > > > > tzuli...@apache.org
> > > > > > >
> > > > > > > wrote:
> > > > > > >
> > > > > > > > Hi devs,
> > > > > > > >
> > > > > > > > Since Stateful Functions 2.0 was released early April,
> > > > > > > > we've been getting some good feedback from various channels,
> > > > > > > > including the Flink mailing lists, JIRA issues, as well as
> > Stack
> > > > > > Overflow
> > > > > > > > questions.
> > > > > > > >
> > > > > > > > Some of the discussions have actually translated into new
> > > features
> > > > > > > > currently being implemented into the project, such as:
> > > > > > > >
> > > > > > > >- State TTL for the state primitives in Stateful Functions
> > > (for
> > > > > both
> > > > > > > >embedded/remote functions)
> > > > > > > >- Transport for remote functions via UNIX domain sockets,
> > > which
> > > > > > would
> > > > > > > be
> > > > > > > >useful when remote functions are co-located with Flink
> > > StateFun
> > > > > > > workers
> > > > > > > >(i.e. the "sidecar" deployment mode)
> > > > > > > >
> > > > > > > >
> > > > > > > > Besides that, some critical shortcomings have already been
> > > > addressed
> > > > > > > since
> > > > > > > > the last release:
> > > > > > > >
> > > > > > > >- After upgrading to Flink 1.10.1, failure recovery in
> > > Stateful
> > > > > > > >Functions now works properly with the new scheduler.
> > > > > > > >- Support for concurrent checkpoints
> > > > > > > >
> > > > > > > >
> > > > > > > > With these ongoing threads, while it's only been just short
> of
> > 2
> > > > > months
> > > > > > > > since the last release,
> > > > > > > > we (Igal Shilman and I) have been thinking about aiming to
> > > already
> > > > > > start
> > > > > > > > the next feature release (2.1.0) soon.
> > > > > > > > This is relatively shorter than the release cycle of what the
> > > > > community
> > > > > > > is
> > > > > > > > used to in Flink (usually 3 months at least),
> > > > > > > > but we think with the StateFun project in its early phases,
> > > having
> > > > > > > smaller
> > > > > > > > and more frequent feature releases could potentially help
> drive
> > > > user
> > > > > > > > adoption.
> > > > > > > >
> > > > > > > > So, what do you think about setting feature freeze for
> StateFun
> > > > 

[jira] [Created] (FLINK-17933) TaskManager was terminated on Yarn - investigate

2020-05-25 Thread Roman Khachatryan (Jira)
Roman Khachatryan created FLINK-17933:
-

 Summary: TaskManager was terminated on Yarn - investigate
 Key: FLINK-17933
 URL: https://issues.apache.org/jira/browse/FLINK-17933
 Project: Flink
  Issue Type: Task
  Components: Deployment / YARN, Runtime / Task
Affects Versions: 1.11.0
Reporter: Roman Khachatryan
Assignee: Roman Khachatryan






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[VOTE] Release flink-shaded 11.0, release candidate #1

2020-05-25 Thread Chesnay Schepler

Hi everyone,
Please review and vote on the release candidate #1 for the version 11.0, 
as follows:

[ ] +1, Approve the release
[ ] -1, Do not approve the release (please provide specific comments)


The complete staging area is available for your review, which includes:
* JIRA release notes [1],
* the official Apache source release to be deployed to dist.apache.org 
[2], which are signed with the key with fingerprint 11D464BA [3],

* all artifacts to be deployed to the Maven Central Repository [4],
* source code tag "release-11.0-rc1" [5],
* website pull request listing the new release [6].

The vote will be open for at least 72 hours. It is adopted by majority 
approval, with at least 3 PMC affirmative votes.


Thanks,
Chesnay

[1] 
https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12347784

[2] https://dist.apache.org/repos/dist/dev/flink/flink-shaded-11.0-rc1/
[3] https://dist.apache.org/repos/dist/release/flink/KEYS
[4] https://repository.apache.org/content/repositories/orgapacheflink-1372/
[5] https://github.com/apache/flink-shaded/tree/release-11.0-rc1
[6] https://github.com/apache/flink-web/pull/340



[jira] [Created] (FLINK-17932) All hbase config Options can not work in HBaseTableSource

2020-05-25 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-17932:
--

 Summary: All hbase config Options can not work in HBaseTableSource
 Key: FLINK-17932
 URL: https://issues.apache.org/jira/browse/FLINK-17932
 Project: Flink
  Issue Type: Bug
  Components: Connectors / HBase
Affects Versions: 1.10.0, 1.11.0, 1.12.0
Reporter: Leonard Xu


All hbase config Options can not work in HBaseTableSource, because field `conf` 
in HBaseRowInputFormat it not serializable, and HBaseTableSource will fallback 
to load hbase-site.xml from classpath. I.E, many table  config Options that 
user defined in DDL is useless and HbaseTableSource only load config options 
from classpath.

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-17931) Document fromValues

2020-05-25 Thread Dawid Wysakowicz (Jira)
Dawid Wysakowicz created FLINK-17931:


 Summary: Document fromValues
 Key: FLINK-17931
 URL: https://issues.apache.org/jira/browse/FLINK-17931
 Project: Flink
  Issue Type: Sub-task
  Components: Documentation, Table SQL / API
Reporter: Dawid Wysakowicz
Assignee: Dawid Wysakowicz






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [DISCUSS] Semantics of our JIRA fields

2020-05-25 Thread Robert Metzger
Hi all,
thanks a lot for the feedback. The majority of responses are very positive
to my proposal.

I have put my proposal into our wiki:
https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=154995514

Regarding the comments so far:
@Jark: I clarified this in the wiki.

@Israel: I have not considered build changing all 3000 resolved tickets to
closed yet, but after consideration I don't think it is necessary. If
others in the community would like to change them, please speak up in this
thread :)

@Flavio: I agree that we can not ask new or infrequent users to fully
adhere to these definitions. I added a note in the Wiki.
Using the resolved state for indicating "PR available" is problematic
because there are plenty of cases where PRs are stale (and this ticket
would then appear as a "resolved"). The Apache tools are adding a link to
the PR, and some contributors are setting the ticket to "In Progress". I
don't see a problem that we need to solve here.

@Yun: Thank you for your comment. I added an example clarifying how I would
handle such a case. It is slightly different from your proposal: You
suggested using x.y.0 versions, I used "the next supported, unreleased
version", because that's how we've done it so far (and I don't want to
change things, I just want to document how the majority of the core
contributors are using JIRA).

Here are all the changes (in green, blue are just formatting changes) I
made compared to my initial proposal:
https://cwiki.apache.org/confluence/pages/diffpagesbyversion.action?pageId=154995514=4=1



On Mon, May 25, 2020 at 2:28 PM Congxian Qiu  wrote:

> @ches...@apache.org   Thanks for the confirmation
>
> Best,
> Congxian
>
>
> Zhu Zhu  于2020年5月25日周一 下午4:13写道:
>
> > This is very helpful!
> > +1
> >
> > Thanks,
> > Zhu Zhu
> >
> > Yang Wang  于2020年5月25日周一 下午4:04写道:
> >
> > > +1 from this useful proposal.
> > >
> > > This makes me clearer about "Resolve" and "Close" since I used to be
> > > confused by this two button.
> > >
> > > Best,
> > > Yang
> > >
> > > Jingsong Li  于2020年5月25日周一 下午3:10写道:
> > >
> > > > +1 for the proposal.
> > > > It makes me clearer.
> > > >
> > > > Best,
> > > > Jingsong Lee
> > > >
> > > > On Mon, May 25, 2020 at 2:51 PM Zhijiang  > > > .invalid>
> > > > wrote:
> > > >
> > > > > Thanks for launching this discussion and giving so detailed infos,
> > > > > Robert!  +1 on my side for the proposal.
> > > > >
> > > > > For "Affects Version", I previously thought it was only for the
> > already
> > > > > released versions, so it can give a reminder that the fix should
> also
> > > > pick
> > > > > into the related released branches for future minor versions.
> > > > > I saw that Jark had somehow similar concerns for this field in
> below
> > > > > replies.  Either way makes sense for me as long as we give a
> > determined
> > > > > rule in Wiki.
> > > > >
> > > > > Re Flavio' s comments, I agree that the Jira reporter can leave
> most
> > of
> > > > > the fields empty if not confirmed of them, then the respective
> > > component
> > > > > maintainer or committer can update them accordingly later.
> > > > > But the state of Jira should not be marked as "resolved" when the
> PR
> > is
> > > > > detected, that is not fitting into the resolved semantic I guess.
> If
> > > > > possible, the Jira can be updated as "in progress" automatically if
> > > > > the respective PR is ready, then it will save some time to stat
> > > progress
> > > > > of related issues during release process.
> > > > >
> > > > > Best,
> > > > > Zhijiang
> > > > > --
> > > > > From:Congxian Qiu 
> > > > > Send Time:2020年5月25日(星期一) 13:57
> > > > > To:dev@flink.apache.org 
> > > > > Subject:Re: [DISCUSS] Semantics of our JIRA fields
> > > > >
> > > > > Hi
> > > > >
> > > > > Currently, when I'm going to create an issue for Project-website.
> I'm
> > > not
> > > > > very sure what the "Affect Version/s" should be set. The problem is
> > > that
> > > > > the current Dockfile[1] in flink-web is broken because of the EOL
> of
> > > > Ubuntu
> > > > > 18.10[2], the project-web does not affect any release of Flink, it
> > does
> > > > > affect the process to build website, so what's the version should I
> > set
> > > > to?
> > > > >
> > > > > [1]
> > > > >
> > > > >
> > > >
> > >
> >
> https://github.com/apache/flink-web/blob/bc66f0f0f463ab62a22e81df7d7efd301b76a6b4/docker/Dockerfile#L17
> > > > > [2] https://wiki.ubuntu.com/Releases
> > > > >
> > > > > Best,
> > > > > Congxian
> > > > >
> > > > >
> > > > > Flavio Pompermaier  于2020年5月24日周日 下午1:23写道:
> > > > >
> > > > > > In my experience it's quite complicated for a normal reporter to
> be
> > > > able
> > > > > to
> > > > > > fill all the fields correctly (especially for new users).
> > > > > > Usually you just wanto to report a problem, remember to add a new
> > > > feature
> > > > > > or improve code/documentation but you can't really give a
> priority,
> > > > > 

[jira] [Created] (FLINK-17928) Incorrect state size reported when using unaligned checkpoints

2020-05-25 Thread Piotr Nowojski (Jira)
Piotr Nowojski created FLINK-17928:
--

 Summary: Incorrect state size reported when using unaligned 
checkpoints 
 Key: FLINK-17928
 URL: https://issues.apache.org/jira/browse/FLINK-17928
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Checkpointing
Affects Versions: 1.11.0
Reporter: Piotr Nowojski
 Fix For: 1.11.0


Even when checkpoints on HDFS are between 100-300MBs, the reported state size 
is in orders of magnitude larger with values like:
{noformat}
1GiB  1.5TiB  2.0TiB  2.1TiB  2.1TiB
148GiB  148GiB  148GiB  148GiB  148GiB  148GiB
{noformat}
it's probably because we have multiple {{Collection}}, 
and each of the individual handle returns the same value from 
{{AbstractChannelStateHandle#getStateSize}} - the full size of the spilled 
data, ignoring that only small portion of those data belong to a single input 
channel/result subpartition. In other words {{
org.apache.flink.runtime.state.AbstractChannelStateHandle#getStateSize}} should 
be taking the offsets into account and return only the size of the data that 
belong exclusively to this handle.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-17929) Fix invalid liquid expressions

2020-05-25 Thread Dawid Wysakowicz (Jira)
Dawid Wysakowicz created FLINK-17929:


 Summary: Fix invalid liquid expressions
 Key: FLINK-17929
 URL: https://issues.apache.org/jira/browse/FLINK-17929
 Project: Flink
  Issue Type: Bug
  Components: Documentation
Affects Versions: 1.11.0
Reporter: Dawid Wysakowicz
Assignee: Dawid Wysakowicz


The .ID expression in ops/deployment/docker.md should be escaped, 
otherwise it is not properly rendered.

{code}
  Generating... 
Liquid Warning: Liquid syntax error (line 331): [:dot, "."] is not a valid 
expression in "{{.ID}}" in ops/deployment/docker.md
Liquid Warning: Liquid syntax error (line 357): [:dot, "."] is not a valid 
expression in "{{.ID}}" in ops/deployment/docker.md
Liquid Warning: Liquid syntax error (line 331): [:dot, "."] is not a valid 
expression in "{{.ID}}" in ops/deployment/docker.zh.md
Liquid Warning: Liquid syntax error (line 357): [:dot, "."] is not a valid 
expression in "{{.ID}}" in ops/deployment/docker.zh.md
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-17927) Default value of "sink.partition-commit.trigger" should be "process-time"

2020-05-25 Thread Jingsong Lee (Jira)
Jingsong Lee created FLINK-17927:


 Summary: Default value of "sink.partition-commit.trigger" should 
be "process-time"
 Key: FLINK-17927
 URL: https://issues.apache.org/jira/browse/FLINK-17927
 Project: Flink
  Issue Type: Bug
  Components: Connectors / FileSystem
Reporter: Jingsong Lee
 Fix For: 1.11.0


Users are hard to figure out what is wrong when they don't have watermark. We 
can set "sink.partition-commit.trigger" to "process-time" to have better 
out-of-box experience.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-17926) Can't build flink-web docker image because of EOL of Ubuntu:18.10

2020-05-25 Thread Congxian Qiu(klion26) (Jira)
Congxian Qiu(klion26) created FLINK-17926:
-

 Summary: Can't build flink-web docker image because of EOL of 
Ubuntu:18.10
 Key: FLINK-17926
 URL: https://issues.apache.org/jira/browse/FLINK-17926
 Project: Flink
  Issue Type: Bug
  Components: Project Website
Reporter: Congxian Qiu(klion26)


Currently, the Dockerfile[1] in flink-web project is broken because of the EOL 
of Ubuntu 18.10[2], will encounter the error such as bellow when executing 
{{./run.sh}}
{code:java}
Err:3 http://security.ubuntu.com/ubuntu cosmic-security Release
  404  Not Found [IP: 91.189.88.152 80]
Ign:4 http://archive.ubuntu.com/ubuntu cosmic-updates InRelease
Ign:5 http://archive.ubuntu.com/ubuntu cosmic-backports InRelease
Err:6 http://archive.ubuntu.com/ubuntu cosmic Release
  404  Not Found [IP: 91.189.88.142 80]
Err:7 http://archive.ubuntu.com/ubuntu cosmic-updates Release
  404  Not Found [IP: 91.189.88.142 80]
Err:8 http://archive.ubuntu.com/ubuntu cosmic-backports Release
  404  Not Found [IP: 91.189.88.142 80]
Reading package lists...
{code}
The current LTS versions can be found in release website[2].

Apache Flink docker image uses fedora:28[3], so it unaffected.

As fedora does not have LTS release[4], I proposal to use Ubuntu for website 
here, and change the version from {{18.10}} to the closest LTS version {{18.04, 
tried locally, it works successfully.}}

 [1] 
[https://github.com/apache/flink-web/blob/bc66f0f0f463ab62a22e81df7d7efd301b76a6b4/docker/Dockerfile#L17]

[2] [https://wiki.ubuntu.com/Releases]

 
[3]https://github.com/apache/flink/blob/e92b2bf19bdf03ad3bae906dc5fa3781aeddb3ee/docs/docker/Dockerfile#L17

 [4] 
https://fedoraproject.org/wiki/Fedora_Release_Life_Cycle#Maintenance_Schedule



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [DISCUSS] Semantics of our JIRA fields

2020-05-25 Thread Congxian Qiu
@ches...@apache.org   Thanks for the confirmation

Best,
Congxian


Zhu Zhu  于2020年5月25日周一 下午4:13写道:

> This is very helpful!
> +1
>
> Thanks,
> Zhu Zhu
>
> Yang Wang  于2020年5月25日周一 下午4:04写道:
>
> > +1 from this useful proposal.
> >
> > This makes me clearer about "Resolve" and "Close" since I used to be
> > confused by this two button.
> >
> > Best,
> > Yang
> >
> > Jingsong Li  于2020年5月25日周一 下午3:10写道:
> >
> > > +1 for the proposal.
> > > It makes me clearer.
> > >
> > > Best,
> > > Jingsong Lee
> > >
> > > On Mon, May 25, 2020 at 2:51 PM Zhijiang  > > .invalid>
> > > wrote:
> > >
> > > > Thanks for launching this discussion and giving so detailed infos,
> > > > Robert!  +1 on my side for the proposal.
> > > >
> > > > For "Affects Version", I previously thought it was only for the
> already
> > > > released versions, so it can give a reminder that the fix should also
> > > pick
> > > > into the related released branches for future minor versions.
> > > > I saw that Jark had somehow similar concerns for this field in below
> > > > replies.  Either way makes sense for me as long as we give a
> determined
> > > > rule in Wiki.
> > > >
> > > > Re Flavio' s comments, I agree that the Jira reporter can leave most
> of
> > > > the fields empty if not confirmed of them, then the respective
> > component
> > > > maintainer or committer can update them accordingly later.
> > > > But the state of Jira should not be marked as "resolved" when the PR
> is
> > > > detected, that is not fitting into the resolved semantic I guess. If
> > > > possible, the Jira can be updated as "in progress" automatically if
> > > > the respective PR is ready, then it will save some time to stat
> > progress
> > > > of related issues during release process.
> > > >
> > > > Best,
> > > > Zhijiang
> > > > --
> > > > From:Congxian Qiu 
> > > > Send Time:2020年5月25日(星期一) 13:57
> > > > To:dev@flink.apache.org 
> > > > Subject:Re: [DISCUSS] Semantics of our JIRA fields
> > > >
> > > > Hi
> > > >
> > > > Currently, when I'm going to create an issue for Project-website. I'm
> > not
> > > > very sure what the "Affect Version/s" should be set. The problem is
> > that
> > > > the current Dockfile[1] in flink-web is broken because of the EOL of
> > > Ubuntu
> > > > 18.10[2], the project-web does not affect any release of Flink, it
> does
> > > > affect the process to build website, so what's the version should I
> set
> > > to?
> > > >
> > > > [1]
> > > >
> > > >
> > >
> >
> https://github.com/apache/flink-web/blob/bc66f0f0f463ab62a22e81df7d7efd301b76a6b4/docker/Dockerfile#L17
> > > > [2] https://wiki.ubuntu.com/Releases
> > > >
> > > > Best,
> > > > Congxian
> > > >
> > > >
> > > > Flavio Pompermaier  于2020年5月24日周日 下午1:23写道:
> > > >
> > > > > In my experience it's quite complicated for a normal reporter to be
> > > able
> > > > to
> > > > > fill all the fields correctly (especially for new users).
> > > > > Usually you just wanto to report a problem, remember to add a new
> > > feature
> > > > > or improve code/documentation but you can't really give a priority,
> > > > assign
> > > > > the correct label or decide which releases will benefit of the
> > fix/new
> > > > > feature. This is something that only core developers could decide
> > > (IMHO).
> > > > >
> > > > > My experiece says that it's better if normal users could just open
> > > > tickets
> > > > > with some default (or mark ticket as new) and leave them in such a
> > > state
> > > > > until an experienced user, one that can assign tickets, have the
> time
> > > to
> > > > > read it and immediately reject the ticket or accept it and properly
> > > > assign
> > > > > priorities, fix version, etc.
> > > > >
> > > > > With respect to resolve/close I think that a good practice could be
> > to
> > > > mark
> > > > > automatically a ticket as resolved once that a PR is detected for
> > that
> > > > > ticket, while marking it as closed should be done by the commiter
> who
> > > > merge
> > > > > the PR.
> > > > >
> > > > > Probably this process would slightly increase the work of a
> committer
> > > but
> > > > > this would make things a lot cleaner and will bring the benefit of
> > > > having a
> > > > > reliable and semantically sound JIRA state.
> > > > >
> > > > > Cheers,
> > > > > Flavio
> > > > >
> > > > > Il Dom 24 Mag 2020, 05:05 Israel Ekpo  ha
> > > scritto:
> > > > >
> > > > > > +1 for the proposal
> > > > > >
> > > > > > This will bring some consistency to the process
> > > > > >
> > > > > > Regarding Closed vs Resolved, should we go back and switch all
> the
> > > > > Resolved
> > > > > > issues to Closed so that is is not confusing to new people to the
> > > > project
> > > > > > in the future that may not have seen this discussion?
> > > > > >
> > > > > > I would recommend changing it to Closed just to be consistent
> since
> > > > that
> > > > > is
> > > > > > the majority and the new process will be using Closed 

Re: [DISCUSS] Switch to Azure Pipelines as the primary CI tool / switch off Travis

2020-05-25 Thread Yun Tang
If we still need to accept PRs for Flink-1.9/1.10, that could explain why we 
still need that command hint.
Chesnay, thanks for your explanation.

From: Chesnay Schepler 
Sent: Monday, May 25, 2020 18:17
To: dev@flink.apache.org ; Yun Tang 
Subject: Re: [DISCUSS] Switch to Azure Pipelines as the primary CI tool / 
switch off Travis

The travis bot commands must be retained so long as we accept PRs for
1.9/1.10 .

On 25/05/2020 10:50, Yun Tang wrote:
> I noticed that there still existed travis related bot commands in the github 
> PR page, and I think we should remove the command hint now.
> 
> From: Robert Metzger 
> Sent: Thursday, April 23, 2020 15:44
> To: dev 
> Subject: Re: [DISCUSS] Switch to Azure Pipelines as the primary CI tool / 
> switch off Travis
>
> FYI: I have moved the Flink PR and master builds from my personal Azure
> account to a PMC controlled account:
> https://dev.azure.com/apache-flink/apache-flink/_build
>
> On Fri, Apr 17, 2020 at 8:28 PM Robert Metzger  wrote:
>
>> Thanks a lot for bringing up this topic again.
>> The reason why I was hesitant to decommission Travis was that we were
>> still facing some issues with the Azure infrastructure that I want to
>> resolve, so that we have a strong test coverage.
>>
>> In the last few weeks, we had the following issues:
>> - unstable e2e tests (we are running the e2e tests much more frequently,
>> thus we see more failures (and discover actual bugs!))
>> - network issues (mostly around downloading maven artifacts. This is
>> solved at the cost of slower builds. I'm preparing a fix to have stable &
>> fast maven downloads)
>> - the private builds were never really stable (this is work in progress.
>> the situation is definitely better than the private Travis builds)
>> - I haven't followed the overall master stability closely before February,
>> but I have the feeling that April so far was a pretty unstable month on
>> master. Piotr is regularly reverting commits that somehow broke master. The
>> problem with unstable master is that is causes a "CI fatigue", were people
>> assume that failing builds are not worth investigating anymore, leading to
>> more instability. This is not a problem of the CI infrastructure itself,
>> but it makes me less confident switching systems :)
>>
>>
>> Unless something unexpected happens, I'm proposing to disable pull request
>> processing on Travis next week.
>>
>>
>>
>> On Fri, Apr 17, 2020 at 10:05 AM Gary Yao  wrote:
>>
>>> I am in favour of decommissioning Travis.
>>>
>>> Moreover, I wanted to use this thread to raise another issue with Travis
>>> that I
>>> have discovered recently; many of the builds running in my private Travis
>>> account are timing out in the compilation stage (i.e., compilation takes
>>> more
>>> than 50 minutes). This means that I am not able to reliably run a full
>>> build on
>>> a CI server without creating a pull request. If other developers also
>>> experience
>>> this issue, it would speak for putting more effort into making Azure
>>> Pipelines
>>> the project-wide default.
>>>
>>> Best,
>>> Gary
>>>
>>> On Thu, Mar 26, 2020 at 12:26 PM Yu Li  wrote:
>>>
 Thanks for the clarification Robert.

 Since the first step plan is to replace the travis PR runs, I checked
>>> all
 PR builds from 2020-01-01 (PR#10735-11526) [1], and below is the result:

 * Travis FAILURE: 298
 * Travis SUCCESS: 649 (68.5%)
 * Azure FAILURE: 420
 * Azure SUCCESS: 571 (57.6%)

 Since the patch for each run is equivalent for Travis and Azure, there
 seems to be slightly higher failure rate (~10%) when running in Azure.

 However, with the just-merged fix for uploading logs (FLINK-16480), I
 believe the success rate of Azure could compete with Travis now
>>> (uploading
 files contribute to 20% of the failures according to the report [2]).

 So I'm +1 to disable travis runs according to the numbers.

 Best Regards,
 Yu

 [1]

>>> https://github.com/apache/flink/pulls?q=is%3Apr+created%3A%3E%3D2020-01-01
 [2]


>>> https://dev.azure.com/rmetzger/Flink/_pipeline/analytics/stageawareoutcome?definitionId=4
 On Thu, 26 Mar 2020 at 03:28, Robert Metzger 
>>> wrote:
> Thank you for your responses.
>
> @Yu Li: In the current master, the log upload always fails, if the e2e
 job
> failed. I just merged a PR that fixes this issue [1]. The problem was
>>> not
> really the network stability, rather a problem with the interaction of
 the
> jobs in the pipeline (the e2e job did not set the right variables for
>>> the
> log upload)
> Secondly, you are looking at the report of the "flink-ci.flink"
>>> pipeline,
> where pull requests are build. Naturally, pull request builds fail all
 the
> time, because the PRs are not yet perfect.
>
> "flink-ci.flink-master" is the right pipeline to look 

[jira] [Created] (FLINK-17925) Throws unsupported exception when using metastore commit policy for filesystem table

2020-05-25 Thread Jingsong Lee (Jira)
Jingsong Lee created FLINK-17925:


 Summary: Throws unsupported exception when using metastore commit 
policy for filesystem table
 Key: FLINK-17925
 URL: https://issues.apache.org/jira/browse/FLINK-17925
 Project: Flink
  Issue Type: Bug
  Components: Connectors / FileSystem
Reporter: Jingsong Lee
 Fix For: 1.11.0


Filesystem connector has an empty implementation in 
\{{TableMetaStoreFactory}}. We should avoid user configuring this policy.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-17924) HBase connector:we can write data to HBase table, but con`t read data from HBase

2020-05-25 Thread pengnengsong (Jira)
pengnengsong created FLINK-17924:


 Summary: HBase connector:we can write data to HBase table, but 
con`t read data from HBase
 Key: FLINK-17924
 URL: https://issues.apache.org/jira/browse/FLINK-17924
 Project: Flink
  Issue Type: Bug
Reporter: pengnengsong


CREATE TABLE source(
rowkey INT,
f1 ROW,
f2 ROW
)WITH(
  'connector.type' = 'hbase',
  'connector.version' = '1.4.3',
  'connector.table-name' = 'test_source:flink',
  'connector.zookeeper.quorum' = ':2181',
  'connector.zookeeper.znode.parent' = '/test/hbase'
)

Look at the source code of the Hbase connector, execute the configure method of 
HBaseRowInputformat. Java, this.conf is always null. The default hbase 
configuration information is created, and the configuration parameters in with 
are not in effect.

private void connectToTable(){
  if(this.conf ==null){
this.conf = HBaseConfiguration.create();
   }
 ...
}




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [DISCUSS] Release flink-shaded 11.0 (and include it in 1.11.0)

2020-05-25 Thread Hequn Cheng
+1
It would be nice to have these fixes!

Best,
Hequn

On Mon, May 25, 2020 at 5:00 PM Zhijiang 
wrote:

> Thanks for driving this, Chesnay!
> +1 on my side.
>
> Best,
> Zhijiang
> --
> From:Till Rohrmann 
> Send Time:2020年5月25日(星期一) 16:28
> To:dev 
> Subject:Re: [DISCUSS] Release flink-shaded 11.0 (and include it in 1.11.0)
>
> +1 for the new flink-shaded release.
>
> Cheers,
> Till
>
> On Mon, May 25, 2020 at 9:06 AM Chesnay Schepler 
> wrote:
>
> > Hello,
> >
> > I would like to do another flink-shaded release for 1.11.0, to include a
> > zookeeper 3.4 security fix and resolve a shading issue when working with
> > Gradle.
> >
> >
> >
>
>


Re: Shall we avoid binding "unsafe ports" during random port selection?

2020-05-25 Thread Ufuk Celebi
I agree with Till. I think this should be a concern of the user configuring
the port range.

– Ufuk


On Mon, May 25, 2020 at 10:27 AM Till Rohrmann  wrote:

> Hi Weike,
>
> would it be good enough if the user did not include unsafe ranges when
> specifying `rest.bind-port`? My concern with excluding unsafe ports is that
> it adds some invisible magic which can be hard to understand for the user.
> I think over the past couple of years it has proven that auto magic often
> leads to hard to understand features.
>
> Cheers,
> Till
>
> On Sat, May 23, 2020 at 7:46 AM DONG, Weike 
> wrote:
>
> > Hi dev,
> >
> > Recently we have found that when* `rest.bind-port`* parameter is
> specified
> > as a port range (i.e. "5000-8000"), Flink may bind to some port (like
> 6000)
> > that is not allowed by Chrome (showing a "ERR_UNSAFE_PORT" message and
> > preventing users to continue accessing the website), similarly Firefox
> > blocks these unsafe port as well [1].
> >
> > When I dig further into this issue, I do believe that this restriction is
> > reasonable [2] as Flink may accidentally bind to some port that is
> > generally considered to be used by other services, posing security risks
> > and causing potential confusions to the network administrator.
> >
> > Here I propose that when Flink tries to do port selection in `
> > *NetUtils.getPortRangeFromString*`, we return an iterator that explicitly
> > skips those unsafe ports, so that those unsafe ports would not be used
> > unless users explicitly specify one in *`rest.port`* parameter.
> >
> > I would like to solicit opinions from the community on this matter,
> thanks
> > : )
> >
> > Sincerely,
> > Weike
> >
> > [1] https://www-archive.mozilla.org/projects/netlib/portbanning#portlist
> > [2]
> >
> >
> https://superuser.com/questions/188058/which-ports-are-considered-unsafe-by-chrome
> >
>


Re: [DISCUSS] FLIP-126: Unify (and separate) Watermark Assigners

2020-05-25 Thread Aljoscha Krettek
The specifics of naming diverged a bit from the FLIP during 
implementation but that should be fine. What matters in the end is the 
intention of the FLIP and that the code that is committed in the end is 
good and consistent in itself.


Best,
Aljoscha

On 24.05.20 05:12, Guanghui Zhang wrote:

Hi, @Aljoscha,
the function param currentTimestamp comment does not match the
recordTimestamp  "long extractTimestamp(T element, long recordTimestamp)"
on wiki.

Best,
Zhangguanghui

Dawid Wysakowicz  于2020年5月13日周三 上午12:28写道:


Thank you for the update and sorry again for chiming in so late...

Best,

Dawid


On 12/05/2020 18:21, Aljoscha Krettek wrote:

Yes, I am also ok with a SerializableTimestampAssigner. This only
looks a bit clumsy in the API but as a user (that uses lambdas) you
should not see this. I pushed changes for this to my branch:
https://github.com/aljoscha/flink/tree/flink-xxx-wm-generators-rebased

And yes, recordTimestamp sounds good for the TimestampAssigner. I
admit I didn't read this well enough and only saw nativeTimestamp.

Best,
Aljoscha

On 12.05.20 17:16, Dawid Wysakowicz wrote:

I have similar thoughts to @Stephan

Ad. 1 I tried something like this on your branch:

  /**
   * Adds the given {@link TimestampAssigner} to this {@link
WatermarkStrategies}. For top-level classes that implement both
Serializable and TimestampAssigner
   */
  public  & Serializable>
WatermarkStrategies withTimestampAssigner(TA timestampAssigner) {
  checkNotNull(timestampAssigner, "timestampAssigner");
  this.timestampAssigner = timestampAssigner;
  return this;
  }

 @FunctionalInterface
  public interface SerializableTimestampAssigner extends
TimestampAssigner, Serializable {
  }

   /**
* Adds the given {@link TimestampAssigner} to this {@link
WatermarkStrategies}.
   * Helper method for serializable lambdas.
   */
  public WatermarkStrategies
withTimestampAssigner(SerializableTimestampAssigner
timestampAssigner) {
  checkNotNull(timestampAssigner, "timestampAssigner");
  this.timestampAssigner = timestampAssigner;
  return this;
  }

But I understand if that's too hacky. It's just a pity that we must
enforce limitations on an interface that are not strictly necessary.

Ad 2/3

I am aware the watermark assigner/timestamp extractor can be applied
further down the graph. Originally I also wanted to suggest
sourceTimestamp and SourceTimestampAssigner, but then I realized it can
be used also after the sources as you correctly pointed out. Even if the
TimestampAssigner is used after the source there might be some
native/record timestamp in the StreamRecord, that could've been
extracted by previous assigner.

Best,

Dawid

On 12/05/2020 16:47, Stephan Ewen wrote:

@Aljoscha

About (1) could we have an interface SerializableTimestampAssigner that
simply mixes in the java.io.Serializable interface? Or will this be too
clumsy?

About (3) RecordTimeStamp seems to fit both cases (in-source-record
timestamp, in stream-record timestamp).

On Tue, May 12, 2020 at 4:12 PM Aljoscha Krettek 
wrote:


Definitely +1 to point 2) raised by Dawid. I'm not sure on points
1) and
3).

1) I can see the benefit of that but in reality most timestamp
assigners
will probably need to be Serializable. If you look at my (updated) POC
branch [1] you can see how a TimestampAssigner would be specified
on the
WatermarkStrategies helper class: [2]. The signature of this would
have
to be changed to something like:

public  & Serializable>
WatermarkStrategies withTimestampAssigner(TA timestampAssigner)

Then, however, it would not be possible for users to specify a
lambda or
anonymous inner function for the TimestampAssigner like this:

WatermarkStrategy testWmStrategy = WatermarkStrategies
  .forGenerator(new PeriodicTestWatermarkGenerator())
  .withTimestampAssigner((event, timestamp) -> event)
  .build();

3) This makes sense if we only allow WatermarkStrategies on sources,
where the previous timestamp really is the "native" timestamp.
Currently, we also allow setting watermark strategies at arbitrary
points in the graph. I'm thinking we probably should only allow
that in
sources but it's not the reality currently. I'm not against
renaming it,
just voicing those thoughts.

Best,
Aljoscha


[1]


https://github.com/aljoscha/flink/tree/flink-xxx-wm-generators-rebased

[2]



https://github.com/aljoscha/flink/blob/flink-xxx-wm-generators-rebased/flink-core/src/main/java/org/apache/flink/api/common/eventtime/WatermarkStrategies.java#L81



On 12.05.20 15:48, Stephan Ewen wrote:

+1 to all of Dawid's suggestions, makes a lot of sense to me

On Tue, May 12, 2020 at 2:32 PM Dawid Wysakowicz

Hi Aljoscha,

Sorry for adding comments during the vote, but I have some really
minor
suggestions that should not influence the voting thread imo.

1) Does it make sense to have the TimestampAssigner extend 

Re: [DISCUSS] Switch to Azure Pipelines as the primary CI tool / switch off Travis

2020-05-25 Thread Chesnay Schepler
The travis bot commands must be retained so long as we accept PRs for  
1.9/1.10 .


On 25/05/2020 10:50, Yun Tang wrote:

I noticed that there still existed travis related bot commands in the github PR 
page, and I think we should remove the command hint now.

From: Robert Metzger 
Sent: Thursday, April 23, 2020 15:44
To: dev 
Subject: Re: [DISCUSS] Switch to Azure Pipelines as the primary CI tool / 
switch off Travis

FYI: I have moved the Flink PR and master builds from my personal Azure
account to a PMC controlled account:
https://dev.azure.com/apache-flink/apache-flink/_build

On Fri, Apr 17, 2020 at 8:28 PM Robert Metzger  wrote:


Thanks a lot for bringing up this topic again.
The reason why I was hesitant to decommission Travis was that we were
still facing some issues with the Azure infrastructure that I want to
resolve, so that we have a strong test coverage.

In the last few weeks, we had the following issues:
- unstable e2e tests (we are running the e2e tests much more frequently,
thus we see more failures (and discover actual bugs!))
- network issues (mostly around downloading maven artifacts. This is
solved at the cost of slower builds. I'm preparing a fix to have stable &
fast maven downloads)
- the private builds were never really stable (this is work in progress.
the situation is definitely better than the private Travis builds)
- I haven't followed the overall master stability closely before February,
but I have the feeling that April so far was a pretty unstable month on
master. Piotr is regularly reverting commits that somehow broke master. The
problem with unstable master is that is causes a "CI fatigue", were people
assume that failing builds are not worth investigating anymore, leading to
more instability. This is not a problem of the CI infrastructure itself,
but it makes me less confident switching systems :)


Unless something unexpected happens, I'm proposing to disable pull request
processing on Travis next week.



On Fri, Apr 17, 2020 at 10:05 AM Gary Yao  wrote:


I am in favour of decommissioning Travis.

Moreover, I wanted to use this thread to raise another issue with Travis
that I
have discovered recently; many of the builds running in my private Travis
account are timing out in the compilation stage (i.e., compilation takes
more
than 50 minutes). This means that I am not able to reliably run a full
build on
a CI server without creating a pull request. If other developers also
experience
this issue, it would speak for putting more effort into making Azure
Pipelines
the project-wide default.

Best,
Gary

On Thu, Mar 26, 2020 at 12:26 PM Yu Li  wrote:


Thanks for the clarification Robert.

Since the first step plan is to replace the travis PR runs, I checked

all

PR builds from 2020-01-01 (PR#10735-11526) [1], and below is the result:

* Travis FAILURE: 298
* Travis SUCCESS: 649 (68.5%)
* Azure FAILURE: 420
* Azure SUCCESS: 571 (57.6%)

Since the patch for each run is equivalent for Travis and Azure, there
seems to be slightly higher failure rate (~10%) when running in Azure.

However, with the just-merged fix for uploading logs (FLINK-16480), I
believe the success rate of Azure could compete with Travis now

(uploading

files contribute to 20% of the failures according to the report [2]).

So I'm +1 to disable travis runs according to the numbers.

Best Regards,
Yu

[1]


https://github.com/apache/flink/pulls?q=is%3Apr+created%3A%3E%3D2020-01-01

[2]



https://dev.azure.com/rmetzger/Flink/_pipeline/analytics/stageawareoutcome?definitionId=4

On Thu, 26 Mar 2020 at 03:28, Robert Metzger 

wrote:

Thank you for your responses.

@Yu Li: In the current master, the log upload always fails, if the e2e

job

failed. I just merged a PR that fixes this issue [1]. The problem was

not

really the network stability, rather a problem with the interaction of

the

jobs in the pipeline (the e2e job did not set the right variables for

the

log upload)
Secondly, you are looking at the report of the "flink-ci.flink"

pipeline,

where pull requests are build. Naturally, pull request builds fail all

the

time, because the PRs are not yet perfect.

"flink-ci.flink-master" is the right pipeline to look at:



https://dev.azure.com/rmetzger/Flink/_pipeline/analytics/stageawareoutcome?definitionId=8=build

We have a fairly high number of failures there, because we currently

have

some issues downloading the maven artifacts [2]. I'm working already

with

Chesnay on merging a fix for that.


[1]



https://github.com/apache/flink/commit/1c86b8b9dd05615a3b2600984db738a9bf388259

[2]https://issues.apache.org/jira/browse/FLINK-16720



On Wed, Mar 25, 2020 at 4:48 PM Chesnay Schepler 
wrote:


The easiest way to disable travis for pushes is to remove all builds
from the .travis.yml with a push/pr condition.

On 25/03/2020 15:03, Robert Metzger wrote:

Thank you for the feedback so far.

Responses to the items Chesnay raised:

- by virtue of maintaining the past 

[jira] [Created] (FLINK-17923) It will throw MemoryAllocationException if rocksdb statebackend and Python UDF are used in the same slot

2020-05-25 Thread Dian Fu (Jira)
Dian Fu created FLINK-17923:
---

 Summary: It will throw MemoryAllocationException if rocksdb 
statebackend and Python UDF are used in the same slot  
 Key: FLINK-17923
 URL: https://issues.apache.org/jira/browse/FLINK-17923
 Project: Flink
  Issue Type: Bug
  Components: API / Python, Runtime / State Backends
Affects Versions: 1.10.0, 1.11.0
Reporter: Dian Fu
 Fix For: 1.11.0


For the following job:
{code}
import logging
import os
import shutil
import sys
import tempfile

from pyflink.datastream import StreamExecutionEnvironment
from pyflink.table import TableConfig, StreamTableEnvironment, DataTypes
from pyflink.table.udf import udf


def word_count():
content = "line Licensed to the Apache Software Foundation ASF under one " \
  "line or more contributor license agreements See the NOTICE file 
" \
  "line distributed with this work for additional information " \
  "line regarding copyright ownership The ASF licenses this file " \
  "to you under the Apache License Version the " \
  "License you may not use this file except in compliance " \
  "with the License"

t_config = TableConfig()
env = StreamExecutionEnvironment.get_execution_environment()
t_env = StreamTableEnvironment.create(env, t_config)

# register Results table in table environment
tmp_dir = tempfile.gettempdir()
result_path = tmp_dir + '/result'
if os.path.exists(result_path):
try:
if os.path.isfile(result_path):
os.remove(result_path)
else:
shutil.rmtree(result_path)
except OSError as e:
logging.error("Error removing directory: %s - %s.", e.filename, 
e.strerror)

logging.info("Results directory: %s", result_path)

sink_ddl = """
create table Results(
word VARCHAR,
`count` BIGINT
) with (
'connector' = 'blackhole'
)
"""
t_env.sql_update(sink_ddl)

@udf(input_types=[DataTypes.BIGINT()], result_type=DataTypes.BIGINT())
def inc(count):
return count + 1
t_env.register_function("inc", inc)

elements = [(word, 1) for word in content.split(" ")]
t_env.from_elements(elements, ["word", "count"]) \
 .group_by("word") \
 .select("word, count(1) as count") \
 .select("word, inc(count) as count") \
 .insert_into("Results")

t_env.execute("word_count")


if __name__ == '__main__':
logging.basicConfig(stream=sys.stdout, level=logging.INFO, 
format="%(message)s")

word_count()
{code}

It will throw the following exception:
{code}
Caused by: org.apache.flink.util.FlinkException: Could not restore keyed state 
backend for KeyedProcessOperator_c27dcf7b54ef6bfd6cff02ca8870b681_(1/1) from 
any of the 1 provided restore options.
at 
org.apache.flink.streaming.api.operators.BackendRestorerProcedure.createAndRestore(BackendRestorerProcedure.java:135)
at 
org.apache.flink.streaming.api.operators.StreamTaskStateInitializerImpl.keyedStatedBackend(StreamTaskStateInitializerImpl.java:317)
at 
org.apache.flink.streaming.api.operators.StreamTaskStateInitializerImpl.streamOperatorStateContext(StreamTaskStateInitializerImpl.java:144)
... 9 more
Caused by: java.io.IOException: Failed to acquire shared cache resource for 
RocksDB
at 
org.apache.flink.contrib.streaming.state.RocksDBOperationUtils.allocateSharedCachesIfConfigured(RocksDBOperationUtils.java:212)
at 
org.apache.flink.contrib.streaming.state.RocksDBStateBackend.createKeyedStateBackend(RocksDBStateBackend.java:516)
at 
org.apache.flink.streaming.api.operators.StreamTaskStateInitializerImpl.lambda$keyedStatedBackend$1(StreamTaskStateInitializerImpl.java:301)
at 
org.apache.flink.streaming.api.operators.BackendRestorerProcedure.attemptCreateAndRestore(BackendRestorerProcedure.java:142)
at 
org.apache.flink.streaming.api.operators.BackendRestorerProcedure.createAndRestore(BackendRestorerProcedure.java:121)
... 11 more
Caused by: org.apache.flink.runtime.memory.MemoryAllocationException: Could not 
created the shared memory resource of size 536870920. Not enough memory left to 
reserve from the slot's managed memory.
at 
org.apache.flink.runtime.memory.MemoryManager.lambda$getSharedMemoryResourceForManagedMemory$8(MemoryManager.java:603)
at 
org.apache.flink.runtime.memory.SharedResources.createResource(SharedResources.java:130)
at 
org.apache.flink.runtime.memory.SharedResources.getOrAllocateSharedResource(SharedResources.java:72)
at 
org.apache.flink.runtime.memory.MemoryManager.getSharedMemoryResourceForManagedMemory(MemoryManager.java:617)
at 

[jira] [Created] (FLINK-17922) Refactor generic and hive table rules for HiveCatalog

2020-05-25 Thread Jingsong Lee (Jira)
Jingsong Lee created FLINK-17922:


 Summary: Refactor generic and hive table rules for HiveCatalog
 Key: FLINK-17922
 URL: https://issues.apache.org/jira/browse/FLINK-17922
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Hive
Reporter: Jingsong Lee
 Fix For: 1.12.0


Now the underlying rules are:
 * HiveCatalog.createTable , hive table has false of {{isGeneric}} flag, 
generic table has no flag.
 * HiveCatalog.getTable , generic table has true of {{isGeneric}} flag, hive 
table has no flag.

It is very difficult to understand and easy to cause bugs.

Whether it's {{createTable}} or {{getTable}} of {{HiveCatalog}}, there should 
be two types of objects that are consistent: {{HiveCatalogTable}} and Flink 
{{CatalogTableImpl}}.

Timo: From a logical perspective, the {{HiveCatalog}} should deal with this 
additional property while storing and retrieving a {{CatalogTable}}. The 
{{isGeneric}} should not travel through the stack of regular Flink connector 
options. It is an internal property from the {{HiveCatalog}} through a custom 
{{HiveCatalogTable}} to a custom {{HiveFactory}}. It should be stored as a 
member variable in a {{HiveCatalogTable}}.

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: Please update the Requirements of PyPI

2020-05-25 Thread jincheng sun
Thanks Ray and Till,

Could you provide more information about the errors you encountered? Is it
the same error message as reported in
https://issues.apache.org/jira/browse/BEAM-8368?

Best,
Jincheng


Till Rohrmann  于2020年5月25日周一 下午4:28写道:

> Thanks for the pointer Ray. I've pulled in Jincheng who works on Flink's
> Python integration.
>
> Cheers,
> Till
>
> On Mon, May 25, 2020 at 8:41 AM Ray Chen 
> wrote:
>
> > Hi PyFlink Developers,
> >
> >
> > I got some errors from the pyarrow depended on apache-flink v1.10.1.
> >
> > It seems updating pyarrow version  greater than 0.14 will solve the
> > problem.
> >
> > The version of Flink on PyPI depends on < 0.14, so cause the error.
> >
> >
> > Best,
> > Ray Chen
> >
>


[jira] [Created] (FLINK-17921) RpcGlobalAggregateManager#updateGlobalAggregate would cause akka.timeout

2020-05-25 Thread zhangminglei (Jira)
zhangminglei created FLINK-17921:


 Summary: RpcGlobalAggregateManager#updateGlobalAggregate would 
cause akka.timeout
 Key: FLINK-17921
 URL: https://issues.apache.org/jira/browse/FLINK-17921
 Project: Flink
  Issue Type: Improvement
Reporter: zhangminglei


As described in summary, {{RpcGlobalAggregateManager#updateGlobalAggregate}} 
would cause akka.timeout. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [ANNOUNCE] Apache Flink 1.11.0, release candidate #1

2020-05-25 Thread Yang Wang
 Since the official image for 1.11 has not been release, I just want to
post how to
build the image with provided Flink binary. I think many users are running
Flink
in container environment(e.g. docker, Kubernetes, etc.) and hope this could
help
them with the testing.



git clone 'https://github.com/apache/flink-docker.git'
cd flink-docker/ && git checkout dev-master

export image_name="flink:1.11.0-rc1-scala_2.11"
./add-custom.sh -u
https://dist.apache.org/repos/dist/dev/flink/flink-1.11.0-rc1/flink-1.11.0-bin-scala_2.11.tgz
-n ${image_name}
docker build --no-cache --network="host" -t ${image_name}
dev/${image_name}-debian

docker push ${image_name}



Best,
Yang

Zhijiang  于2020年5月25日周一 下午12:09写道:

> Hi all,
>
> Apache Flink-1.11.0-RC1 has been created. It has all the artifacts that we
> would typically have for a release.
>
> This preview-only RC is created only to drive the current testing efforts,
> and no official vote will take place. It includes the following:
>
>* The preview source release and binary convenience releases [1], which
> are signed with the key with fingerprint
> 2DA85B93244FDFA19A6244500653C0A2CEA00D0E [2],
>* All artifacts that would normally be deployed to the Maven Central
> Repository [3]
>
> To test with these artifacts, you can create a settings.xml file with the
> content shown below [4]. This settings file can be referenced in your maven
> commands
> via --settings /path/to/settings.xml. This is useful for creating a
> quickstart project based on the staged release and also for building
> against the staged jars.
>
> Happy testing!
>
> Best,
> Zhijiang
>
> [1] https://dist.apache.org/repos/dist/dev/flink/flink-1.11.0-rc1/
> [2] https://dist.apache.org/repos/dist/release/flink/KEYS
> [3]
> https://repository.apache.org/content/repositories/orgapacheflink-1370/
> [4]
> 
> 
>  flink-1.11.0
> 
> 
> 
> flink-1.11.0
> 
>   
> flink-1.11.0
> 
> https://repository.apache.org/content/repositories/orgapacheflink-1370/
> 
>  
>  
>archetype
>
> https://repository.apache.org/content/repositories/orgapacheflink-1370/
> 
>  
>  
> 
> 
> 


[jira] [Created] (FLINK-17920) Add the Python example of Time-windowed join in Table API doc

2020-05-25 Thread Huang Xingbo (Jira)
Huang Xingbo created FLINK-17920:


 Summary: Add the Python example of Time-windowed join in Table API 
doc
 Key: FLINK-17920
 URL: https://issues.apache.org/jira/browse/FLINK-17920
 Project: Flink
  Issue Type: Improvement
  Components: API / Python, Documentation
Reporter: Huang Xingbo
 Fix For: 1.11.0, 1.10.2


We have supported Time-windowed join in Python Table API since 1.10. So we 
should add the corresponding example in Table API doc.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [DISCUSS] Release flink-shaded 11.0 (and include it in 1.11.0)

2020-05-25 Thread Zhijiang
Thanks for driving this, Chesnay! 
+1 on my side.

Best,
Zhijiang
--
From:Till Rohrmann 
Send Time:2020年5月25日(星期一) 16:28
To:dev 
Subject:Re: [DISCUSS] Release flink-shaded 11.0 (and include it in 1.11.0)

+1 for the new flink-shaded release.

Cheers,
Till

On Mon, May 25, 2020 at 9:06 AM Chesnay Schepler  wrote:

> Hello,
>
> I would like to do another flink-shaded release for 1.11.0, to include a
> zookeeper 3.4 security fix and resolve a shading issue when working with
> Gradle.
>
>
>



[jira] [Created] (FLINK-17919) KafkaConsumerThread should add ratelimiter functionality as well

2020-05-25 Thread zhangminglei (Jira)
zhangminglei created FLINK-17919:


 Summary: KafkaConsumerThread should add ratelimiter functionality 
as well
 Key: FLINK-17919
 URL: https://issues.apache.org/jira/browse/FLINK-17919
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / Kafka
Reporter: zhangminglei


Currently, {{KafkaConsumerThread}} within {{flink-connector-kafka-09}} has the 
ability of rateLimiter. However, under {{flink-connector-kafka}} does not own 
it. I would suggest we can add it as well if we could.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [DISCUSS] Switch to Azure Pipelines as the primary CI tool / switch off Travis

2020-05-25 Thread Yun Tang

I noticed that there still existed travis related bot commands in the github PR 
page, and I think we should remove the command hint now.

From: Robert Metzger 
Sent: Thursday, April 23, 2020 15:44
To: dev 
Subject: Re: [DISCUSS] Switch to Azure Pipelines as the primary CI tool / 
switch off Travis

FYI: I have moved the Flink PR and master builds from my personal Azure
account to a PMC controlled account:
https://dev.azure.com/apache-flink/apache-flink/_build

On Fri, Apr 17, 2020 at 8:28 PM Robert Metzger  wrote:

> Thanks a lot for bringing up this topic again.
> The reason why I was hesitant to decommission Travis was that we were
> still facing some issues with the Azure infrastructure that I want to
> resolve, so that we have a strong test coverage.
>
> In the last few weeks, we had the following issues:
> - unstable e2e tests (we are running the e2e tests much more frequently,
> thus we see more failures (and discover actual bugs!))
> - network issues (mostly around downloading maven artifacts. This is
> solved at the cost of slower builds. I'm preparing a fix to have stable &
> fast maven downloads)
> - the private builds were never really stable (this is work in progress.
> the situation is definitely better than the private Travis builds)
> - I haven't followed the overall master stability closely before February,
> but I have the feeling that April so far was a pretty unstable month on
> master. Piotr is regularly reverting commits that somehow broke master. The
> problem with unstable master is that is causes a "CI fatigue", were people
> assume that failing builds are not worth investigating anymore, leading to
> more instability. This is not a problem of the CI infrastructure itself,
> but it makes me less confident switching systems :)
>
>
> Unless something unexpected happens, I'm proposing to disable pull request
> processing on Travis next week.
>
>
>
> On Fri, Apr 17, 2020 at 10:05 AM Gary Yao  wrote:
>
>> I am in favour of decommissioning Travis.
>>
>> Moreover, I wanted to use this thread to raise another issue with Travis
>> that I
>> have discovered recently; many of the builds running in my private Travis
>> account are timing out in the compilation stage (i.e., compilation takes
>> more
>> than 50 minutes). This means that I am not able to reliably run a full
>> build on
>> a CI server without creating a pull request. If other developers also
>> experience
>> this issue, it would speak for putting more effort into making Azure
>> Pipelines
>> the project-wide default.
>>
>> Best,
>> Gary
>>
>> On Thu, Mar 26, 2020 at 12:26 PM Yu Li  wrote:
>>
>> > Thanks for the clarification Robert.
>> >
>> > Since the first step plan is to replace the travis PR runs, I checked
>> all
>> > PR builds from 2020-01-01 (PR#10735-11526) [1], and below is the result:
>> >
>> > * Travis FAILURE: 298
>> > * Travis SUCCESS: 649 (68.5%)
>> > * Azure FAILURE: 420
>> > * Azure SUCCESS: 571 (57.6%)
>> >
>> > Since the patch for each run is equivalent for Travis and Azure, there
>> > seems to be slightly higher failure rate (~10%) when running in Azure.
>> >
>> > However, with the just-merged fix for uploading logs (FLINK-16480), I
>> > believe the success rate of Azure could compete with Travis now
>> (uploading
>> > files contribute to 20% of the failures according to the report [2]).
>> >
>> > So I'm +1 to disable travis runs according to the numbers.
>> >
>> > Best Regards,
>> > Yu
>> >
>> > [1]
>> >
>> https://github.com/apache/flink/pulls?q=is%3Apr+created%3A%3E%3D2020-01-01
>> > [2]
>> >
>> >
>> https://dev.azure.com/rmetzger/Flink/_pipeline/analytics/stageawareoutcome?definitionId=4
>> >
>> > On Thu, 26 Mar 2020 at 03:28, Robert Metzger 
>> wrote:
>> >
>> > > Thank you for your responses.
>> > >
>> > > @Yu Li: In the current master, the log upload always fails, if the e2e
>> > job
>> > > failed. I just merged a PR that fixes this issue [1]. The problem was
>> not
>> > > really the network stability, rather a problem with the interaction of
>> > the
>> > > jobs in the pipeline (the e2e job did not set the right variables for
>> the
>> > > log upload)
>> > > Secondly, you are looking at the report of the "flink-ci.flink"
>> pipeline,
>> > > where pull requests are build. Naturally, pull request builds fail all
>> > the
>> > > time, because the PRs are not yet perfect.
>> > >
>> > > "flink-ci.flink-master" is the right pipeline to look at:
>> > >
>> > >
>> >
>> https://dev.azure.com/rmetzger/Flink/_pipeline/analytics/stageawareoutcome?definitionId=8=build
>> > > We have a fairly high number of failures there, because we currently
>> have
>> > > some issues downloading the maven artifacts [2]. I'm working already
>> with
>> > > Chesnay on merging a fix for that.
>> > >
>> > >
>> > > [1]
>> > >
>> > >
>> >
>> https://github.com/apache/flink/commit/1c86b8b9dd05615a3b2600984db738a9bf388259
>> > > [2]https://issues.apache.org/jira/browse/FLINK-16720
>> > >
>> > >
>> > >
>> > > 

Re: [DISCUSS] Release flink-shaded 11.0 (and include it in 1.11.0)

2020-05-25 Thread Till Rohrmann
+1 for the new flink-shaded release.

Cheers,
Till

On Mon, May 25, 2020 at 9:06 AM Chesnay Schepler  wrote:

> Hello,
>
> I would like to do another flink-shaded release for 1.11.0, to include a
> zookeeper 3.4 security fix and resolve a shading issue when working with
> Gradle.
>
>
>


Re: Please update the Requirements of PyPI

2020-05-25 Thread Till Rohrmann
Thanks for the pointer Ray. I've pulled in Jincheng who works on Flink's
Python integration.

Cheers,
Till

On Mon, May 25, 2020 at 8:41 AM Ray Chen 
wrote:

> Hi PyFlink Developers,
>
>
> I got some errors from the pyarrow depended on apache-flink v1.10.1.
>
> It seems updating pyarrow version  greater than 0.14 will solve the
> problem.
>
> The version of Flink on PyPI depends on < 0.14, so cause the error.
>
>
> Best,
> Ray Chen
>


Re: Shall we avoid binding "unsafe ports" during random port selection?

2020-05-25 Thread Till Rohrmann
Hi Weike,

would it be good enough if the user did not include unsafe ranges when
specifying `rest.bind-port`? My concern with excluding unsafe ports is that
it adds some invisible magic which can be hard to understand for the user.
I think over the past couple of years it has proven that auto magic often
leads to hard to understand features.

Cheers,
Till

On Sat, May 23, 2020 at 7:46 AM DONG, Weike  wrote:

> Hi dev,
>
> Recently we have found that when* `rest.bind-port`* parameter is specified
> as a port range (i.e. "5000-8000"), Flink may bind to some port (like 6000)
> that is not allowed by Chrome (showing a "ERR_UNSAFE_PORT" message and
> preventing users to continue accessing the website), similarly Firefox
> blocks these unsafe port as well [1].
>
> When I dig further into this issue, I do believe that this restriction is
> reasonable [2] as Flink may accidentally bind to some port that is
> generally considered to be used by other services, posing security risks
> and causing potential confusions to the network administrator.
>
> Here I propose that when Flink tries to do port selection in `
> *NetUtils.getPortRangeFromString*`, we return an iterator that explicitly
> skips those unsafe ports, so that those unsafe ports would not be used
> unless users explicitly specify one in *`rest.port`* parameter.
>
> I would like to solicit opinions from the community on this matter, thanks
> : )
>
> Sincerely,
> Weike
>
> [1] https://www-archive.mozilla.org/projects/netlib/portbanning#portlist
> [2]
>
> https://superuser.com/questions/188058/which-ports-are-considered-unsafe-by-chrome
>


Re: java.lang.NoSuchMethodError while writing to Kafka from Flink

2020-05-25 Thread Chesnay Schepler

Please double-check that your distribution and application jar were
built against the same Flink version.

This looks related to a binary-compatibility issues reporter in
FLINK-13586 .



Re: [DISCUSS] Semantics of our JIRA fields

2020-05-25 Thread Zhu Zhu
This is very helpful!
+1

Thanks,
Zhu Zhu

Yang Wang  于2020年5月25日周一 下午4:04写道:

> +1 from this useful proposal.
>
> This makes me clearer about "Resolve" and "Close" since I used to be
> confused by this two button.
>
> Best,
> Yang
>
> Jingsong Li  于2020年5月25日周一 下午3:10写道:
>
> > +1 for the proposal.
> > It makes me clearer.
> >
> > Best,
> > Jingsong Lee
> >
> > On Mon, May 25, 2020 at 2:51 PM Zhijiang  > .invalid>
> > wrote:
> >
> > > Thanks for launching this discussion and giving so detailed infos,
> > > Robert!  +1 on my side for the proposal.
> > >
> > > For "Affects Version", I previously thought it was only for the already
> > > released versions, so it can give a reminder that the fix should also
> > pick
> > > into the related released branches for future minor versions.
> > > I saw that Jark had somehow similar concerns for this field in below
> > > replies.  Either way makes sense for me as long as we give a determined
> > > rule in Wiki.
> > >
> > > Re Flavio' s comments, I agree that the Jira reporter can leave most of
> > > the fields empty if not confirmed of them, then the respective
> component
> > > maintainer or committer can update them accordingly later.
> > > But the state of Jira should not be marked as "resolved" when the PR is
> > > detected, that is not fitting into the resolved semantic I guess. If
> > > possible, the Jira can be updated as "in progress" automatically if
> > > the respective PR is ready, then it will save some time to stat
> progress
> > > of related issues during release process.
> > >
> > > Best,
> > > Zhijiang
> > > --
> > > From:Congxian Qiu 
> > > Send Time:2020年5月25日(星期一) 13:57
> > > To:dev@flink.apache.org 
> > > Subject:Re: [DISCUSS] Semantics of our JIRA fields
> > >
> > > Hi
> > >
> > > Currently, when I'm going to create an issue for Project-website. I'm
> not
> > > very sure what the "Affect Version/s" should be set. The problem is
> that
> > > the current Dockfile[1] in flink-web is broken because of the EOL of
> > Ubuntu
> > > 18.10[2], the project-web does not affect any release of Flink, it does
> > > affect the process to build website, so what's the version should I set
> > to?
> > >
> > > [1]
> > >
> > >
> >
> https://github.com/apache/flink-web/blob/bc66f0f0f463ab62a22e81df7d7efd301b76a6b4/docker/Dockerfile#L17
> > > [2] https://wiki.ubuntu.com/Releases
> > >
> > > Best,
> > > Congxian
> > >
> > >
> > > Flavio Pompermaier  于2020年5月24日周日 下午1:23写道:
> > >
> > > > In my experience it's quite complicated for a normal reporter to be
> > able
> > > to
> > > > fill all the fields correctly (especially for new users).
> > > > Usually you just wanto to report a problem, remember to add a new
> > feature
> > > > or improve code/documentation but you can't really give a priority,
> > > assign
> > > > the correct label or decide which releases will benefit of the
> fix/new
> > > > feature. This is something that only core developers could decide
> > (IMHO).
> > > >
> > > > My experiece says that it's better if normal users could just open
> > > tickets
> > > > with some default (or mark ticket as new) and leave them in such a
> > state
> > > > until an experienced user, one that can assign tickets, have the time
> > to
> > > > read it and immediately reject the ticket or accept it and properly
> > > assign
> > > > priorities, fix version, etc.
> > > >
> > > > With respect to resolve/close I think that a good practice could be
> to
> > > mark
> > > > automatically a ticket as resolved once that a PR is detected for
> that
> > > > ticket, while marking it as closed should be done by the commiter who
> > > merge
> > > > the PR.
> > > >
> > > > Probably this process would slightly increase the work of a committer
> > but
> > > > this would make things a lot cleaner and will bring the benefit of
> > > having a
> > > > reliable and semantically sound JIRA state.
> > > >
> > > > Cheers,
> > > > Flavio
> > > >
> > > > Il Dom 24 Mag 2020, 05:05 Israel Ekpo  ha
> > scritto:
> > > >
> > > > > +1 for the proposal
> > > > >
> > > > > This will bring some consistency to the process
> > > > >
> > > > > Regarding Closed vs Resolved, should we go back and switch all the
> > > > Resolved
> > > > > issues to Closed so that is is not confusing to new people to the
> > > project
> > > > > in the future that may not have seen this discussion?
> > > > >
> > > > > I would recommend changing it to Closed just to be consistent since
> > > that
> > > > is
> > > > > the majority and the new process will be using Closed going forward
> > > > >
> > > > > Those are my thoughts for now
> > > > >
> > > > > On Sat, May 23, 2020 at 7:48 AM Congxian Qiu <
> qcx978132...@gmail.com
> > >
> > > > > wrote:
> > > > >
> > > > > > +1 for the proposal. It's good to have a unified description and
> > > write
> > > > it
> > > > > > down in the wiki, so that every contributor has the same
> > > understanding
> > > > of
> > > > > > all the 

Re: [DISCUSS] Semantics of our JIRA fields

2020-05-25 Thread Yang Wang
+1 from this useful proposal.

This makes me clearer about "Resolve" and "Close" since I used to be
confused by this two button.

Best,
Yang

Jingsong Li  于2020年5月25日周一 下午3:10写道:

> +1 for the proposal.
> It makes me clearer.
>
> Best,
> Jingsong Lee
>
> On Mon, May 25, 2020 at 2:51 PM Zhijiang  .invalid>
> wrote:
>
> > Thanks for launching this discussion and giving so detailed infos,
> > Robert!  +1 on my side for the proposal.
> >
> > For "Affects Version", I previously thought it was only for the already
> > released versions, so it can give a reminder that the fix should also
> pick
> > into the related released branches for future minor versions.
> > I saw that Jark had somehow similar concerns for this field in below
> > replies.  Either way makes sense for me as long as we give a determined
> > rule in Wiki.
> >
> > Re Flavio' s comments, I agree that the Jira reporter can leave most of
> > the fields empty if not confirmed of them, then the respective component
> > maintainer or committer can update them accordingly later.
> > But the state of Jira should not be marked as "resolved" when the PR is
> > detected, that is not fitting into the resolved semantic I guess. If
> > possible, the Jira can be updated as "in progress" automatically if
> > the respective PR is ready, then it will save some time to stat progress
> > of related issues during release process.
> >
> > Best,
> > Zhijiang
> > --
> > From:Congxian Qiu 
> > Send Time:2020年5月25日(星期一) 13:57
> > To:dev@flink.apache.org 
> > Subject:Re: [DISCUSS] Semantics of our JIRA fields
> >
> > Hi
> >
> > Currently, when I'm going to create an issue for Project-website. I'm not
> > very sure what the "Affect Version/s" should be set. The problem is that
> > the current Dockfile[1] in flink-web is broken because of the EOL of
> Ubuntu
> > 18.10[2], the project-web does not affect any release of Flink, it does
> > affect the process to build website, so what's the version should I set
> to?
> >
> > [1]
> >
> >
> https://github.com/apache/flink-web/blob/bc66f0f0f463ab62a22e81df7d7efd301b76a6b4/docker/Dockerfile#L17
> > [2] https://wiki.ubuntu.com/Releases
> >
> > Best,
> > Congxian
> >
> >
> > Flavio Pompermaier  于2020年5月24日周日 下午1:23写道:
> >
> > > In my experience it's quite complicated for a normal reporter to be
> able
> > to
> > > fill all the fields correctly (especially for new users).
> > > Usually you just wanto to report a problem, remember to add a new
> feature
> > > or improve code/documentation but you can't really give a priority,
> > assign
> > > the correct label or decide which releases will benefit of the fix/new
> > > feature. This is something that only core developers could decide
> (IMHO).
> > >
> > > My experiece says that it's better if normal users could just open
> > tickets
> > > with some default (or mark ticket as new) and leave them in such a
> state
> > > until an experienced user, one that can assign tickets, have the time
> to
> > > read it and immediately reject the ticket or accept it and properly
> > assign
> > > priorities, fix version, etc.
> > >
> > > With respect to resolve/close I think that a good practice could be to
> > mark
> > > automatically a ticket as resolved once that a PR is detected for that
> > > ticket, while marking it as closed should be done by the commiter who
> > merge
> > > the PR.
> > >
> > > Probably this process would slightly increase the work of a committer
> but
> > > this would make things a lot cleaner and will bring the benefit of
> > having a
> > > reliable and semantically sound JIRA state.
> > >
> > > Cheers,
> > > Flavio
> > >
> > > Il Dom 24 Mag 2020, 05:05 Israel Ekpo  ha
> scritto:
> > >
> > > > +1 for the proposal
> > > >
> > > > This will bring some consistency to the process
> > > >
> > > > Regarding Closed vs Resolved, should we go back and switch all the
> > > Resolved
> > > > issues to Closed so that is is not confusing to new people to the
> > project
> > > > in the future that may not have seen this discussion?
> > > >
> > > > I would recommend changing it to Closed just to be consistent since
> > that
> > > is
> > > > the majority and the new process will be using Closed going forward
> > > >
> > > > Those are my thoughts for now
> > > >
> > > > On Sat, May 23, 2020 at 7:48 AM Congxian Qiu  >
> > > > wrote:
> > > >
> > > > > +1 for the proposal. It's good to have a unified description and
> > write
> > > it
> > > > > down in the wiki, so that every contributor has the same
> > understanding
> > > of
> > > > > all the fields.
> > > > >
> > > > > Best,
> > > > > Congxian
> > > > >
> > > > >
> > > > > Till Rohrmann  于2020年5月23日周六 上午12:04写道:
> > > > >
> > > > > > Thanks for drafting this proposal Robert. +1 for the proposal.
> > > > > >
> > > > > > Cheers,
> > > > > > Till
> > > > > >
> > > > > > On Fri, May 22, 2020 at 5:39 PM Leonard Xu 
> > > wrote:
> > > > > >
> > > > > > > Thanks bringing up this topic 

[jira] [Created] (FLINK-17918) Jobs with two input operators are loosing data in unaligned checkpoints

2020-05-25 Thread Piotr Nowojski (Jira)
Piotr Nowojski created FLINK-17918:
--

 Summary: Jobs with two input operators are loosing data in 
unaligned checkpoints
 Key: FLINK-17918
 URL: https://issues.apache.org/jira/browse/FLINK-17918
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Network
Affects Versions: 1.11.0
Reporter: Piotr Nowojski
 Fix For: 1.11.0


After trying to enable unaligned checkpoints by default, a lot of Blink 
streaming SQL/Table API tests containing joins or set operations are throwing 
errors that are indicating we are loosing some data (full records, without 
deserialisation errors). Example errors:

{noformat}
[ERROR] Failures: 
[ERROR]   JoinITCase.testFullJoinWithEqualPk:775 expected: but was:
[ERROR]   JoinITCase.testStreamJoinWithSameRecord:391 expected: but was:
[ERROR]   SemiAntiJoinStreamITCase.testAntiJoin:352 expected:<0> but was:<1>
[ERROR]   SetOperatorsITCase.testIntersect:55 expected: but was:
[ERROR]   JoinITCase.testJoinPushThroughJoin:1272 expected: but was:
{noformat}
 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[ANNOUNCE] Weekly Community Update 2020/21

2020-05-25 Thread Konstantin Knauf
Dear community,

happy to share this week's community update!

Flink Development
==

* [releases]  Zhijiang has published a first release candidate for Apache
Flink 1.11.0. This is a "preview-only" release candidate to facilitate
current testing efforts. There will not be a vote on this RC. [1]

* [releases] Gordon proposes to release Apache Flink Stateful Function
2.1.0 soon. This would mean a faster release cycle compared to the main
project, which seems to make sense given the early state of the project.
The proposal received a lot positive feedback and the feature freeze date
was set for this Wednesday (27th of May). [2]

* [deployment] Klou proposes to remove support for shipping nested JARs
during job submission, a so far hidden feature of Apache Flink from the
early days of the project. [3]

* [security] Weike Dong proposes to exclude "unsafe ports", i.e. ports
considered to be unsafe by major browsers, as candidates for the Flink REST
API, if no port was specified explicitly. [4]

* [development process] Robert has started a thread to clarify the
semantics and usage of JIRA fields in the Apache Flink project. If you are
unsure about some fields, check out this thread and join the conversation.
[5]

[1]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/ANNOUNCE-Apache-Flink-1-11-0-release-candidate-1-tp41840.html
[2]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Releasing-Stateful-Functions-2-1-0-soon-tp41734.html
[3]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Remove-dependency-shipping-through-nested-jars-during-job-submission-tp41711.html
[4]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/Shall-we-avoid-binding-unsafe-ports-during-random-port-selection-tp41821.html
[5]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Semantics-of-our-JIRA-fields-tp41786p41806.html

Notable Bugs
==

* [FLINK-17189] [1.10.1] HiveCatalog does not support processing time
columns. [6]
* [FLINK-17800] [1.10.1] There might be a state loss issue when using the
Rocksdb with optimizeForPointLookup. The community is still investigating
the bug report. [7]

[6] https://issues.apache.org/jira/browse/FLINK-17189
[7] https://issues.apache.org/jira/browse/FLINK-17800

Events, Blog Posts, Misc
===

* Ankit Jhalaria of GoDaddy has published a blog post about their
Flink-based streaming data platform on the Ververica Blog [8].
* If you have not heard about Lyft's streaming platform before, here is
another great talk by Sherin Thomas and blog on InfoQ [9].

[8]
https://www.ververica.com/blog/how-godaddy-uses-flink-to-run-real-time-streaming-pipelines
[9] https://www.infoq.com/presentations/ml-streaming-lyft/

Cheers,

Konstantin

-- 

Konstantin Knauf

https://twitter.com/snntrable

https://github.com/knaufk


Re: [DISCUSS] Semantics of our JIRA fields

2020-05-25 Thread Jingsong Li
+1 for the proposal.
It makes me clearer.

Best,
Jingsong Lee

On Mon, May 25, 2020 at 2:51 PM Zhijiang 
wrote:

> Thanks for launching this discussion and giving so detailed infos,
> Robert!  +1 on my side for the proposal.
>
> For "Affects Version", I previously thought it was only for the already
> released versions, so it can give a reminder that the fix should also pick
> into the related released branches for future minor versions.
> I saw that Jark had somehow similar concerns for this field in below
> replies.  Either way makes sense for me as long as we give a determined
> rule in Wiki.
>
> Re Flavio' s comments, I agree that the Jira reporter can leave most of
> the fields empty if not confirmed of them, then the respective component
> maintainer or committer can update them accordingly later.
> But the state of Jira should not be marked as "resolved" when the PR is
> detected, that is not fitting into the resolved semantic I guess. If
> possible, the Jira can be updated as "in progress" automatically if
> the respective PR is ready, then it will save some time to stat progress
> of related issues during release process.
>
> Best,
> Zhijiang
> --
> From:Congxian Qiu 
> Send Time:2020年5月25日(星期一) 13:57
> To:dev@flink.apache.org 
> Subject:Re: [DISCUSS] Semantics of our JIRA fields
>
> Hi
>
> Currently, when I'm going to create an issue for Project-website. I'm not
> very sure what the "Affect Version/s" should be set. The problem is that
> the current Dockfile[1] in flink-web is broken because of the EOL of Ubuntu
> 18.10[2], the project-web does not affect any release of Flink, it does
> affect the process to build website, so what's the version should I set to?
>
> [1]
>
> https://github.com/apache/flink-web/blob/bc66f0f0f463ab62a22e81df7d7efd301b76a6b4/docker/Dockerfile#L17
> [2] https://wiki.ubuntu.com/Releases
>
> Best,
> Congxian
>
>
> Flavio Pompermaier  于2020年5月24日周日 下午1:23写道:
>
> > In my experience it's quite complicated for a normal reporter to be able
> to
> > fill all the fields correctly (especially for new users).
> > Usually you just wanto to report a problem, remember to add a new feature
> > or improve code/documentation but you can't really give a priority,
> assign
> > the correct label or decide which releases will benefit of the fix/new
> > feature. This is something that only core developers could decide (IMHO).
> >
> > My experiece says that it's better if normal users could just open
> tickets
> > with some default (or mark ticket as new) and leave them in such a state
> > until an experienced user, one that can assign tickets, have the time to
> > read it and immediately reject the ticket or accept it and properly
> assign
> > priorities, fix version, etc.
> >
> > With respect to resolve/close I think that a good practice could be to
> mark
> > automatically a ticket as resolved once that a PR is detected for that
> > ticket, while marking it as closed should be done by the commiter who
> merge
> > the PR.
> >
> > Probably this process would slightly increase the work of a committer but
> > this would make things a lot cleaner and will bring the benefit of
> having a
> > reliable and semantically sound JIRA state.
> >
> > Cheers,
> > Flavio
> >
> > Il Dom 24 Mag 2020, 05:05 Israel Ekpo  ha scritto:
> >
> > > +1 for the proposal
> > >
> > > This will bring some consistency to the process
> > >
> > > Regarding Closed vs Resolved, should we go back and switch all the
> > Resolved
> > > issues to Closed so that is is not confusing to new people to the
> project
> > > in the future that may not have seen this discussion?
> > >
> > > I would recommend changing it to Closed just to be consistent since
> that
> > is
> > > the majority and the new process will be using Closed going forward
> > >
> > > Those are my thoughts for now
> > >
> > > On Sat, May 23, 2020 at 7:48 AM Congxian Qiu 
> > > wrote:
> > >
> > > > +1 for the proposal. It's good to have a unified description and
> write
> > it
> > > > down in the wiki, so that every contributor has the same
> understanding
> > of
> > > > all the fields.
> > > >
> > > > Best,
> > > > Congxian
> > > >
> > > >
> > > > Till Rohrmann  于2020年5月23日周六 上午12:04写道:
> > > >
> > > > > Thanks for drafting this proposal Robert. +1 for the proposal.
> > > > >
> > > > > Cheers,
> > > > > Till
> > > > >
> > > > > On Fri, May 22, 2020 at 5:39 PM Leonard Xu 
> > wrote:
> > > > >
> > > > > > Thanks bringing up this topic @Robert,  +1 to the proposal.
> > > > > >
> > > > > > It clarifies the JIRA fields well and should be a rule to follow.
> > > > > >
> > > > > > Best,
> > > > > > Leonard Xu
> > > > > > > 在 2020年5月22日,20:24,Aljoscha Krettek  写道:
> > > > > > >
> > > > > > > +1 That's also how I think of the semantics of the fields.
> > > > > > >
> > > > > > > Aljoscha
> > > > > > >
> > > > > > > On 22.05.20 08:07, Robert Metzger wrote:
> > > > > > >> Hi all,
> > > > > > >> I have the feeling 

Re: [DISCUSS] FLIP-84 Feedback Summary

2020-05-25 Thread Timo Walther

+1

Personally, I would declare it as `collect(): ClosableIterator` to 
avoid an additional class in the API and reuse existing Flink utils.


Regards,
Timo

On 17.05.20 10:21, Jingsong Li wrote:

+1

Best,
Jingsong Lee

On Sun, May 17, 2020 at 3:42 PM Kurt Young  wrote:


+1 from my side.

Best,
Kurt


On Sun, May 17, 2020 at 12:41 PM godfrey he  wrote:


hi everyone,

I would like to bring up another topic about the return value of
TableResult#collect() method.
Currently, the return type is `Iterator`, we meet some problems when
implementing FLINK-14807[1].

In current design, the sink operator has a buffer pool which buffers the
data from upstream,
and waits the client to consume the data. The client will pull the data
when `Iterator#next()` method is called.
If the client submits a select job, consumes a part of data and exits.

The

job will not be finished.
This will cause resource leak. We can't require the client must consume

all

data. for unbounded stream job, it's also impossible.
Currently, users can also cancel the job via
`TableResult.getJobClient().get().cancel()` method.
But this approach is not intuitive and convenient.

So, I want to change the return type from `Iterator` to
`CloseableRowIterator`,
the new method likes like:

public interface TableResult {

   CloseableRowIterator collect();

}

public interface CloseableRowIterator extends Iterator,

AutoCloseable

{

}

Prefixing the name with "Closeable" is intended to remind the users that
this iterator should be closed,
users can conveniently use try-with-resources statement to close the
resources.
The resource leak problem is still there if users do not close the

iterator

or cancel the job through job client,
  we just provide an easier way for users to avoid this.

I also notice that there is a `CloseableIterator` interface in
`org.apache.flink.util` package.
But I still tend to introduce `CloseableRowIterator`. My point of view

is:

1) `CloseableIterator` is in a util package, not a public interface.
2) `CloseableRowIterator` is more convenient, users do not need to define
generic type ``.

What do you think?

Best,
Godfrey


[1] https://issues.apache.org/jira/browse/FLINK-14807


Fabian Hueske  于2020年5月7日周四 下午3:59写道:


Thanks for the update Godfrey!

+1 to this approach.

Since there can be only one primary key, I'd also be fine to just use
`PRI` even if it is composite, but `PRI(f0, f5)` might be more

convenient

for users.

Thanks, Fabian

Am Do., 7. Mai 2020 um 09:31 Uhr schrieb godfrey he <

godfre...@gmail.com

:


Hi fabian,
Thanks for you suggestions.

Agree with you that `UNQ(f2, f3)` is more clear.

A table can have only ONE primary key,
this primary key can consist of single or multiple columns. [1]
if primary key consists of single column,
we can simply use `PRI` (or `PRI(xx)`) to represent it.
if primary key have multiple columns,
we should use `PRI(xx, yy, ...)` to represent it.

A table may have multiple unique keys,
each unique key can consist of single or multiple columns. [2]
if there is only one unique key and this unique key has only single
column,
we can simply use `UNQ` (or `UNQ(xx)`) to represent it.
otherwise, we should use `UNQ(xx, yy, ...)` to represent it.
(a corner case: two unique keys with same columns, like `UNQ(f2, f3)`,
`UNQ(f2, f3)`,
we can forbid this case or add a unique name for each key in the

future)


primary key and unique key with multiple columns example:
create table MyTable (
   f0 BIGINT NOT NULL,
   f1 ROW,
   f2 VARCHAR<256>,
   f3 AS f0 + 1,
   f4 TIMESTAMP(3) NOT NULL,
   f5 BIGINT NOT NULL,
  * PRIMARY KEY (f0, f5)*,
   *UNIQUE (f3, f2)*,
   WATERMARK f4 AS f4 - INTERVAL '3' SECOND
) with (...)






++--+---++---+--+

| name | type

   |

null   | key  | compute column | watermark
|





++--+---++---+--+

| f0   | BIGINT

  |

false | PRI(f0, f5)   |  (NULL)   |   (NULL)
 |





++--+---++---+--+

| f1   | ROW | true   | (NULL)

|

(NULL)   |  (NULL) |





++--+---++---+--+

| f2   | VARCHAR<256> | true

  |

UNQ(f2, f3) |  (NULL)   |  (NULL)
|





++--+---++---+--+

| f3   | BIGINT

  |

false | UNQ(f2, f3) |  f0 + 1  |  (NULL)
 

[DISCUSS] Release flink-shaded 11.0 (and include it in 1.11.0)

2020-05-25 Thread Chesnay Schepler

Hello,

I would like to do another flink-shaded release for 1.11.0, to include a 
zookeeper 3.4 security fix and resolve a shading issue when working with 
Gradle.





Re: [DISCUSS] Semantics of our JIRA fields

2020-05-25 Thread Zhijiang
Thanks for launching this discussion and giving so detailed infos, Robert!  +1 
on my side for the proposal. 

For "Affects Version", I previously thought it was only for the already 
released versions, so it can give a reminder that the fix should also pick into 
the related released branches for future minor versions.
I saw that Jark had somehow similar concerns for this field in below replies.  
Either way makes sense for me as long as we give a determined rule in Wiki.

Re Flavio' s comments, I agree that the Jira reporter can leave most of the 
fields empty if not confirmed of them, then the respective component maintainer 
or committer can update them accordingly later.
But the state of Jira should not be marked as "resolved" when the PR is 
detected, that is not fitting into the resolved semantic I guess. If possible, 
the Jira can be updated as "in progress" automatically if
the respective PR is ready, then it will save some time to stat progress of 
related issues during release process.

Best,
Zhijiang
--
From:Congxian Qiu 
Send Time:2020年5月25日(星期一) 13:57
To:dev@flink.apache.org 
Subject:Re: [DISCUSS] Semantics of our JIRA fields

Hi

Currently, when I'm going to create an issue for Project-website. I'm not
very sure what the "Affect Version/s" should be set. The problem is that
the current Dockfile[1] in flink-web is broken because of the EOL of Ubuntu
18.10[2], the project-web does not affect any release of Flink, it does
affect the process to build website, so what's the version should I set to?

[1]
https://github.com/apache/flink-web/blob/bc66f0f0f463ab62a22e81df7d7efd301b76a6b4/docker/Dockerfile#L17
[2] https://wiki.ubuntu.com/Releases

Best,
Congxian


Flavio Pompermaier  于2020年5月24日周日 下午1:23写道:

> In my experience it's quite complicated for a normal reporter to be able to
> fill all the fields correctly (especially for new users).
> Usually you just wanto to report a problem, remember to add a new feature
> or improve code/documentation but you can't really give a priority, assign
> the correct label or decide which releases will benefit of the fix/new
> feature. This is something that only core developers could decide (IMHO).
>
> My experiece says that it's better if normal users could just open tickets
> with some default (or mark ticket as new) and leave them in such a state
> until an experienced user, one that can assign tickets, have the time to
> read it and immediately reject the ticket or accept it and properly assign
> priorities, fix version, etc.
>
> With respect to resolve/close I think that a good practice could be to mark
> automatically a ticket as resolved once that a PR is detected for that
> ticket, while marking it as closed should be done by the commiter who merge
> the PR.
>
> Probably this process would slightly increase the work of a committer but
> this would make things a lot cleaner and will bring the benefit of having a
> reliable and semantically sound JIRA state.
>
> Cheers,
> Flavio
>
> Il Dom 24 Mag 2020, 05:05 Israel Ekpo  ha scritto:
>
> > +1 for the proposal
> >
> > This will bring some consistency to the process
> >
> > Regarding Closed vs Resolved, should we go back and switch all the
> Resolved
> > issues to Closed so that is is not confusing to new people to the project
> > in the future that may not have seen this discussion?
> >
> > I would recommend changing it to Closed just to be consistent since that
> is
> > the majority and the new process will be using Closed going forward
> >
> > Those are my thoughts for now
> >
> > On Sat, May 23, 2020 at 7:48 AM Congxian Qiu 
> > wrote:
> >
> > > +1 for the proposal. It's good to have a unified description and write
> it
> > > down in the wiki, so that every contributor has the same understanding
> of
> > > all the fields.
> > >
> > > Best,
> > > Congxian
> > >
> > >
> > > Till Rohrmann  于2020年5月23日周六 上午12:04写道:
> > >
> > > > Thanks for drafting this proposal Robert. +1 for the proposal.
> > > >
> > > > Cheers,
> > > > Till
> > > >
> > > > On Fri, May 22, 2020 at 5:39 PM Leonard Xu 
> wrote:
> > > >
> > > > > Thanks bringing up this topic @Robert,  +1 to the proposal.
> > > > >
> > > > > It clarifies the JIRA fields well and should be a rule to follow.
> > > > >
> > > > > Best,
> > > > > Leonard Xu
> > > > > > 在 2020年5月22日,20:24,Aljoscha Krettek  写道:
> > > > > >
> > > > > > +1 That's also how I think of the semantics of the fields.
> > > > > >
> > > > > > Aljoscha
> > > > > >
> > > > > > On 22.05.20 08:07, Robert Metzger wrote:
> > > > > >> Hi all,
> > > > > >> I have the feeling that the semantics of some of our JIRA fields
> > > > (mostly
> > > > > >> "affects versions", "fix versions" and resolve / close) are not
> > > > defined
> > > > > in
> > > > > >> the same way by all the core Flink contributors, which leads to
> > > cases
> > > > > where
> > > > > >> I spend quite some time on filling the fields correctly (at
> least
> > > > what 

Please update the Requirements of PyPI

2020-05-25 Thread Ray Chen
Hi PyFlink Developers,


I got some errors from the pyarrow depended on apache-flink v1.10.1.

It seems updating pyarrow version  greater than 0.14 will solve the problem.

The version of Flink on PyPI depends on < 0.14, so cause the error.


Best,
Ray Chen


Re: [DISCUSS] Semantics of our JIRA fields

2020-05-25 Thread Yun Tang
Hi

I like the idea to give clear description of JIRA fields in Flink community 
when creating tickets.

After reading from Robert's explanation, I found my previous understanding 
differs from the field at "Affect Version/s",
and is close to general understanding as JIRA official guide said[1]: Affects 
version is the version in which a bug or problem was found.

If the bug is introduced and found in Flink-1.8.1 and still existing in current 
release-1.11 branch. I would like to fill in the first version and
next major versions, eg. Flink-1.8.1, Flink-1.9.0, Flink-1.10.0 and 
Flink-1.11.0. Integrated with Robert's explanation, I think a better choice 
might be
filling in the version which brought in and all currently supported and 
unreleased Flink versions known to be affected by this.

I think it's okay to give different explanations on the same field for 
different communities.
If so, I think we should provide this notice in the official web-site, JIRA 
template or any other place instead of just dev mail list to reach 
developer/users.

[1] https://www.atlassian.com/agile/tutorials/versions

Best
Yun Tang


From: Chesnay Schepler 
Sent: Monday, May 25, 2020 14:36
To: dev@flink.apache.org ; Congxian Qiu 

Subject: Re: [DISCUSS] Semantics of our JIRA fields

flink-web is independent from any release, so there should be no
affects/fix version.

On 25/05/2020 07:56, Congxian Qiu wrote:
> Hi
>
> Currently, when I'm going to create an issue for Project-website. I'm not
> very sure what the "Affect Version/s" should be set. The problem is that
> the current Dockfile[1] in flink-web is broken because of the EOL of Ubuntu
> 18.10[2], the project-web does not affect any release of Flink, it does
> affect the process to build website, so what's the version should I set to?
>
> [1]
> https://github.com/apache/flink-web/blob/bc66f0f0f463ab62a22e81df7d7efd301b76a6b4/docker/Dockerfile#L17
> [2] https://wiki.ubuntu.com/Releases
>
> Best,
> Congxian
>
>
> Flavio Pompermaier  于2020年5月24日周日 下午1:23写道:
>
>> In my experience it's quite complicated for a normal reporter to be able to
>> fill all the fields correctly (especially for new users).
>> Usually you just wanto to report a problem, remember to add a new feature
>> or improve code/documentation but you can't really give a priority, assign
>> the correct label or decide which releases will benefit of the fix/new
>> feature. This is something that only core developers could decide (IMHO).
>>
>> My experiece says that it's better if normal users could just open tickets
>> with some default (or mark ticket as new) and leave them in such a state
>> until an experienced user, one that can assign tickets, have the time to
>> read it and immediately reject the ticket or accept it and properly assign
>> priorities, fix version, etc.
>>
>> With respect to resolve/close I think that a good practice could be to mark
>> automatically a ticket as resolved once that a PR is detected for that
>> ticket, while marking it as closed should be done by the commiter who merge
>> the PR.
>>
>> Probably this process would slightly increase the work of a committer but
>> this would make things a lot cleaner and will bring the benefit of having a
>> reliable and semantically sound JIRA state.
>>
>> Cheers,
>> Flavio
>>
>> Il Dom 24 Mag 2020, 05:05 Israel Ekpo  ha scritto:
>>
>>> +1 for the proposal
>>>
>>> This will bring some consistency to the process
>>>
>>> Regarding Closed vs Resolved, should we go back and switch all the
>> Resolved
>>> issues to Closed so that is is not confusing to new people to the project
>>> in the future that may not have seen this discussion?
>>>
>>> I would recommend changing it to Closed just to be consistent since that
>> is
>>> the majority and the new process will be using Closed going forward
>>>
>>> Those are my thoughts for now
>>>
>>> On Sat, May 23, 2020 at 7:48 AM Congxian Qiu 
>>> wrote:
>>>
 +1 for the proposal. It's good to have a unified description and write
>> it
 down in the wiki, so that every contributor has the same understanding
>> of
 all the fields.

 Best,
 Congxian


 Till Rohrmann  于2020年5月23日周六 上午12:04写道:

> Thanks for drafting this proposal Robert. +1 for the proposal.
>
> Cheers,
> Till
>
> On Fri, May 22, 2020 at 5:39 PM Leonard Xu 
>> wrote:
>> Thanks bringing up this topic @Robert,  +1 to the proposal.
>>
>> It clarifies the JIRA fields well and should be a rule to follow.
>>
>> Best,
>> Leonard Xu
>>> 在 2020年5月22日,20:24,Aljoscha Krettek  写道:
>>>
>>> +1 That's also how I think of the semantics of the fields.
>>>
>>> Aljoscha
>>>
>>> On 22.05.20 08:07, Robert Metzger wrote:
 Hi all,
 I have the feeling that the semantics of some of our JIRA fields
> (mostly
 "affects versions", "fix versions" and resolve / close) are not
> defined
>> in
 the 

Re: [DISCUSS] Semantics of our JIRA fields

2020-05-25 Thread Chesnay Schepler
flink-web is independent from any release, so there should be no 
affects/fix version.


On 25/05/2020 07:56, Congxian Qiu wrote:

Hi

Currently, when I'm going to create an issue for Project-website. I'm not
very sure what the "Affect Version/s" should be set. The problem is that
the current Dockfile[1] in flink-web is broken because of the EOL of Ubuntu
18.10[2], the project-web does not affect any release of Flink, it does
affect the process to build website, so what's the version should I set to?

[1]
https://github.com/apache/flink-web/blob/bc66f0f0f463ab62a22e81df7d7efd301b76a6b4/docker/Dockerfile#L17
[2] https://wiki.ubuntu.com/Releases

Best,
Congxian


Flavio Pompermaier  于2020年5月24日周日 下午1:23写道:


In my experience it's quite complicated for a normal reporter to be able to
fill all the fields correctly (especially for new users).
Usually you just wanto to report a problem, remember to add a new feature
or improve code/documentation but you can't really give a priority, assign
the correct label or decide which releases will benefit of the fix/new
feature. This is something that only core developers could decide (IMHO).

My experiece says that it's better if normal users could just open tickets
with some default (or mark ticket as new) and leave them in such a state
until an experienced user, one that can assign tickets, have the time to
read it and immediately reject the ticket or accept it and properly assign
priorities, fix version, etc.

With respect to resolve/close I think that a good practice could be to mark
automatically a ticket as resolved once that a PR is detected for that
ticket, while marking it as closed should be done by the commiter who merge
the PR.

Probably this process would slightly increase the work of a committer but
this would make things a lot cleaner and will bring the benefit of having a
reliable and semantically sound JIRA state.

Cheers,
Flavio

Il Dom 24 Mag 2020, 05:05 Israel Ekpo  ha scritto:


+1 for the proposal

This will bring some consistency to the process

Regarding Closed vs Resolved, should we go back and switch all the

Resolved

issues to Closed so that is is not confusing to new people to the project
in the future that may not have seen this discussion?

I would recommend changing it to Closed just to be consistent since that

is

the majority and the new process will be using Closed going forward

Those are my thoughts for now

On Sat, May 23, 2020 at 7:48 AM Congxian Qiu 
wrote:


+1 for the proposal. It's good to have a unified description and write

it

down in the wiki, so that every contributor has the same understanding

of

all the fields.

Best,
Congxian


Till Rohrmann  于2020年5月23日周六 上午12:04写道:


Thanks for drafting this proposal Robert. +1 for the proposal.

Cheers,
Till

On Fri, May 22, 2020 at 5:39 PM Leonard Xu 

wrote:

Thanks bringing up this topic @Robert,  +1 to the proposal.

It clarifies the JIRA fields well and should be a rule to follow.

Best,
Leonard Xu

在 2020年5月22日,20:24,Aljoscha Krettek  写道:

+1 That's also how I think of the semantics of the fields.

Aljoscha

On 22.05.20 08:07, Robert Metzger wrote:

Hi all,
I have the feeling that the semantics of some of our JIRA fields

(mostly

"affects versions", "fix versions" and resolve / close) are not

defined

in

the same way by all the core Flink contributors, which leads to

cases

where

I spend quite some time on filling the fields correctly (at

least

what I

consider correctly), and then others changing them again to

match

their

semantics.
In an effort to increase our efficiency, and since I'm creating

a

lot

of

(test instability-related) tickets these days, I would like to

discuss

the

semantics, come to a conclusion and document this in our Wiki.
*Proposal:*
*Priority:*
"Blocker": needs to be resolved before a release (matched based

on

fix

versions)
"Critical": strongly considered before a release
other priorities have no practical meaning in Flink.
*Component/s:*
Primary component relevant for this feature / fix.
For test-related issues, add the component the test belongs to

(for

example

"Connectors / Kafka" for Kafka test failures) + "Test".
The same applies for documentation tickets. For example, if

there's

something wrong with the DataStream API, add it to the "API /

DataStream"

and "Documentation" components.
*Affects Version/s:*Only for Bug / Task-type tickets: We list

all

currently

supported and unreleased Flink versions known to be affected by

this.

Example: If I see a test failure that happens on "master" and
"release-1.11", I set "affects version" to "1.12.0" and

"1.11.0".

*Fix Version/s:*
For closed/resolved tickets, this field lists the released Flink

versions

that contain a fix or feature for the first time.
For open tickets, it indicates that a fix / feature should be

contained

in

the listed versions. Only blocker issues can block a release,

all

other

tickets which have "fix version/s" set at the time of a release

and

are


[jira] [Created] (FLINK-17917) ResourceInformationReflector#getExternalResources should ignore the external resource with a value of 0

2020-05-25 Thread Yangze Guo (Jira)
Yangze Guo created FLINK-17917:
--

 Summary: ResourceInformationReflector#getExternalResources should 
ignore the external resource with a value of 0
 Key: FLINK-17917
 URL: https://issues.apache.org/jira/browse/FLINK-17917
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Coordination
Affects Versions: 1.11.0
Reporter: Yangze Guo
 Fix For: 1.11.0


*Background*: In FLINK-17390, we leverage 
{{WorkerSpecContainerResourceAdapter.InternalContainerResource}} to handle 
container matching logic. In FLINK-17407, we introduce external resources in 
{{WorkerSpecContainerResourceAdapter.InternalContainerResource}}.
On containers returned by Yarn, we try to get the corresponding worker specs by:
- Convert the container to {{InternalContainerResource}}
- Get the WorkerResourceSpec from {{containerResourceToWorkerSpecs}} map.

Container mismatch could happen in the below scenario:
- Flink does not allocate any external resources, the {{externalResources}} of 
{{InternalContainerResource}} is an empty map.
- The returned container contains all the resources (with a value of 0) defined 
in Yarn's {{resource-types.xml}}. The {{externalResources}} of 
{{InternalContainerResource}} has one or more entries with a value of 0.
- These two {{InternalContainerResource}} do not match.

To solve this problem, we could ignore all the external resources with a value 
of 0 in "ResourceInformationReflector#getExternalResources".

cc [~trohrmann] Could you assign this to me.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-17915) TransitiveClosureITCase>JavaProgramTestBase.testJobWithObjectReuse:113 Error while calling the test program: Could not retrieve JobResult

2020-05-25 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-17915:
--

 Summary: 
TransitiveClosureITCase>JavaProgramTestBase.testJobWithObjectReuse:113 Error 
while calling the test program: Could not retrieve JobResult
 Key: FLINK-17915
 URL: https://issues.apache.org/jira/browse/FLINK-17915
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Coordination, Tests
Affects Versions: 1.11.0
Reporter: Robert Metzger


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2096=logs=5c8e7682-d68f-54d1-16a2-a09310218a49=45cc9205-bdb7-5b54-63cd-89fdc0983323

{code}
2020-05-25T03:26:08.8891832Z [INFO] Tests run: 3, Failures: 0, Errors: 0, 
Skipped: 0, Time elapsed: 4.287 s - in 
org.apache.flink.test.example.java.WordCountSimplePOJOITCase
2020-05-25T03:26:09.6452511Z Could not retrieve JobResult.
2020-05-25T03:26:09.6454291Z 
org.apache.flink.runtime.client.JobExecutionException: Could not retrieve 
JobResult.
2020-05-25T03:26:09.6482505Zat 
org.apache.flink.runtime.minicluster.MiniCluster.executeJobBlocking(MiniCluster.java:673)
2020-05-25T03:26:09.6483671Zat 
org.apache.flink.test.util.TestEnvironment.execute(TestEnvironment.java:115)
2020-05-25T03:26:09.6625490Zat 
org.apache.flink.examples.java.graph.TransitiveClosureNaive.main(TransitiveClosureNaive.java:120)
2020-05-25T03:26:09.6752644Zat 
org.apache.flink.test.example.java.TransitiveClosureITCase.testProgram(TransitiveClosureITCase.java:51)
2020-05-25T03:26:09.6754368Zat 
org.apache.flink.test.util.JavaProgramTestBase.testJobWithObjectReuse(JavaProgramTestBase.java:107)
2020-05-25T03:26:09.6756679Zat 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
2020-05-25T03:26:09.6757511Zat 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
2020-05-25T03:26:09.6759607Zat 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
2020-05-25T03:26:09.6760692Zat 
java.lang.reflect.Method.invoke(Method.java:498)
2020-05-25T03:26:09.6761519Zat 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
2020-05-25T03:26:09.6762382Zat 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
2020-05-25T03:26:09.6763246Zat 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
2020-05-25T03:26:09.6778288Zat 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
2020-05-25T03:26:09.6779479Zat 
org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
2020-05-25T03:26:09.6780187Zat 
org.junit.rules.RunRules.evaluate(RunRules.java:20)
2020-05-25T03:26:09.6780851Zat 
org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
2020-05-25T03:26:09.6781843Zat 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
2020-05-25T03:26:09.6782583Zat 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
2020-05-25T03:26:09.6783485Zat 
org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
2020-05-25T03:26:09.6784670Zat 
org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
2020-05-25T03:26:09.6785320Zat 
org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
2020-05-25T03:26:09.6786034Zat 
org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
2020-05-25T03:26:09.6786670Zat 
org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
2020-05-25T03:26:09.6787550Zat 
org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
2020-05-25T03:26:09.6788233Zat 
org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
2020-05-25T03:26:09.6789050Zat 
org.junit.rules.RunRules.evaluate(RunRules.java:20)
2020-05-25T03:26:09.6789698Zat 
org.junit.runners.ParentRunner.run(ParentRunner.java:363)
2020-05-25T03:26:09.6790701Zat 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
2020-05-25T03:26:09.6791797Zat 
org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
2020-05-25T03:26:09.6792592Zat 
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
2020-05-25T03:26:09.6793535Zat 
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
2020-05-25T03:26:09.6794429Zat 
org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
2020-05-25T03:26:09.6795279Zat 
org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
2020-05-25T03:26:09.6796109Zat 
org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
2020-05-25T03:26:09.6797133Zat 
org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)
2020-05-25T03:26:09.6798317Z Caused by: 

[jira] [Created] (FLINK-17916) Separate KafkaShuffle read/write to different environments

2020-05-25 Thread Yuan Mei (Jira)
Yuan Mei created FLINK-17916:


 Summary: Separate KafkaShuffle read/write to different environments
 Key: FLINK-17916
 URL: https://issues.apache.org/jira/browse/FLINK-17916
 Project: Flink
  Issue Type: Improvement
  Components: API / DataStream, Connectors / Kafka
Affects Versions: 1.11.0
Reporter: Yuan Mei
 Fix For: 1.12.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-17914) HistoryServerTest.testCleanExpiredJob:158->runArchiveExpirationTest:214 expected:<2> but was:<0>

2020-05-25 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-17914:
--

 Summary: 
HistoryServerTest.testCleanExpiredJob:158->runArchiveExpirationTest:214 
expected:<2> but was:<0>
 Key: FLINK-17914
 URL: https://issues.apache.org/jira/browse/FLINK-17914
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Web Frontend
Affects Versions: 1.12.0
Reporter: Robert Metzger


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2047=logs=0da23115-68bb-5dcd-192c-bd4c8adebde1=4ed44b66-cdd6-5dcf-5f6a-88b07dda665d

{code}
[ERROR] Tests run: 6, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 4.697 s 
<<< FAILURE! - in org.apache.flink.runtime.webmonitor.history.HistoryServerTest
[ERROR] testCleanExpiredJob[Flink version less than 1.4: 
false](org.apache.flink.runtime.webmonitor.history.HistoryServerTest)  Time 
elapsed: 0.483 s  <<< FAILURE!
java.lang.AssertionError: expected:<2> but was:<0>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:645)
at org.junit.Assert.assertEquals(Assert.java:631)
at 
org.apache.flink.runtime.webmonitor.history.HistoryServerTest.runArchiveExpirationTest(HistoryServerTest.java:214)
at 
org.apache.flink.runtime.webmonitor.history.HistoryServerTest.testCleanExpiredJob(HistoryServerTest.java:158)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)