[jira] [Commented] (FLINK-8149) Replace usages of deprecated SerializationSchema

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8149?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16272291#comment-16272291
 ] 

ASF GitHub Bot commented on FLINK-8149:
---

Github user tzulitai commented on the issue:

https://github.com/apache/flink/pull/5069
  
LGTM, merging this ..


> Replace usages of deprecated SerializationSchema
> 
>
> Key: FLINK-8149
> URL: https://issues.apache.org/jira/browse/FLINK-8149
> Project: Flink
>  Issue Type: Improvement
>  Components: Kinesis Connector
>Affects Versions: 1.4.0
>Reporter: Hai Zhou UTC+8
>Assignee: Hai Zhou UTC+8
> Fix For: 1.5.0
>
>
> The deprecated {{SerializationSchema}}  in {{flink-streaming-java}}, has been 
> moved to {{flink-core}}.
> But, the deprecate {{SerializationSchema}} is still used in 
> {{flink-connector-kinesis}}.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5069: [FLINK-8149][kinesis] Replace usages of deprecated Serial...

2017-11-29 Thread tzulitai
Github user tzulitai commented on the issue:

https://github.com/apache/flink/pull/5069
  
LGTM, merging this ..


---


[jira] [Comment Edited] (FLINK-8167) Failing WikipediaEditsSourceTest on Travis

2017-11-29 Thread Hai Zhou UTC+8 (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8167?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16272233#comment-16272233
 ] 

Hai Zhou UTC+8 edited comment on FLINK-8167 at 11/30/17 6:59 AM:
-

I agree with "move some low popularity connectors to Apache Bahir". about this, 
I mentioned a ticket FLINK-7510 before.

BTW, It is very common for test suites or build scripts to hang. Travis CI has 
specific time limits for each job, and will stop. {{A job on travis-ci.org 
takes longer than 50 minutes, will be stoped.}}
Currently, the "misc" profile mvn script executes very close to 50 minutes and 
occasionally builds timeouts. 
Do we have a better way to solve this problem?



was (Author: yew1eb):
I agree with "move some low popularity connectors to Apache Bahir". about this, 
I mentioned a ticket FLINK-7510 before.

BTW, It is very common for test suites or build scripts to hang. Travis CI has 
specific time limits for each job, and will stop. {{A job on travis-ci.org 
takes longer than 50 minutes, will be stoped.}}
Currently, the "misc" profile mvn script executes very close to 50 minutes and 
occasionally builds timeouts. 
We have no better way to solve this problem?

> Failing WikipediaEditsSourceTest on Travis
> --
>
> Key: FLINK-8167
> URL: https://issues.apache.org/jira/browse/FLINK-8167
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.5.0
>Reporter: Till Rohrmann
>Assignee: Ufuk Celebi
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.5.0
>
>
> I've observed the {{WikipediaEditsSourceTest}} failing on Travis.
> https://travis-ci.org/tillrohrmann/flink/jobs/308351298



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-8167) Failing WikipediaEditsSourceTest on Travis

2017-11-29 Thread Hai Zhou UTC+8 (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8167?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16272233#comment-16272233
 ] 

Hai Zhou UTC+8 commented on FLINK-8167:
---

I agree with "move some low popularity connectors to Apache Bahir". about this, 
I mentioned a ticket FLINK-7510 before.

BTW, It is very common for test suites or build scripts to hang. Travis CI has 
specific time limits for each job, and will stop. {{A job on travis-ci.org 
takes longer than 50 minutes, will be stoped.}}
Currently, the "misc" profile mvn script executes very close to 50 minutes and 
occasionally builds timeouts. 
We have no better way to solve this problem?

> Failing WikipediaEditsSourceTest on Travis
> --
>
> Key: FLINK-8167
> URL: https://issues.apache.org/jira/browse/FLINK-8167
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.5.0
>Reporter: Till Rohrmann
>Assignee: Ufuk Celebi
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.5.0
>
>
> I've observed the {{WikipediaEditsSourceTest}} failing on Travis.
> https://travis-ci.org/tillrohrmann/flink/jobs/308351298



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-8158) Rowtime window inner join emits late data

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8158?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16272204#comment-16272204
 ] 

ASF GitHub Bot commented on FLINK-8158:
---

Github user hequn8128 commented on the issue:

https://github.com/apache/flink/pull/5094
  
Thanks for your replies.
I agree with that it is valid to join late input data.What I concern is the 
watermark has not been hold back correctly.

Take `testRowTimeJoinWithCommonBounds2` as an example.
1. Output watermark with timestamp 1000 (hold back with 5000 from 6000)
2. Output record1 with timestamp 6000 (valid)
3. Output record2 with timestamp 1000 (invalid)

The record2 is outputted invalidly because it's timestamp is equal to the 
previous output watermark (1000). In my pr, I hold the input watermark back 
with 5001 to make record2 valid. So the test will output with:

1. Output watermark with timestamp 999 (hold back with 5001 from 6000)
2. Output record1 with timestamp 6000 (valid)
3. Output record2 with timestamp 1000 (valid) 


> Rowtime window inner join emits late data
> -
>
> Key: FLINK-8158
> URL: https://issues.apache.org/jira/browse/FLINK-8158
> Project: Flink
>  Issue Type: Bug
>  Components: Table API & SQL
>Reporter: Hequn Cheng
>Assignee: Hequn Cheng
> Attachments: screenshot-1xxx.png
>
>
> When executing the join, the join operator needs to make sure that no late 
> data is emitted. Currently, this achieved by holding back watermarks. 
> However, the window border is not handled correctly. For the sql bellow: 
> {quote}
> val sqlQuery =
>   """
> SELECT t2.key, t2.id, t1.id
> FROM T1 as t1 join T2 as t2 ON
>   t1.key = t2.key AND
>   t1.rt BETWEEN t2.rt - INTERVAL '5' SECOND AND
> t2.rt + INTERVAL '1' SECOND
> """.stripMargin
> val data1 = new mutable.MutableList[(String, String, Long)]
> // for boundary test
> data1.+=(("A", "LEFT1", 6000L))
> val data2 = new mutable.MutableList[(String, String, Long)]
> data2.+=(("A", "RIGHT1", 6000L))
> {quote}
> Join will output a watermark with timestamp 1000, but if left comes with 
> another data ("A", "LEFT1", 1000L), join will output a record with timestamp 
> 1000 which equals previous watermark.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5094: [FLINK-8158] [table] Fix rowtime window inner join emits ...

2017-11-29 Thread hequn8128
Github user hequn8128 commented on the issue:

https://github.com/apache/flink/pull/5094
  
Thanks for your replies.
I agree with that it is valid to join late input data.What I concern is the 
watermark has not been hold back correctly.

Take `testRowTimeJoinWithCommonBounds2` as an example.
1. Output watermark with timestamp 1000 (hold back with 5000 from 6000)
2. Output record1 with timestamp 6000 (valid)
3. Output record2 with timestamp 1000 (invalid)

The record2 is outputted invalidly because it's timestamp is equal to the 
previous output watermark (1000). In my pr, I hold the input watermark back 
with 5001 to make record2 valid. So the test will output with:

1. Output watermark with timestamp 999 (hold back with 5001 from 6000)
2. Output record1 with timestamp 6000 (valid)
3. Output record2 with timestamp 1000 (valid) 


---


[jira] [Updated] (FLINK-8175) remove flink-streaming-contrib and migrate its classes to flink-streaming-java/scala

2017-11-29 Thread Bowen Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-8175?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bowen Li updated FLINK-8175:

Description: 
I propose removing flink-streaming-contrib from flink-contrib, and migrating 
its classes to flink-streaming-java/scala for the following reasons:

- flink-streaming-contrib is so small that it only has 4 classes (3 java and 1 
scala), and they don't need a dedicated jar for Flink to distribute and 
maintain it and for users to deal with the overhead of dependency management
- the 4 classes in flink-streaming-contrib are logically more tied to 
flink-streaming-java/scala, and thus can be easily migrated
- flink-contrib is already too crowded and noisy. It contains lots of sub 
modules with different purposes which confuse developers and users, and they 
lack a proper project hierarchy

To take a step even forward, I would argue that even flink-contrib should be 
removed and all its submodules should be migrated to other top-level modules 
for the following reasons: 1) Apache Flink the whole project itself is a result 
of contributions from many developers, there's no reason to highlight some 
contributions in a dedicated module named 'contrib' 2) flink-contrib inherently 
doesn't have a good hierarchy to hold submodules

  was:
I propose removing flink-streaming-contrib from flink-contrib, and migrating 
its classes to flink-streaming-java/scala for the following reasons:

- flink-streaming-contrib is so small that it only has 4 classes (3 java and 1 
scala), and they don't need a dedicated jar for Flink to distribute and 
maintain it and for users to deal with the overhead of dependency management
- the 4 classes in flink-streaming-contrib are logically more tied to 
flink-streaming-java/scala, and thus can be easily migrated
- flink-contrib is already too crowded and noisy. It contains lots of sub 
modules with different purposes which confuse developers and users, and they 
lack a proper project hierarchy

To take a step even forward, I would argue that even flink-contrib should be 
removed and its submodules should be migrated to other top-level modules for 
the following reasons: 1) Apache Flink the whole project itself is a result of 
contributions from all developers, there's no reason to highlight some 
contributions in a dedicated module named 'contrib' 2) flink-contrib inherently 
doesn't have a good hierarchy to hold submodules


> remove flink-streaming-contrib and migrate its classes to 
> flink-streaming-java/scala
> 
>
> Key: FLINK-8175
> URL: https://issues.apache.org/jira/browse/FLINK-8175
> Project: Flink
>  Issue Type: Improvement
>Affects Versions: 1.5.0
>Reporter: Bowen Li
>
> I propose removing flink-streaming-contrib from flink-contrib, and migrating 
> its classes to flink-streaming-java/scala for the following reasons:
> - flink-streaming-contrib is so small that it only has 4 classes (3 java and 
> 1 scala), and they don't need a dedicated jar for Flink to distribute and 
> maintain it and for users to deal with the overhead of dependency management
> - the 4 classes in flink-streaming-contrib are logically more tied to 
> flink-streaming-java/scala, and thus can be easily migrated
> - flink-contrib is already too crowded and noisy. It contains lots of sub 
> modules with different purposes which confuse developers and users, and they 
> lack a proper project hierarchy
> To take a step even forward, I would argue that even flink-contrib should be 
> removed and all its submodules should be migrated to other top-level modules 
> for the following reasons: 1) Apache Flink the whole project itself is a 
> result of contributions from many developers, there's no reason to highlight 
> some contributions in a dedicated module named 'contrib' 2) flink-contrib 
> inherently doesn't have a good hierarchy to hold submodules



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (FLINK-8175) remove flink-streaming-contrib and migrate its classes to flink-streaming-java/scala

2017-11-29 Thread Bowen Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-8175?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bowen Li updated FLINK-8175:

Description: 
I propose removing flink-streaming-contrib from flink-contrib, and migrating 
its classes to flink-streaming-java/scala for the following reasons:

- flink-streaming-contrib is so small that it only has 4 classes (3 java and 1 
scala), and they don't need a dedicated jar for Flink to distribute and 
maintain it and for users to deal with the overhead of dependency management
- the 4 classes in flink-streaming-contrib are logically more tied to 
flink-streaming-java/scala, and thus can be easily migrated
- flink-contrib is already too crowded and noisy. It contains lots of sub 
modules with different purposes which confuse developers and users, and they 
lack a proper project hierarchy

To take a step even forward, I would argue that even flink-contrib should be 
removed and its submodules should be migrated to other top-level modules for 
the following reasons: 1) Apache Flink the whole project itself is a result of 
contributions from all developers, there's no reason to highlight some 
contributions in a dedicated module named 'contrib' 2) flink-contrib inherently 
doesn't have a good hierarchy to hold submodules

  was:
I propose removing flink-streaming-contrib from flink-contrib, and migrating 
its classes to flink-streaming-java/scala for the following reasons:

- flink-streaming-contrib is so small that it only has 4 classes (3 java and 1 
scala), and they don't need a dedicated jar for Flink to distribute and 
maintain it and for users to deal with the overhead of dependency management
- the 4 classes in flink-streaming-contrib are logically more tied to 
flink-streaming-java/scala, and thus can be easily migrated
- flink-contrib is already too crowded and noisy. It contains lots of sub 
modules with different purposes which confuse developers and users, and they 
lack a proper project hierarchy


According to the history of flink-streaming-contrib, it has things to do with 
storm-compatibility and I believe that's not blocking us from migrating it 
anymore.


> remove flink-streaming-contrib and migrate its classes to 
> flink-streaming-java/scala
> 
>
> Key: FLINK-8175
> URL: https://issues.apache.org/jira/browse/FLINK-8175
> Project: Flink
>  Issue Type: Improvement
>Affects Versions: 1.5.0
>Reporter: Bowen Li
>
> I propose removing flink-streaming-contrib from flink-contrib, and migrating 
> its classes to flink-streaming-java/scala for the following reasons:
> - flink-streaming-contrib is so small that it only has 4 classes (3 java and 
> 1 scala), and they don't need a dedicated jar for Flink to distribute and 
> maintain it and for users to deal with the overhead of dependency management
> - the 4 classes in flink-streaming-contrib are logically more tied to 
> flink-streaming-java/scala, and thus can be easily migrated
> - flink-contrib is already too crowded and noisy. It contains lots of sub 
> modules with different purposes which confuse developers and users, and they 
> lack a proper project hierarchy
> To take a step even forward, I would argue that even flink-contrib should be 
> removed and its submodules should be migrated to other top-level modules for 
> the following reasons: 1) Apache Flink the whole project itself is a result 
> of contributions from all developers, there's no reason to highlight some 
> contributions in a dedicated module named 'contrib' 2) flink-contrib 
> inherently doesn't have a good hierarchy to hold submodules



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Comment Edited] (FLINK-7294) mesos.resourcemanager.framework.role not working

2017-11-29 Thread Bhumika Bayani (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7294?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16272190#comment-16272190
 ] 

Bhumika Bayani edited comment on FLINK-7294 at 11/30/17 5:34 AM:
-

Hi [~eronwright], 
As per this link 
http://mesos.apache.org/documentation/latest/roles/#default-role, the 
unreserved resources are assigned to "*" role by default. So the fix should 
have worked even for unreserved resources as per me. I remember testing this 
with mesos 1.0.

If that is not the case, i.e. unreserved resources are not assigned to * role 
by default, this fix wont work and we will need a fix that works for both 
conditions.



was (Author: bbayani):
Hi [~eronwright], 
As per this link 
http://mesos.apache.org/documentation/latest/roles/#default-role, the 
unreserved resources are assigned to "*" role by default. So the fix should 
have worked even for unreserved resources as per me.

If that is not the case, i.e. unreserved resources are not assigned to * role 
by default, this fix wont work and we will need a fix that works for both 
conditions.


> mesos.resourcemanager.framework.role not working
> 
>
> Key: FLINK-7294
> URL: https://issues.apache.org/jira/browse/FLINK-7294
> Project: Flink
>  Issue Type: Bug
>  Components: Mesos
>Affects Versions: 1.3.1
>Reporter: Bhumika Bayani
>Assignee: Eron Wright 
>Priority: Critical
> Fix For: 1.4.0, 1.3.3
>
>
> I am using the above said setting in flink-conf.yaml
> e.g.
> mesos.resourcemanager.framework.role: mesos_role_tasks
> I see a flink-scheduler registered in mesos/frameworks tab with above said 
> role.
> But the scheduler fails to launch any tasks inspite of getting 
> resource-offers from mesos-agents with correct role.
> The error seen is:
> {code}
> 2017-07-28 13:23:00,683 INFO  
> org.apache.flink.mesos.runtime.clusterframework.MesosFlinkResourceManager  - 
> Mesos task taskmanager-03768 failed, with a TaskManager in launch or 
> registration. State: TASK_ERROR Reason: REASON_TASK_INVALID (Task uses more 
> resources cpus(\*):1; mem(\*):1024; ports(\*):[4006-4007] than available 
> cpus(mesos_role_tasks):7.4; mem(mesos_role_tasks):45876; 
> ports(mesos_role_tasks):[4002-4129, 4131-4380, 4382-4809, 4811-4957, 
> 4959-4966, 4968-4979, 4981-5049, 31000-31196, 31198-31431, 31433-31607, 
> 31609-32000]; ephemeral_storage(mesos_role_tasks):37662; 
> efs_storage(mesos_role_tasks):8.79609e+12; disk(mesos_role_tasks):5115)
> {code}
> The request is made for resources with * role. We do not have mesos running 
> anywhere with * role. Thus task manager never come up. 
> Am I missing any configuration?
> I am using flink version 1.3.1



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-7294) mesos.resourcemanager.framework.role not working

2017-11-29 Thread Bhumika Bayani (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7294?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16272190#comment-16272190
 ] 

Bhumika Bayani commented on FLINK-7294:
---

Hi [~eronwright], 
As per this link 
http://mesos.apache.org/documentation/latest/roles/#default-role, the 
unreserved resources are assigned to "*" role by default. So the fix should 
have worked even for unreserved resources as per me.

If that is not the case, i.e. unreserved resources are not assigned to * role 
by default, this fix wont work and we will need a fix that works for both 
conditions.


> mesos.resourcemanager.framework.role not working
> 
>
> Key: FLINK-7294
> URL: https://issues.apache.org/jira/browse/FLINK-7294
> Project: Flink
>  Issue Type: Bug
>  Components: Mesos
>Affects Versions: 1.3.1
>Reporter: Bhumika Bayani
>Assignee: Eron Wright 
>Priority: Critical
> Fix For: 1.4.0, 1.3.3
>
>
> I am using the above said setting in flink-conf.yaml
> e.g.
> mesos.resourcemanager.framework.role: mesos_role_tasks
> I see a flink-scheduler registered in mesos/frameworks tab with above said 
> role.
> But the scheduler fails to launch any tasks inspite of getting 
> resource-offers from mesos-agents with correct role.
> The error seen is:
> {code}
> 2017-07-28 13:23:00,683 INFO  
> org.apache.flink.mesos.runtime.clusterframework.MesosFlinkResourceManager  - 
> Mesos task taskmanager-03768 failed, with a TaskManager in launch or 
> registration. State: TASK_ERROR Reason: REASON_TASK_INVALID (Task uses more 
> resources cpus(\*):1; mem(\*):1024; ports(\*):[4006-4007] than available 
> cpus(mesos_role_tasks):7.4; mem(mesos_role_tasks):45876; 
> ports(mesos_role_tasks):[4002-4129, 4131-4380, 4382-4809, 4811-4957, 
> 4959-4966, 4968-4979, 4981-5049, 31000-31196, 31198-31431, 31433-31607, 
> 31609-32000]; ephemeral_storage(mesos_role_tasks):37662; 
> efs_storage(mesos_role_tasks):8.79609e+12; disk(mesos_role_tasks):5115)
> {code}
> The request is made for resources with * role. We do not have mesos running 
> anywhere with * role. Thus task manager never come up. 
> Am I missing any configuration?
> I am using flink version 1.3.1



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-8167) Failing WikipediaEditsSourceTest on Travis

2017-11-29 Thread Bowen Li (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8167?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16272182#comment-16272182
 ] 

Bowen Li commented on FLINK-8167:
-

I also opened FLINK-8175. Let me know what you guys think

> Failing WikipediaEditsSourceTest on Travis
> --
>
> Key: FLINK-8167
> URL: https://issues.apache.org/jira/browse/FLINK-8167
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.5.0
>Reporter: Till Rohrmann
>Assignee: Ufuk Celebi
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.5.0
>
>
> I've observed the {{WikipediaEditsSourceTest}} failing on Travis.
> https://travis-ci.org/tillrohrmann/flink/jobs/308351298



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (FLINK-8175) remove flink-streaming-contrib and migrate its classes to flink-streaming-java/scala

2017-11-29 Thread Bowen Li (JIRA)
Bowen Li created FLINK-8175:
---

 Summary: remove flink-streaming-contrib and migrate its classes to 
flink-streaming-java/scala
 Key: FLINK-8175
 URL: https://issues.apache.org/jira/browse/FLINK-8175
 Project: Flink
  Issue Type: Improvement
Affects Versions: 1.5.0
Reporter: Bowen Li


I propose removing flink-streaming-contrib from flink-contrib, and migrating 
its classes to flink-streaming-java/scala for the following reasons:

- flink-streaming-contrib is so small that it only has 4 classes (3 java and 1 
scala), and they don't need a dedicated jar for Flink to distribute and 
maintain it and for users to deal with the overhead of dependency management
- the 4 classes in flink-streaming-contrib are logically more tied to 
flink-streaming-java/scala, and thus can be easily migrated
- flink-contrib is already too crowded and noisy. It contains lots of sub 
modules with different purposes which confuse developers and users, and they 
lack a proper project hierarchy


According to the history of flink-streaming-contrib, it has things to do with 
storm-compatibility and I believe that's not blocking us from migrating it 
anymore.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-8167) Failing WikipediaEditsSourceTest on Travis

2017-11-29 Thread Bowen Li (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8167?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16272162#comment-16272162
 ] 

Bowen Li commented on FLINK-8167:
-

I'm wondering why wikiedit has to be part of Flink itself? I don't see much 
necessity given its low popularity in real use case in Flink community. 
Furthermore, flink-contrib is too crowded and noisy with lots of things which 
are of different purposes, and those all result in the growing build time :(.

Can we move wikiedit to Apache Bahir?

> Failing WikipediaEditsSourceTest on Travis
> --
>
> Key: FLINK-8167
> URL: https://issues.apache.org/jira/browse/FLINK-8167
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.5.0
>Reporter: Till Rohrmann
>Assignee: Ufuk Celebi
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.5.0
>
>
> I've observed the {{WikipediaEditsSourceTest}} failing on Travis.
> https://travis-ci.org/tillrohrmann/flink/jobs/308351298



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-7294) mesos.resourcemanager.framework.role not working

2017-11-29 Thread Eron Wright (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7294?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16272120#comment-16272120
 ] 

Eron Wright  commented on FLINK-7294:
-

[~bbayani] I believe the fix we came up with is not correct.   Looking at this 
bug again, it appears that you were using a reservation (static or dynamic, 
doesn't matter) and finding that Flink was unable to use _reserved _resources.  
Unfortunately the fix has reversed the situation, and Flink is now unable to 
use _unreserved _resources.   We must find a solution that works for both 
situations.  Do you agree with my characterization?  Thanks.

See FLINK-8174.

> mesos.resourcemanager.framework.role not working
> 
>
> Key: FLINK-7294
> URL: https://issues.apache.org/jira/browse/FLINK-7294
> Project: Flink
>  Issue Type: Bug
>  Components: Mesos
>Affects Versions: 1.3.1
>Reporter: Bhumika Bayani
>Assignee: Eron Wright 
>Priority: Critical
> Fix For: 1.4.0, 1.3.3
>
>
> I am using the above said setting in flink-conf.yaml
> e.g.
> mesos.resourcemanager.framework.role: mesos_role_tasks
> I see a flink-scheduler registered in mesos/frameworks tab with above said 
> role.
> But the scheduler fails to launch any tasks inspite of getting 
> resource-offers from mesos-agents with correct role.
> The error seen is:
> {code}
> 2017-07-28 13:23:00,683 INFO  
> org.apache.flink.mesos.runtime.clusterframework.MesosFlinkResourceManager  - 
> Mesos task taskmanager-03768 failed, with a TaskManager in launch or 
> registration. State: TASK_ERROR Reason: REASON_TASK_INVALID (Task uses more 
> resources cpus(\*):1; mem(\*):1024; ports(\*):[4006-4007] than available 
> cpus(mesos_role_tasks):7.4; mem(mesos_role_tasks):45876; 
> ports(mesos_role_tasks):[4002-4129, 4131-4380, 4382-4809, 4811-4957, 
> 4959-4966, 4968-4979, 4981-5049, 31000-31196, 31198-31431, 31433-31607, 
> 31609-32000]; ephemeral_storage(mesos_role_tasks):37662; 
> efs_storage(mesos_role_tasks):8.79609e+12; disk(mesos_role_tasks):5115)
> {code}
> The request is made for resources with * role. We do not have mesos running 
> anywhere with * role. Thus task manager never come up. 
> Am I missing any configuration?
> I am using flink version 1.3.1



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (FLINK-8174) Mesos RM unable to accept offers for unreserved resources

2017-11-29 Thread Eron Wright (JIRA)
Eron Wright  created FLINK-8174:
---

 Summary: Mesos RM unable to accept offers for unreserved resources
 Key: FLINK-8174
 URL: https://issues.apache.org/jira/browse/FLINK-8174
 Project: Flink
  Issue Type: Bug
  Components: Mesos
Affects Versions: 1.4.0, 1.3.3
Reporter: Eron Wright 
Assignee: Eron Wright 
Priority: Blocker
 Fix For: 1.4.0


Flink has suffered a regression due to FLINK-7294.   Any attempt to accept a 
resource offer that is based on unreserved resources will fail, because Flink 
(as of FLINK-7294) erroneously insists that the resource come from a prior 
reservation.

Looking at the original issue, the problem may have been misdiagnosed.  Ideally 
Flink should work with both reserved and unreserved resources, but the latter 
is a more common situation that is now broken.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-8173) InvalidProgramException: Table program cannot be compiled. This is a bug. Please file an issue.

2017-11-29 Thread Aegeaner (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8173?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271988#comment-16271988
 ] 

Aegeaner commented on FLINK-8173:
-

Assignment conversion not possible from type "java.lang.CharSequence" to type 
"org.apache.avro.util.Utf8" 

> InvalidProgramException: Table program cannot be compiled. This is a bug. 
> Please file an issue.
> ---
>
> Key: FLINK-8173
> URL: https://issues.apache.org/jira/browse/FLINK-8173
> Project: Flink
>  Issue Type: Bug
>Reporter: Tao Xia
>
> It is a stream of Avro objects, simply select a String field and trying to 
> print out
> val query = "SELECT nd_key FROM table1"
> val result = tableEnv.sql(query)
> tableEnv.toAppendStream[org.apache.avro.util.Utf8](result).print()
> 11/29/2017 16:07:36   Source: Custom Source -> from: (accepted_cohort_id, 
> admin_id, after_submission, amount_paid, anonymous_id, application_id, 
> atom_key, bd_group_key, biz_geo, braavos_purchase_id, category, cohort_id, 
> concept_key, concept_rank, context, context_campaign, context_experiment, 
> coupon_code, course_key, course_rank, cta_destination, cta_location, 
> cta_message, cta_type, currency, decision_group_id, device_browser, 
> device_os, device_os_version, device_type, duration, evaluation_id, 
> event_type, fin_geo, in_collaboration_with, lab_id, lab_rank, label, 
> lesson_key, lesson_rank, locale, max_pause_duration, message, message_id, 
> module_key, module_rank, nd_key, nd_unit_id, nd_unit_rank, new_cohort_id, 
> notification_id, num_concepts_completed, num_interactions, 
> num_lessons_completed, old_cohort_id, part_key, part_rank, pause_duration, 
> pause_reason, payment_plan, payment_provider, points_earned, points_possible, 
> price, price_sheet, product_key, product_type, provider_charge_id, 
> provider_refund_id, quiz_type, referrer, refund_amount, requested_cohort_id, 
> results, scholarship_group_key, search_term, skill_level, subscription_id, 
> suspension_length, suspension_reason, technology, timestamp, total_concepts, 
> total_lessons, total_time_sec, type, unenroll_reason, user_id, user_locale, 
> user_response, variant, version, workspace_id, workspace_session, 
> workspace_type) -> select: (nd_key) -> to: Utf8 -> Sink: Unnamed(5/8) 
> switched to FAILED 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.
>   at 
> org.apache.flink.table.codegen.Compiler$class.compile(Compiler.scala:36)
>   at 
> org.apache.flink.table.runtime.CRowOutputMapRunner.compile(CRowOutputMapRunner.scala:33)
>   at 
> org.apache.flink.table.runtime.CRowOutputMapRunner.open(CRowOutputMapRunner.scala:48)
>   at 
> org.apache.flink.api.common.functions.util.FunctionUtils.openFunction(FunctionUtils.java:36)
>   at 
> org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.open(AbstractUdfStreamOperator.java:111)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.openAllOperators(StreamTask.java:376)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:253)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:702)
>   at java.lang.Thread.run(Thread.java:748)
> Caused by: org.codehaus.commons.compiler.CompileException: Line 790, Column 
> 15: Assignment conversion not possible from type "java.lang.CharSequence" to 
> type "org.apache.avro.util.Utf8"
>   at 
> org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:11672)
>   at 
> org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10528)
>   at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:2534)
>   at org.codehaus.janino.UnitCompiler.access$2600(UnitCompiler.java:212)
>   at 
> org.codehaus.janino.UnitCompiler$6.visitLocalVariableDeclarationStatement(UnitCompiler.java:1459)
>   at 
> org.codehaus.janino.UnitCompiler$6.visitLocalVariableDeclarationStatement(UnitCompiler.java:1443)
>   at 
> org.codehaus.janino.Java$LocalVariableDeclarationStatement.accept(Java.java:3348)
>   at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:1443)
>   at 
> org.codehaus.janino.UnitCompiler.compileStatements(UnitCompiler.java:1523)
>   at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:3052)
>   at 
> org.codehaus.janino.UnitCompiler.compileDeclaredMethods(UnitCompiler.java:1313)
>   at 
> org.codehaus.janino.UnitCompiler.compileDeclaredMethods(UnitCompiler.java:1286)
>   at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:785)
>   at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:436)
>   at org.codehaus.janino.UnitCompiler.access$400(UnitCompiler.java:212)
>   at

[jira] [Created] (FLINK-8173) InvalidProgramException: Table program cannot be compiled. This is a bug. Please file an issue.

2017-11-29 Thread Tao Xia (JIRA)
Tao Xia created FLINK-8173:
--

 Summary: InvalidProgramException: Table program cannot be 
compiled. This is a bug. Please file an issue.
 Key: FLINK-8173
 URL: https://issues.apache.org/jira/browse/FLINK-8173
 Project: Flink
  Issue Type: Bug
Reporter: Tao Xia


It is a stream of Avro objects, simply select a String field and trying to 
print out
val query = "SELECT nd_key FROM table1"
val result = tableEnv.sql(query)
tableEnv.toAppendStream[org.apache.avro.util.Utf8](result).print()


11/29/2017 16:07:36 Source: Custom Source -> from: (accepted_cohort_id, 
admin_id, after_submission, amount_paid, anonymous_id, application_id, 
atom_key, bd_group_key, biz_geo, braavos_purchase_id, category, cohort_id, 
concept_key, concept_rank, context, context_campaign, context_experiment, 
coupon_code, course_key, course_rank, cta_destination, cta_location, 
cta_message, cta_type, currency, decision_group_id, device_browser, device_os, 
device_os_version, device_type, duration, evaluation_id, event_type, fin_geo, 
in_collaboration_with, lab_id, lab_rank, label, lesson_key, lesson_rank, 
locale, max_pause_duration, message, message_id, module_key, module_rank, 
nd_key, nd_unit_id, nd_unit_rank, new_cohort_id, notification_id, 
num_concepts_completed, num_interactions, num_lessons_completed, old_cohort_id, 
part_key, part_rank, pause_duration, pause_reason, payment_plan, 
payment_provider, points_earned, points_possible, price, price_sheet, 
product_key, product_type, provider_charge_id, provider_refund_id, quiz_type, 
referrer, refund_amount, requested_cohort_id, results, scholarship_group_key, 
search_term, skill_level, subscription_id, suspension_length, 
suspension_reason, technology, timestamp, total_concepts, total_lessons, 
total_time_sec, type, unenroll_reason, user_id, user_locale, user_response, 
variant, version, workspace_id, workspace_session, workspace_type) -> select: 
(nd_key) -> to: Utf8 -> Sink: Unnamed(5/8) switched to FAILED 
org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
compiled. This is a bug. Please file an issue.
at 
org.apache.flink.table.codegen.Compiler$class.compile(Compiler.scala:36)
at 
org.apache.flink.table.runtime.CRowOutputMapRunner.compile(CRowOutputMapRunner.scala:33)
at 
org.apache.flink.table.runtime.CRowOutputMapRunner.open(CRowOutputMapRunner.scala:48)
at 
org.apache.flink.api.common.functions.util.FunctionUtils.openFunction(FunctionUtils.java:36)
at 
org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.open(AbstractUdfStreamOperator.java:111)
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.openAllOperators(StreamTask.java:376)
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:253)
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:702)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.codehaus.commons.compiler.CompileException: Line 790, Column 15: 
Assignment conversion not possible from type "java.lang.CharSequence" to type 
"org.apache.avro.util.Utf8"
at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:11672)
at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10528)
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:2534)
at org.codehaus.janino.UnitCompiler.access$2600(UnitCompiler.java:212)
at 
org.codehaus.janino.UnitCompiler$6.visitLocalVariableDeclarationStatement(UnitCompiler.java:1459)
at 
org.codehaus.janino.UnitCompiler$6.visitLocalVariableDeclarationStatement(UnitCompiler.java:1443)
at 
org.codehaus.janino.Java$LocalVariableDeclarationStatement.accept(Java.java:3348)
at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:1443)
at 
org.codehaus.janino.UnitCompiler.compileStatements(UnitCompiler.java:1523)
at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:3052)
at 
org.codehaus.janino.UnitCompiler.compileDeclaredMethods(UnitCompiler.java:1313)
at 
org.codehaus.janino.UnitCompiler.compileDeclaredMethods(UnitCompiler.java:1286)
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:785)
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:436)
at org.codehaus.janino.UnitCompiler.access$400(UnitCompiler.java:212)
at 
org.codehaus.janino.UnitCompiler$2.visitPackageMemberClassDeclaration(UnitCompiler.java:390)
at 
org.codehaus.janino.UnitCompiler$2.visitPackageMemberClassDeclaration(UnitCompiler.java:385)
at 
org.codehaus.janino.Java$PackageMemberClassDeclaration.accept(Java.java:1405)
at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:385)
at org.codehaus.janino.UnitCompiler.compileUnit(UnitCompiler.java:357)
at org.codehaus.j

[jira] [Commented] (FLINK-7907) Flink Metrics documentation missing Scala examples

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271580#comment-16271580
 ] 

ASF GitHub Bot commented on FLINK-7907:
---

Github user zentol commented on a diff in the pull request:

https://github.com/apache/flink/pull/5071#discussion_r153923624
  
--- Diff: docs/monitoring/metrics.md ---
@@ -94,8 +94,8 @@ Alternatively you can also use your own `Counter` 
implementation:
 
 {% highlight java %}
 
-public class MyMapper extends RichMapFunction {
-  private Counter counter;
+RichMapFunction {
--- End diff --

looks like you removed more than intended?


> Flink Metrics documentation missing Scala examples
> --
>
> Key: FLINK-7907
> URL: https://issues.apache.org/jira/browse/FLINK-7907
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Colin Williams
>Assignee: Hai Zhou UTC+8
>Priority: Minor
>
> The Flink metrics documentation is missing Scala examples for many of the 
> metrics types. To be consistent there should be Scala examples for all the 
> types.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink pull request #5071: [FLINK-7907][docs] The metrics documentation missi...

2017-11-29 Thread zentol
Github user zentol commented on a diff in the pull request:

https://github.com/apache/flink/pull/5071#discussion_r153923624
  
--- Diff: docs/monitoring/metrics.md ---
@@ -94,8 +94,8 @@ Alternatively you can also use your own `Counter` 
implementation:
 
 {% highlight java %}
 
-public class MyMapper extends RichMapFunction {
-  private Counter counter;
+RichMapFunction {
--- End diff --

looks like you removed more than intended?


---


[jira] [Commented] (FLINK-4812) Report Watermark metrics in all operators

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271487#comment-16271487
 ] 

ASF GitHub Bot commented on FLINK-4812:
---

Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5100
  
yes, we _could_ setup all IO metrics in the operator chain. But that will 
require some more changes there since the methods there never access 2 
operators at the same time. 
For a chain A->B, to setup the output of A you need metrics for both A and 
B.
You cannot be "clever" and re-use the output metrics of A as the input 
metrics for B, since this will cause odd results in forking chain scenarios 
where the input metrics for both chained operators change at the same time even 
though they are always processed sequentially. This would get even more funky 
if an operator decides that it's input/output metrics are always supposed to be 
identical, at which point you have metrics for 3 operators changing at the same 
time.

But just thinking about modifying the `OperatorChain` sends shivers down my 
spine ❄️ 


> Report Watermark metrics in all operators
> -
>
> Key: FLINK-4812
> URL: https://issues.apache.org/jira/browse/FLINK-4812
> Project: Flink
>  Issue Type: Improvement
>  Components: Metrics
>Reporter: Robert Metzger
>Assignee: Chesnay Schepler
>Priority: Critical
> Fix For: 1.5.0
>
>
> As reported by a user, Flink does currently not export the current low 
> watermark for sources 
> (http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/currentLowWatermark-metric-not-reported-for-all-tasks-td13770.html).
> This JIRA is for adding such a metric for the sources as well.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5100: [FLINK-4812][metrics] Expose currentLowWatermark for all ...

2017-11-29 Thread zentol
Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5100
  
yes, we _could_ setup all IO metrics in the operator chain. But that will 
require some more changes there since the methods there never access 2 
operators at the same time. 
For a chain A->B, to setup the output of A you need metrics for both A and 
B.
You cannot be "clever" and re-use the output metrics of A as the input 
metrics for B, since this will cause odd results in forking chain scenarios 
where the input metrics for both chained operators change at the same time even 
though they are always processed sequentially. This would get even more funky 
if an operator decides that it's input/output metrics are always supposed to be 
identical, at which point you have metrics for 3 operators changing at the same 
time.

But just thinking about modifying the `OperatorChain` sends shivers down my 
spine ❄️ 


---


[jira] [Commented] (FLINK-8151) [Table] Map equality check to use entrySet equality

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8151?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271417#comment-16271417
 ] 

ASF GitHub Bot commented on FLINK-8151:
---

Github user walterddr commented on the issue:

https://github.com/apache/flink/pull/5070
  
@twalthr I think my previous unit test method was not correct. Since 
TableAPI always use `java.util.HashMap` per flink documentation. I think it is 
safe to directly use `$leftTerm.equals($rightTerm)` and it should always be 
using `java.util.AbstractMap.equals` implementation. Do you think I should add 
extra check or do you think the `isMap(resultType)` should be suffice? 


> [Table] Map equality check to use entrySet equality
> ---
>
> Key: FLINK-8151
> URL: https://issues.apache.org/jira/browse/FLINK-8151
> Project: Flink
>  Issue Type: Bug
>  Components: Table API & SQL
>Reporter: Rong Rong
>Assignee: Rong Rong
>
> Following up with FLINK-8038. The equality check currently is broken. Plan to 
> support element-wise equality check by always using the base class: 
> "java.util.Map.equals" method.
>  



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5070: [FLINK-8151][table]Map equality check to use entrySet equ...

2017-11-29 Thread walterddr
Github user walterddr commented on the issue:

https://github.com/apache/flink/pull/5070
  
@twalthr I think my previous unit test method was not correct. Since 
TableAPI always use `java.util.HashMap` per flink documentation. I think it is 
safe to directly use `$leftTerm.equals($rightTerm)` and it should always be 
using `java.util.AbstractMap.equals` implementation. Do you think I should add 
extra check or do you think the `isMap(resultType)` should be suffice? 


---


[jira] [Updated] (FLINK-8151) [Table] Map equality check to use entrySet equality

2017-11-29 Thread Rong Rong (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-8151?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rong Rong updated FLINK-8151:
-
Summary: [Table] Map equality check to use entrySet equality  (was: [Table] 
Clean up Map equality check)

> [Table] Map equality check to use entrySet equality
> ---
>
> Key: FLINK-8151
> URL: https://issues.apache.org/jira/browse/FLINK-8151
> Project: Flink
>  Issue Type: Bug
>  Components: Table API & SQL
>Reporter: Rong Rong
>Assignee: Rong Rong
>
> Following up with FLINK-8038. The equality check currently is broken. Plan to 
> support element-wise equality check by always using the base class: 
> "java.util.Map.equals" method.
>  



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (FLINK-8151) [Table] Clean up Map equality check

2017-11-29 Thread Rong Rong (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-8151?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rong Rong updated FLINK-8151:
-
Description: 
Following up with FLINK-8038. The equality check currently is broken. Plan to 
support element-wise equality check by always using the base class: 
"java.util.Map.equals" method.
 

  was:Following up with FLINK-8038. The equality check is not working as Map 
does not support element-wise equality. Suggest to remove it.


> [Table] Clean up Map equality check
> ---
>
> Key: FLINK-8151
> URL: https://issues.apache.org/jira/browse/FLINK-8151
> Project: Flink
>  Issue Type: Bug
>  Components: Table API & SQL
>Reporter: Rong Rong
>Assignee: Rong Rong
>
> Following up with FLINK-8038. The equality check currently is broken. Plan to 
> support element-wise equality check by always using the base class: 
> "java.util.Map.equals" method.
>  



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (FLINK-8151) [Table] Clean up Map equality check

2017-11-29 Thread Rong Rong (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-8151?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rong Rong updated FLINK-8151:
-
Summary: [Table] Clean up Map equality check  (was: [Table] removing map 
value equality check)

> [Table] Clean up Map equality check
> ---
>
> Key: FLINK-8151
> URL: https://issues.apache.org/jira/browse/FLINK-8151
> Project: Flink
>  Issue Type: Bug
>  Components: Table API & SQL
>Reporter: Rong Rong
>Assignee: Rong Rong
>
> Following up with FLINK-8038. The equality check is not working as Map does 
> not support element-wise equality. Suggest to remove it.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-7595) Removing stateless task from task chain breaks savepoint restore

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7595?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271279#comment-16271279
 ] 

ASF GitHub Bot commented on FLINK-7595:
---

GitHub user zentol opened a pull request:

https://github.com/apache/flink/pull/5103

 [FLINK-7595] [Savepoints] Allow removing stateless operators 

This is a fixed version of #4651 for 1.4. It adds some checkstyle fixes and 
modifies the SavepointLoaderTest to actually have a stateful task. The test is 
currently only passing due to the broken behavior.

## What is the purpose of the change

This PR reverts a regression where stateless operators could no longer be 
removed from a job when loading a savepoint without setting the 
`--allowNonRestoredState` flag. The check now explicitly checks whether the 
state of an operator, that could not be mapped to the new program, is empty.

## Brief change log

* Modify `SavepointLoader` to check whether the unmapped state is actually 
empty
* Modify `AbstractOperatorRestoreTestBase` to allow subclasses to set the 
`--allowNonRestoredState` flag
* Add a modified version of `ChainLengthDecreaseTest` to prevent this issue 
from re-emerging.


## Verifying this change

This change added tests and can be verified as follows:

Run `ChainLengthStatelessDecreaseTest`. Alternatively, run the reproducer 
from the JIRA before and after the change.

## Does this pull request potentially affect one of the following parts:

  - Dependencies (does it add or upgrade a dependency): (no)
  - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
  - The serializers: (no)
  - The runtime per-record code paths (performance sensitive): no)
  - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (yes)

## Documentation

  - Does this pull request introduce a new feature? (no)
  - If yes, how is the feature documented? (not applicable)

This should be merged to 1.3 and master. Note that for 1.3 it may be 
necessary to backport the `OperatorSubtaskState#hasState()` method.

@StefanRRichter @uce 


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/zentol/flink 7595_14

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/5103.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #5103


commit 083d43b0365705c6d6355d5609da6b812d3ac909
Author: zentol 
Date:   2017-09-06T13:38:20Z

[FLINK-7595] [Savepoints] Allow removing stateless operators

commit 2e6bad3ee473aeb76c323d826d1f57b51f2968a2
Author: zentol 
Date:   2017-11-29T17:57:51Z

checkstyle

commit d0674e8c135ab195ed91ef0ebbcb82f9e9aec79e
Author: zentol 
Date:   2017-11-29T18:14:18Z

make task stateful in SavepointLoaderTest




> Removing stateless task from task chain breaks savepoint restore
> 
>
> Key: FLINK-7595
> URL: https://issues.apache.org/jira/browse/FLINK-7595
> Project: Flink
>  Issue Type: Bug
>  Components: State Backends, Checkpointing
>Reporter: Ufuk Celebi
>Assignee: Chesnay Schepler
> Attachments: ChainedTaskRemoveTest.java
>
>
> When removing a stateless operator from a 2-task chain where the head 
> operator is stateful breaks savepoint restore with 
> {code}
> Caused by: java.lang.IllegalStateException: Failed to rollback to savepoint 
> /var/folders/py/s_1l8vln6f19ygc77m8c4zhrgn/T/junit1167397515334838028/junit8006766303945373008/savepoint-cb0bcf-3cfa67865ac0.
>  Cannot map savepoint state...
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink pull request #5103: [FLINK-7595] [Savepoints] Allow removing stateless...

2017-11-29 Thread zentol
GitHub user zentol opened a pull request:

https://github.com/apache/flink/pull/5103

 [FLINK-7595] [Savepoints] Allow removing stateless operators 

This is a fixed version of #4651 for 1.4. It adds some checkstyle fixes and 
modifies the SavepointLoaderTest to actually have a stateful task. The test is 
currently only passing due to the broken behavior.

## What is the purpose of the change

This PR reverts a regression where stateless operators could no longer be 
removed from a job when loading a savepoint without setting the 
`--allowNonRestoredState` flag. The check now explicitly checks whether the 
state of an operator, that could not be mapped to the new program, is empty.

## Brief change log

* Modify `SavepointLoader` to check whether the unmapped state is actually 
empty
* Modify `AbstractOperatorRestoreTestBase` to allow subclasses to set the 
`--allowNonRestoredState` flag
* Add a modified version of `ChainLengthDecreaseTest` to prevent this issue 
from re-emerging.


## Verifying this change

This change added tests and can be verified as follows:

Run `ChainLengthStatelessDecreaseTest`. Alternatively, run the reproducer 
from the JIRA before and after the change.

## Does this pull request potentially affect one of the following parts:

  - Dependencies (does it add or upgrade a dependency): (no)
  - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
  - The serializers: (no)
  - The runtime per-record code paths (performance sensitive): no)
  - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (yes)

## Documentation

  - Does this pull request introduce a new feature? (no)
  - If yes, how is the feature documented? (not applicable)

This should be merged to 1.3 and master. Note that for 1.3 it may be 
necessary to backport the `OperatorSubtaskState#hasState()` method.

@StefanRRichter @uce 


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/zentol/flink 7595_14

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/5103.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #5103


commit 083d43b0365705c6d6355d5609da6b812d3ac909
Author: zentol 
Date:   2017-09-06T13:38:20Z

[FLINK-7595] [Savepoints] Allow removing stateless operators

commit 2e6bad3ee473aeb76c323d826d1f57b51f2968a2
Author: zentol 
Date:   2017-11-29T17:57:51Z

checkstyle

commit d0674e8c135ab195ed91ef0ebbcb82f9e9aec79e
Author: zentol 
Date:   2017-11-29T18:14:18Z

make task stateful in SavepointLoaderTest




---


[jira] [Commented] (FLINK-8151) [Table] removing map value equality check

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8151?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271238#comment-16271238
 ] 

ASF GitHub Bot commented on FLINK-8151:
---

Github user walterddr commented on the issue:

https://github.com/apache/flink/pull/5070
  
Ah. I see. `AbstractMap.equals` look very much similar to `Arrays.equals`. 
I will add it in and some more testings. thanks @fhueske for the great tips


> [Table] removing map value equality check
> -
>
> Key: FLINK-8151
> URL: https://issues.apache.org/jira/browse/FLINK-8151
> Project: Flink
>  Issue Type: Bug
>  Components: Table API & SQL
>Reporter: Rong Rong
>Assignee: Rong Rong
>
> Following up with FLINK-8038. The equality check is not working as Map does 
> not support element-wise equality. Suggest to remove it.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5070: [FLINK-8151][table]Remove Map type equality comparison.

2017-11-29 Thread walterddr
Github user walterddr commented on the issue:

https://github.com/apache/flink/pull/5070
  
Ah. I see. `AbstractMap.equals` look very much similar to `Arrays.equals`. 
I will add it in and some more testings. thanks @fhueske for the great tips


---


[jira] [Comment Edited] (FLINK-5506) Java 8 - CommunityDetection.java:158 - java.lang.NullPointerException

2017-11-29 Thread Christos Hadjinikolis (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-5506?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271183#comment-16271183
 ] 

Christos Hadjinikolis edited comment on FLINK-5506 at 11/29/17 5:42 PM:


Same issue here. I am running:

{{{color:red}return {color:#205081}graph{color}.run(new 
CommunityDetection<>(MAX_ITERATIONS_FOR_SEGMENTATION, DELTA)){color}}}

where my *graph *is of type:{{{color:#205081}Graph{color}}}.

The *graph *is generated with: 

{{{color:#14892c}Graph.fromDataSet(env.fromCollection(vertices), 
env.fromCollection(edges), env);{color}}}

where *vertices* are:{{{color:#14892c} List> 
vertices{color}}}
and edges:{{ {color:#14892c}Set> edges{color}}}

It is worth noting that I am intentionally creating a vertices with distinct 
Long ids to create List> vertices. After running the code 
I am getting the below *nullPointerException*: 

org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:933)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40)
at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397)
at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
at 
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
at 
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
at 
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Caused by: java.lang.NullPointerException
at 
*{color:#d04437}org.apache.flink.graph.library.CommunityDetection$VertexLabelUpdater.updateVertex(CommunityDetection.java:158{color}*)
at 
org.apache.flink.graph.spargel.ScatterGatherIteration$GatherUdfSimpleVV.coGroup(ScatterGatherIteration.java:389)
at 
org.apache.flink.runtime.operators.CoGroupWithSolutionSetSecondDriver.run(CoGroupWithSolutionSetSecondDriver.java:218)
at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:490)
at 
org.apache.flink.runtime.iterative.task.AbstractIterativeTask.run(AbstractIterativeTask.java:146)
at 
org.apache.flink.runtime.iterative.task.IterationTailTask.run(IterationTailTask.java:107)
at 
org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:355)
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:702)
at java.lang.Thread.run(Thread.java:745)

Is there a quick fix for this? e.g. initialising vertices directly using their 
ids as both their ids and values as per Vasia's advice? Has anyone solved this? 
Anything else I can try before this is resolved?


was (Author: hadjinik):
Same issue here. I am running:

{{{color:red}return {color:#205081}graph{color}.run(new 
CommunityDetection<>(MAX_ITERATIONS_FOR_SEGMENTATION, DELTA)){color}}}

where my *graph *is of type:{{{color:#205081}Graph{color}}}.

The *graph *is generated with: 

{{{color:#14892c}Graph.fromDataSet(env.fromCollection(vertices), 
env.fromCollection(edges), env);{color}}}

where *vertices* are:{{{color:#14892c} List> 
vertices{color}}}
and edges:{{ {color:#14892c}Set> edges{color}}}

It is worth noting that I am intentionally creating a vertices with distinct 
Long ids to create List> vertices. After running the code 
I am getting the below *nullPointerException*: 

org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:933)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40)
at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397)
at scala.concurrent.forkjoin.ForkJoinTask.doE

[jira] [Comment Edited] (FLINK-5506) Java 8 - CommunityDetection.java:158 - java.lang.NullPointerException

2017-11-29 Thread Christos Hadjinikolis (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-5506?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271183#comment-16271183
 ] 

Christos Hadjinikolis edited comment on FLINK-5506 at 11/29/17 5:38 PM:


Same issue here. I am running:

{{{color:red}return {color:#205081}graph{color}.run(new 
CommunityDetection<>(MAX_ITERATIONS_FOR_SEGMENTATION, DELTA)){color}}}

where my *graph *is of type:{{{color:#205081}Graph{color}}}.

The *graph *is generated with: 

{{{color:#14892c}Graph.fromDataSet(env.fromCollection(vertices), 
env.fromCollection(edges), env);{color}}}

where *vertices* are:{{{color:#14892c} List> 
vertices{color}}}
and edges:{{ {color:#14892c}Set> edges{color}}}

It is worth noting that I am intentionally creating a vertices with distinct 
Long ids to create List> vertices. After running the code 
I am getting the below *nullPointerException*: 

org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:933)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40)
at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397)
at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
at 
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
at 
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
at 
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Caused by: java.lang.NullPointerException
at 
org.apache.flink.graph.library.CommunityDetection$VertexLabelUpdater.updateVertex(CommunityDetection.java:158)
at 
org.apache.flink.graph.spargel.ScatterGatherIteration$GatherUdfSimpleVV.coGroup(ScatterGatherIteration.java:389)
at 
org.apache.flink.runtime.operators.CoGroupWithSolutionSetSecondDriver.run(CoGroupWithSolutionSetSecondDriver.java:218)
at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:490)
at 
org.apache.flink.runtime.iterative.task.AbstractIterativeTask.run(AbstractIterativeTask.java:146)
at 
org.apache.flink.runtime.iterative.task.IterationTailTask.run(IterationTailTask.java:107)
at 
org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:355)
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:702)
at java.lang.Thread.run(Thread.java:745)


was (Author: hadjinik):
Same issue here. I am running:

{{{color:red}return {color:#205081}graph{color}.run(new 
CommunityDetection<>(MAX_ITERATIONS_FOR_SEGMENTATION, DELTA)){color}}}

where my *graph *is of type:{{{color:#205081}Graph{color}}}.

The *graph *is generated with: 

{{{color:#14892c}Graph.fromDataSet(env.fromCollection(vertices), 
env.fromCollection(edges), env);{color}}}

where *vertices* are:{{{color:#14892c} List> 
vertices{color}}}
and edges:{{ {color:#14892c}Set> edges{color}}}

It is worth noting that I am intentionally creating a vertices with distinct 
Long ids to create List> vertices. After running the code 
I am getting the below *nullPointerException*: 
{{org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:933)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40)
at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397)
at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
at 
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
at 
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
at 
scala.concurrent.forkjoin.ForkJoinWorkerThrea

[jira] [Comment Edited] (FLINK-5506) Java 8 - CommunityDetection.java:158 - java.lang.NullPointerException

2017-11-29 Thread Christos Hadjinikolis (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-5506?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271183#comment-16271183
 ] 

Christos Hadjinikolis edited comment on FLINK-5506 at 11/29/17 5:38 PM:


Same issue here. I am running:

{{{color:red}return {color:#205081}graph{color}.run(new 
CommunityDetection<>(MAX_ITERATIONS_FOR_SEGMENTATION, DELTA)){color}}}

where my *graph *is of type:{{{color:#205081}Graph{color}}}.

The *graph *is generated with: 

{{{color:#14892c}Graph.fromDataSet(env.fromCollection(vertices), 
env.fromCollection(edges), env);{color}}}

where *vertices* are:{{{color:#14892c} List> 
vertices{color}}}
and edges:{{ {color:#14892c}Set> edges{color}}}

It is worth noting that I am intentionally creating a vertices with distinct 
Long ids to create List> vertices. After running the code 
I am getting the below *nullPointerException*: 
{{org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:933)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40)
at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397)
at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
at 
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
at 
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
at 
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Caused by: java.lang.NullPointerException
at 
org.apache.flink.graph.library.CommunityDetection$VertexLabelUpdater.updateVertex(CommunityDetection.java:158)
at 
org.apache.flink.graph.spargel.ScatterGatherIteration$GatherUdfSimpleVV.coGroup(ScatterGatherIteration.java:389)
at 
org.apache.flink.runtime.operators.CoGroupWithSolutionSetSecondDriver.run(CoGroupWithSolutionSetSecondDriver.java:218)
at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:490)
at 
org.apache.flink.runtime.iterative.task.AbstractIterativeTask.run(AbstractIterativeTask.java:146)
at 
org.apache.flink.runtime.iterative.task.IterationTailTask.run(IterationTailTask.java:107)
at 
org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:355)
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:702)
at java.lang.Thread.run(Thread.java:745)}}


was (Author: hadjinik):
Same issue here. I am running:

{{{color:red}return {color:#205081}graph{color}.run(new 
CommunityDetection<>(MAX_ITERATIONS_FOR_SEGMENTATION, DELTA)){color}}}

where my *graph *is of type:{{{color:#205081}Graph{color}}}.

The *graph *is generated with: 

{{{color:#14892c}Graph.fromDataSet(env.fromCollection(vertices), 
env.fromCollection(edges), env);{color}}}

where *vertices* are:{{{color:#14892c} List> 
vertices{color}}}
and edges:{{ {color:#14892c}Set> edges{color}}}

It is worth noting that I am intentionally creating a vertices with distinct 
Long ids to create List> vertices. After running the code 
I am getting the below *nullPointerException*: 
{{
org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:933)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40)
at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397)
at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
at 
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
at 
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
at 
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThre

[jira] [Commented] (FLINK-5506) Java 8 - CommunityDetection.java:158 - java.lang.NullPointerException

2017-11-29 Thread Christos Hadjinikolis (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-5506?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271183#comment-16271183
 ] 

Christos Hadjinikolis commented on FLINK-5506:
--

Same issue here. I am running:

{{{color:red}return {color:#205081}graph{color}.run(new 
CommunityDetection<>(MAX_ITERATIONS_FOR_SEGMENTATION, DELTA)){color}}}

where my *graph *is of type:{{{color:#205081}Graph{color}}}.

The *graph *is generated with: 

{{{color:#14892c}Graph.fromDataSet(env.fromCollection(vertices), 
env.fromCollection(edges), env);{color}}}

where *vertices* are:{{{color:#14892c} List> 
vertices{color}}}
and edges:{{ {color:#14892c}Set> edges{color}}}

It is worth noting that I am intentionally creating a vertices with distinct 
Long ids to create List> vertices. After running the code 
I am getting the below *nullPointerException*: 
{{
org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:933)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:876)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
at 
scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40)
at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397)
at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
at 
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
at 
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
at 
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Caused by: java.lang.NullPointerException
at 
org.apache.flink.graph.library.CommunityDetection$VertexLabelUpdater.updateVertex(CommunityDetection.java:158)
at 
org.apache.flink.graph.spargel.ScatterGatherIteration$GatherUdfSimpleVV.coGroup(ScatterGatherIteration.java:389)
at 
org.apache.flink.runtime.operators.CoGroupWithSolutionSetSecondDriver.run(CoGroupWithSolutionSetSecondDriver.java:218)
at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:490)
at 
org.apache.flink.runtime.iterative.task.AbstractIterativeTask.run(AbstractIterativeTask.java:146)
at 
org.apache.flink.runtime.iterative.task.IterationTailTask.run(IterationTailTask.java:107)
at 
org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:355)
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:702)
at java.lang.Thread.run(Thread.java:745)}}


> Java 8 - CommunityDetection.java:158 - java.lang.NullPointerException
> -
>
> Key: FLINK-5506
> URL: https://issues.apache.org/jira/browse/FLINK-5506
> Project: Flink
>  Issue Type: Bug
>  Components: Gelly
>Affects Versions: 1.1.4
>Reporter: Miguel E. Coimbra
>  Labels: easyfix, newbie
>   Original Estimate: 48h
>  Remaining Estimate: 48h
>
> Reporting this here as per Vasia's advice.
> I am having the following problem while trying out the 
> org.apache.flink.graph.library.CommunityDetection algorithm of the Gelly API 
> (Java).
> Specs: JDK 1.8.0_102 x64
> Apache Flink: 1.1.4
> Suppose I have a very small (I tried an example with 38 vertices as well) 
> dataset stored in a tab-separated file 3-vertex.tsv:
> #id1 id2 score
> 010
> 020
> 030
> This is just a central vertex with 3 neighbors (disconnected between 
> themselves).
> I am loading the dataset and executing the algorithm with the following code:
> ---
> // Load the data from the .tsv file.
> final DataSet> edgeTuples = 
> env.readCsvFile(inputPath)
> .fieldDelimiter("\t") // node IDs are separated by spaces
> .ignoreComments("#")  // comments start with "%"
> .types(Long.class, Long.class, Double.class);
> // Generate a graph and add reverse edges (undirected).
> final Graph graph = Graph.fromTupleDataSet(edgeTuples, 
> new MapFunction() {
> private static final long serialVersionUID = 8713516577419451509L;
> public Long map(Long value) {
> return value;
> }
> },
> env).getUndirected();
> // CommunityDetection parameters.
> final double hopAttenuationDelta = 0.5d;
> final int iterationCount = 10;
> // Prepare and

[jira] [Commented] (FLINK-7574) Remove unused dependencies from flink-clients

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7574?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271167#comment-16271167
 ] 

ASF GitHub Bot commented on FLINK-7574:
---

Github user yew1eb commented on a diff in the pull request:

https://github.com/apache/flink/pull/5076#discussion_r153858786
  
--- Diff: pom.xml ---
@@ -891,6 +905,41 @@ under the License.
 


+   
+   org.apache.maven.plugins
+   maven-dependency-plugin
+   3.0.2
--- End diff --

will remove.


> Remove unused dependencies from flink-clients
> -
>
> Key: FLINK-7574
> URL: https://issues.apache.org/jira/browse/FLINK-7574
> Project: Flink
>  Issue Type: Sub-task
>  Components: Build System
>Affects Versions: 1.3.2
> Environment: Apache Maven 3.3.9, Java version: 1.8.0_144
>Reporter: Hai Zhou UTC+8
>Assignee: Hai Zhou UTC+8
>
> [INFO] --- maven-dependency-plugin:2.10:analyze (default-cli) @ 
> flink-clients_2.11 ---
> [WARNING] Used undeclared dependencies found:
> [WARNING]org.scala-lang:scala-library:jar:2.11.11:compile
> [WARNING]com.data-artisans:flakka-actor_2.11:jar:2.3-custom:compile
> [WARNING] Unused declared dependencies found:
> [WARNING]org.hamcrest:hamcrest-all:jar:1.3:test
> [WARNING]org.apache.flink:force-shading:jar:1.4-SNAPSHOT:compile
> [WARNING]org.powermock:powermock-module-junit4:jar:1.6.5:test
> [WARNING]com.google.code.findbugs:jsr305:jar:1.3.9:compile
> [WARNING]log4j:log4j:jar:1.2.17:test
> [WARNING]org.powermock:powermock-api-mockito:jar:1.6.5:test
> [WARNING]org.slf4j:slf4j-log4j12:jar:1.7.7:test



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink pull request #5076: [FLINK-7574][build] POM Cleanup flink-clients

2017-11-29 Thread yew1eb
Github user yew1eb commented on a diff in the pull request:

https://github.com/apache/flink/pull/5076#discussion_r153858786
  
--- Diff: pom.xml ---
@@ -891,6 +905,41 @@ under the License.
 


+   
+   org.apache.maven.plugins
+   maven-dependency-plugin
+   3.0.2
--- End diff --

will remove.


---


[jira] [Commented] (FLINK-7762) Make WikipediaEditsSourceTest a proper test

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7762?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271151#comment-16271151
 ] 

ASF GitHub Bot commented on FLINK-7762:
---

Github user uce commented on the issue:

https://github.com/apache/flink/pull/5102
  
I would be in favour of removing this or moving it to Bahir, but it is 
currently used in a [doc 
example](https://ci.apache.org/projects/flink/flink-docs-release-1.4/quickstart/run_example_quickstart.html).
 Besides that I doubt that this is of much value to users.

If you think it's OK with it, let's merge this for now and think about the 
example thing really needs it.


> Make WikipediaEditsSourceTest a proper test
> ---
>
> Key: FLINK-7762
> URL: https://issues.apache.org/jira/browse/FLINK-7762
> Project: Flink
>  Issue Type: Improvement
>  Components: Streaming Connectors
>Reporter: Aljoscha Krettek
>Assignee: Ufuk Celebi
>Priority: Minor
>
> {{WikipediaEditsSourceTest}} is currently an ITCase even though it's called 
> test. Making it a test reduces runtime and also makes it more stable because 
> we don't run a whole Flink job.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5102: [FLINK-7762, FLINK-8167] Clean up and harden WikipediaEdi...

2017-11-29 Thread uce
Github user uce commented on the issue:

https://github.com/apache/flink/pull/5102
  
I would be in favour of removing this or moving it to Bahir, but it is 
currently used in a [doc 
example](https://ci.apache.org/projects/flink/flink-docs-release-1.4/quickstart/run_example_quickstart.html).
 Besides that I doubt that this is of much value to users.

If you think it's OK with it, let's merge this for now and think about the 
example thing really needs it.


---


[jira] [Commented] (FLINK-7907) Flink Metrics documentation missing Scala examples

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271136#comment-16271136
 ] 

ASF GitHub Bot commented on FLINK-7907:
---

Github user yew1eb commented on the issue:

https://github.com/apache/flink/pull/5071
  
@zentol , thanks for your review. I have updated the PR according to your 
comments.


> Flink Metrics documentation missing Scala examples
> --
>
> Key: FLINK-7907
> URL: https://issues.apache.org/jira/browse/FLINK-7907
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Colin Williams
>Assignee: Hai Zhou UTC+8
>Priority: Minor
>
> The Flink metrics documentation is missing Scala examples for many of the 
> metrics types. To be consistent there should be Scala examples for all the 
> types.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5071: [FLINK-7907][docs] The metrics documentation missing scal...

2017-11-29 Thread yew1eb
Github user yew1eb commented on the issue:

https://github.com/apache/flink/pull/5071
  
@zentol , thanks for your review. I have updated the PR according to your 
comments.


---


[jira] [Commented] (FLINK-7762) Make WikipediaEditsSourceTest a proper test

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7762?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271127#comment-16271127
 ] 

ASF GitHub Bot commented on FLINK-7762:
---

Github user aljoscha commented on the issue:

https://github.com/apache/flink/pull/5102
  
The changes look good! But you comment about needing that IRC channel seem 
valid. If we don't have the test, however, we would have no way of knowing that 
the code at least works. (We could also drop the source completely)


> Make WikipediaEditsSourceTest a proper test
> ---
>
> Key: FLINK-7762
> URL: https://issues.apache.org/jira/browse/FLINK-7762
> Project: Flink
>  Issue Type: Improvement
>  Components: Streaming Connectors
>Reporter: Aljoscha Krettek
>Assignee: Ufuk Celebi
>Priority: Minor
>
> {{WikipediaEditsSourceTest}} is currently an ITCase even though it's called 
> test. Making it a test reduces runtime and also makes it more stable because 
> we don't run a whole Flink job.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-8158) Rowtime window inner join emits late data

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8158?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271126#comment-16271126
 ] 

ASF GitHub Bot commented on FLINK-8158:
---

Github user fhueske commented on the issue:

https://github.com/apache/flink/pull/5094
  
I had a look at `testRowTimeJoinWithCommonBounds2`. 

The test processes starts by processing two watermarks for `6000` for the 
first and second input. After that three records are added for `6000` (first), 
`6000` (second), and `1000` (first). All three records are late because their 
timestamp is smaller or equal to the current watermark. Therefore, the join 
result should also be late. If all three records would be aligned with the 
watermarks, the result should not be late.


> Rowtime window inner join emits late data
> -
>
> Key: FLINK-8158
> URL: https://issues.apache.org/jira/browse/FLINK-8158
> Project: Flink
>  Issue Type: Bug
>  Components: Table API & SQL
>Reporter: Hequn Cheng
>Assignee: Hequn Cheng
> Attachments: screenshot-1xxx.png
>
>
> When executing the join, the join operator needs to make sure that no late 
> data is emitted. Currently, this achieved by holding back watermarks. 
> However, the window border is not handled correctly. For the sql bellow: 
> {quote}
> val sqlQuery =
>   """
> SELECT t2.key, t2.id, t1.id
> FROM T1 as t1 join T2 as t2 ON
>   t1.key = t2.key AND
>   t1.rt BETWEEN t2.rt - INTERVAL '5' SECOND AND
> t2.rt + INTERVAL '1' SECOND
> """.stripMargin
> val data1 = new mutable.MutableList[(String, String, Long)]
> // for boundary test
> data1.+=(("A", "LEFT1", 6000L))
> val data2 = new mutable.MutableList[(String, String, Long)]
> data2.+=(("A", "RIGHT1", 6000L))
> {quote}
> Join will output a watermark with timestamp 1000, but if left comes with 
> another data ("A", "LEFT1", 1000L), join will output a record with timestamp 
> 1000 which equals previous watermark.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (FLINK-8172) Remove unnecessary synchronisation in RecordSerializer

2017-11-29 Thread Piotr Nowojski (JIRA)
Piotr Nowojski created FLINK-8172:
-

 Summary: Remove unnecessary synchronisation in RecordSerializer
 Key: FLINK-8172
 URL: https://issues.apache.org/jira/browse/FLINK-8172
 Project: Flink
  Issue Type: Improvement
  Components: Network
Affects Versions: 1.3.2, 1.4.0
Reporter: Piotr Nowojski
Assignee: Piotr Nowojski
 Fix For: 1.5.0


While writing the records, RecordSerializer is the only owner of the `Buffer` 
into which data are written. Yet we are synchronisation twice per record while 
accessing MemorySegment. Removing this synchronisation speeds up the Network 
throughput in point to point benchmark by a factor of two (from 
~12500records/ms up to 23000 records/ms).



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5102: [FLINK-7762, FLINK-8167] Clean up and harden WikipediaEdi...

2017-11-29 Thread aljoscha
Github user aljoscha commented on the issue:

https://github.com/apache/flink/pull/5102
  
The changes look good! But you comment about needing that IRC channel seem 
valid. If we don't have the test, however, we would have no way of knowing that 
the code at least works. (We could also drop the source completely)


---


[GitHub] flink issue #5094: [FLINK-8158] [table] Fix rowtime window inner join emits ...

2017-11-29 Thread fhueske
Github user fhueske commented on the issue:

https://github.com/apache/flink/pull/5094
  
I had a look at `testRowTimeJoinWithCommonBounds2`. 

The test processes starts by processing two watermarks for `6000` for the 
first and second input. After that three records are added for `6000` (first), 
`6000` (second), and `1000` (first). All three records are late because their 
timestamp is smaller or equal to the current watermark. Therefore, the join 
result should also be late. If all three records would be aligned with the 
watermarks, the result should not be late.


---


[GitHub] flink issue #4985: [FLINK-8027] Generalize existing rest handlers to work wi...

2017-11-29 Thread tillrohrmann
Github user tillrohrmann commented on the issue:

https://github.com/apache/flink/pull/4985
  
Thanks for the review @shuai-xu. Will rebase it and once Travis gives green 
light, I'll merge it.


---


[jira] [Commented] (FLINK-8027) Generalize existing rest handlers to work with arbitrary RestfulGateway

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271117#comment-16271117
 ] 

ASF GitHub Bot commented on FLINK-8027:
---

Github user tillrohrmann commented on a diff in the pull request:

https://github.com/apache/flink/pull/4985#discussion_r153852356
  
--- Diff: 
flink-runtime/src/test/java/org/apache/flink/runtime/taskexecutor/TestingTaskExecutorGateway.java
 ---
@@ -0,0 +1,124 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.taskexecutor;
+
+import org.apache.flink.api.common.JobID;
+import org.apache.flink.api.common.time.Time;
+import org.apache.flink.runtime.checkpoint.CheckpointOptions;
+import org.apache.flink.runtime.clusterframework.types.AllocationID;
+import org.apache.flink.runtime.clusterframework.types.ResourceID;
+import org.apache.flink.runtime.clusterframework.types.SlotID;
+import org.apache.flink.runtime.deployment.TaskDeploymentDescriptor;
+import org.apache.flink.runtime.executiongraph.ExecutionAttemptID;
+import org.apache.flink.runtime.executiongraph.PartitionInfo;
+import org.apache.flink.runtime.jobmaster.JobMasterId;
+import org.apache.flink.runtime.messages.Acknowledge;
+import org.apache.flink.runtime.resourcemanager.ResourceManagerId;
+import org.apache.flink.util.Preconditions;
+
+import java.util.concurrent.CompletableFuture;
+
+/**
+ * Simple {@link TaskExecutorGateway} implementation for testing purposes.
+ */
+public class TestingTaskExecutorGateway implements TaskExecutorGateway {
+
+   private final String address;
+
+   private final String hostname;
+
+   public TestingTaskExecutorGateway() {
+   this("foobar:1234", "foobar");
+   }
+
+   public TestingTaskExecutorGateway(String address, String hostname) {
+   this.address = Preconditions.checkNotNull(address);
+   this.hostname = Preconditions.checkNotNull(hostname);
+   }
+
+   @Override
+   public CompletableFuture requestSlot(SlotID slotId, JobID 
jobId, AllocationID allocationId, String targetAddress, ResourceManagerId 
resourceManagerId, Time timeout) {
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public CompletableFuture 
submitTask(TaskDeploymentDescriptor tdd, JobMasterId jobMasterId, Time timeout) 
{
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public CompletableFuture 
updatePartitions(ExecutionAttemptID executionAttemptID, Iterable 
partitionInfos, Time timeout) {
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public void failPartition(ExecutionAttemptID executionAttemptID) {
+   // noop
+   }
+
+   @Override
+   public CompletableFuture 
triggerCheckpoint(ExecutionAttemptID executionAttemptID, long checkpointID, 
long checkpointTimestamp, CheckpointOptions checkpointOptions) {
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public CompletableFuture 
confirmCheckpoint(ExecutionAttemptID executionAttemptID, long checkpointId, 
long checkpointTimestamp) {
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public CompletableFuture stopTask(ExecutionAttemptID 
executionAttemptID, Time timeout) {
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public CompletableFuture cancelTask(ExecutionAttemptID 
executionAttemptID, Time timeout) {
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public void heartbeatFromJobManager(ResourceID heartbeatOrigin) {
+   // noop
+   }
+
+   @Override

[jira] [Commented] (FLINK-8027) Generalize existing rest handlers to work with arbitrary RestfulGateway

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271118#comment-16271118
 ] 

ASF GitHub Bot commented on FLINK-8027:
---

Github user tillrohrmann commented on the issue:

https://github.com/apache/flink/pull/4985
  
Thanks for the review @shuai-xu. Will rebase it and once Travis gives green 
light, I'll merge it.


> Generalize existing rest handlers to work with arbitrary RestfulGateway
> ---
>
> Key: FLINK-8027
> URL: https://issues.apache.org/jira/browse/FLINK-8027
> Project: Flink
>  Issue Type: Sub-task
>  Components: REST
>Affects Versions: 1.5.0
>Reporter: Till Rohrmann
>Assignee: Till Rohrmann
>  Labels: flip-6
> Fix For: 1.5.0
>
>
> In order to reuse the existing {{AbstractRestHandler}} we should refactor 
> them such that they work with arbitrary {{RestfulGateway}}.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink pull request #4985: [FLINK-8027] Generalize existing rest handlers to ...

2017-11-29 Thread tillrohrmann
Github user tillrohrmann commented on a diff in the pull request:

https://github.com/apache/flink/pull/4985#discussion_r153852356
  
--- Diff: 
flink-runtime/src/test/java/org/apache/flink/runtime/taskexecutor/TestingTaskExecutorGateway.java
 ---
@@ -0,0 +1,124 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.taskexecutor;
+
+import org.apache.flink.api.common.JobID;
+import org.apache.flink.api.common.time.Time;
+import org.apache.flink.runtime.checkpoint.CheckpointOptions;
+import org.apache.flink.runtime.clusterframework.types.AllocationID;
+import org.apache.flink.runtime.clusterframework.types.ResourceID;
+import org.apache.flink.runtime.clusterframework.types.SlotID;
+import org.apache.flink.runtime.deployment.TaskDeploymentDescriptor;
+import org.apache.flink.runtime.executiongraph.ExecutionAttemptID;
+import org.apache.flink.runtime.executiongraph.PartitionInfo;
+import org.apache.flink.runtime.jobmaster.JobMasterId;
+import org.apache.flink.runtime.messages.Acknowledge;
+import org.apache.flink.runtime.resourcemanager.ResourceManagerId;
+import org.apache.flink.util.Preconditions;
+
+import java.util.concurrent.CompletableFuture;
+
+/**
+ * Simple {@link TaskExecutorGateway} implementation for testing purposes.
+ */
+public class TestingTaskExecutorGateway implements TaskExecutorGateway {
+
+   private final String address;
+
+   private final String hostname;
+
+   public TestingTaskExecutorGateway() {
+   this("foobar:1234", "foobar");
+   }
+
+   public TestingTaskExecutorGateway(String address, String hostname) {
+   this.address = Preconditions.checkNotNull(address);
+   this.hostname = Preconditions.checkNotNull(hostname);
+   }
+
+   @Override
+   public CompletableFuture requestSlot(SlotID slotId, JobID 
jobId, AllocationID allocationId, String targetAddress, ResourceManagerId 
resourceManagerId, Time timeout) {
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public CompletableFuture 
submitTask(TaskDeploymentDescriptor tdd, JobMasterId jobMasterId, Time timeout) 
{
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public CompletableFuture 
updatePartitions(ExecutionAttemptID executionAttemptID, Iterable 
partitionInfos, Time timeout) {
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public void failPartition(ExecutionAttemptID executionAttemptID) {
+   // noop
+   }
+
+   @Override
+   public CompletableFuture 
triggerCheckpoint(ExecutionAttemptID executionAttemptID, long checkpointID, 
long checkpointTimestamp, CheckpointOptions checkpointOptions) {
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public CompletableFuture 
confirmCheckpoint(ExecutionAttemptID executionAttemptID, long checkpointId, 
long checkpointTimestamp) {
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public CompletableFuture stopTask(ExecutionAttemptID 
executionAttemptID, Time timeout) {
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public CompletableFuture cancelTask(ExecutionAttemptID 
executionAttemptID, Time timeout) {
+   return CompletableFuture.completedFuture(Acknowledge.get());
+   }
+
+   @Override
+   public void heartbeatFromJobManager(ResourceID heartbeatOrigin) {
+   // noop
+   }
+
+   @Override
+   public void heartbeatFromResourceManager(ResourceID heartbeatOrigin) {
+   // noop
+   }
+
+   @Override
+   public void disconnectJobManager(JobID jobId, Exception cause) {
+   // nooop
--- End diff --


[jira] [Commented] (FLINK-7652) Port CurrentJobIdsHandler to new REST endpoint

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7652?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271116#comment-16271116
 ] 

ASF GitHub Bot commented on FLINK-7652:
---

Github user asfgit closed the pull request at:

https://github.com/apache/flink/pull/4734


> Port CurrentJobIdsHandler to new REST endpoint
> --
>
> Key: FLINK-7652
> URL: https://issues.apache.org/jira/browse/FLINK-7652
> Project: Flink
>  Issue Type: Sub-task
>  Components: REST, Webfrontend
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Tzu-Li (Gordon) Tai
>  Labels: flip-6
> Fix For: 1.5.0
>
>
> Port existing {{CurrentJobIdsHandler}} to new REST endpoint



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink pull request #4734: [FLINK-7652] [flip6] Port CurrentJobIdsHandler to ...

2017-11-29 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/flink/pull/4734


---


[jira] [Resolved] (FLINK-7652) Port CurrentJobIdsHandler to new REST endpoint

2017-11-29 Thread Till Rohrmann (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-7652?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Till Rohrmann resolved FLINK-7652.
--
Resolution: Done

Added via 67aad88ee025ce02053ab560f2504762f53b87d9

> Port CurrentJobIdsHandler to new REST endpoint
> --
>
> Key: FLINK-7652
> URL: https://issues.apache.org/jira/browse/FLINK-7652
> Project: Flink
>  Issue Type: Sub-task
>  Components: REST, Webfrontend
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Tzu-Li (Gordon) Tai
>  Labels: flip-6
> Fix For: 1.5.0
>
>
> Port existing {{CurrentJobIdsHandler}} to new REST endpoint



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5100: [FLINK-4812][metrics] Expose currentLowWatermark for all ...

2017-11-29 Thread aljoscha
Github user aljoscha commented on the issue:

https://github.com/apache/flink/pull/5100
  
I think the fact the the operator needs this in `setup()` is already 
problematic, the latency metrics implementation shouldn't have been merged like 
this. Wouldn't we need the output gauge only in `OperatorChain` because that 
creates the `Output`s that operators use to emit watermarks?


---


[jira] [Commented] (FLINK-4812) Report Watermark metrics in all operators

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271100#comment-16271100
 ] 

ASF GitHub Bot commented on FLINK-4812:
---

Github user aljoscha commented on the issue:

https://github.com/apache/flink/pull/5100
  
I think the fact the the operator needs this in `setup()` is already 
problematic, the latency metrics implementation shouldn't have been merged like 
this. Wouldn't we need the output gauge only in `OperatorChain` because that 
creates the `Output`s that operators use to emit watermarks?


> Report Watermark metrics in all operators
> -
>
> Key: FLINK-4812
> URL: https://issues.apache.org/jira/browse/FLINK-4812
> Project: Flink
>  Issue Type: Improvement
>  Components: Metrics
>Reporter: Robert Metzger
>Assignee: Chesnay Schepler
>Priority: Critical
> Fix For: 1.5.0
>
>
> As reported by a user, Flink does currently not export the current low 
> watermark for sources 
> (http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/currentLowWatermark-metric-not-reported-for-all-tasks-td13770.html).
> This JIRA is for adding such a metric for the sources as well.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-8171) Remove work arounds in Flip6LocalStreamEnvironment

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8171?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271097#comment-16271097
 ] 

ASF GitHub Bot commented on FLINK-8171:
---

Github user tillrohrmann commented on the issue:

https://github.com/apache/flink/pull/5101
  
I think we can remove the comment but not enabling the queued scheduling 
mode which is mandatory for Flip-6 to work. I'll remove the comment.


> Remove work arounds in Flip6LocalStreamEnvironment
> --
>
> Key: FLINK-8171
> URL: https://issues.apache.org/jira/browse/FLINK-8171
> Project: Flink
>  Issue Type: Improvement
>  Components: Tests
>Affects Versions: 1.5.0
>Reporter: Till Rohrmann
>Assignee: Till Rohrmann
>  Labels: flip-6
> Fix For: 1.5.0
>
>
> After adding FLINK-7956, it is no longer necessary that the 
> {{Flip6LocalStreamEnvironment}} waits for the registration of TaskManagers 
> before submitting a job. Moreover, it is also possible to use slot sharing 
> when submitting jobs.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5101: [FLINK-8171] [flip6] Remove work arounds from Flip6LocalS...

2017-11-29 Thread tillrohrmann
Github user tillrohrmann commented on the issue:

https://github.com/apache/flink/pull/5101
  
I think we can remove the comment but not enabling the queued scheduling 
mode which is mandatory for Flip-6 to work. I'll remove the comment.


---


[GitHub] flink pull request #5102: [FLINK-7762, FLINK-8167] Clean up and harden Wikip...

2017-11-29 Thread uce
GitHub user uce opened a pull request:

https://github.com/apache/flink/pull/5102

[FLINK-7762, FLINK-8167] Clean up and harden WikipediaEditsSource

## What is the purpose of the change

This pull requests addresses two related issues with the 
WikipediaEditsSource. It makes the WikipediaEditsSourceTest a proper test 
instead of unnecessarily starting a FlinkMiniCluster and addresses a potential 
test instability.

In general, the WikipediaEditsSource is not in good shape and could benefit 
from further refactoring. One potential area for improvement is integration 
with the asynchronous channel listener that reports events like errors or being 
kicked out of a channel, etc.

I did not do this due to time constraints and the fact that this is not a 
production source. In general, it is questionable whether we should keep the 
test as is or remove it since it depends on connectivity to an IRC channel.

## Brief change log
- Harden WikipediaEditsSource with eager sanity checks
- Make WikipediaEditsSourceTest proper test

## Verifying this change

This change is a rework/code cleanup without any new test coverage.

## Does this pull request potentially affect one of the following parts:

- Dependencies (does it add or upgrade a dependency): yes, but only to 
`flink-test-utils-junit`
- The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
- The serializers: no
- The runtime per-record code paths (performance sensitive): no
- Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: no
- The S3 file system connector: no

## Documentation

- Does this pull request introduce a new feature? no
- If yes, how is the feature documented? not applicable


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/uce/flink 7762-8167-wikiedits

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/5102.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #5102


commit b2ab66f05ce545214a8132dc2d46b3143939b015
Author: Ufuk Celebi 
Date:   2017-11-29T15:28:18Z

[FLINK-8167] [connector-wikiedits] Harden WikipediaEditsSource

- Minor eager sanity checks
- Use UUID suffix for nickname. As reported in FLINK-8167, the current
  nickname suffix can result in nickname clashes which lead to test
  failures.

commit 06ec1542963bbe2afaf1ad1fd55a54d13f855304
Author: Ufuk Celebi 
Date:   2017-11-29T15:36:29Z

[FLINK-7762] [connector-wikiedits] Make WikipediaEditsSourceTest proper test

The WikipediaEditsSourceTest unnecessarily implements an integration
test that starts a FlinkMiniCluster and executes a small Flink program.

This simply creates a source and executes run in a separate thread until
a single WikipediaEditEvent is received.




---


[jira] [Commented] (FLINK-7762) Make WikipediaEditsSourceTest a proper test

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7762?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271083#comment-16271083
 ] 

ASF GitHub Bot commented on FLINK-7762:
---

GitHub user uce opened a pull request:

https://github.com/apache/flink/pull/5102

[FLINK-7762, FLINK-8167] Clean up and harden WikipediaEditsSource

## What is the purpose of the change

This pull requests addresses two related issues with the 
WikipediaEditsSource. It makes the WikipediaEditsSourceTest a proper test 
instead of unnecessarily starting a FlinkMiniCluster and addresses a potential 
test instability.

In general, the WikipediaEditsSource is not in good shape and could benefit 
from further refactoring. One potential area for improvement is integration 
with the asynchronous channel listener that reports events like errors or being 
kicked out of a channel, etc.

I did not do this due to time constraints and the fact that this is not a 
production source. In general, it is questionable whether we should keep the 
test as is or remove it since it depends on connectivity to an IRC channel.

## Brief change log
- Harden WikipediaEditsSource with eager sanity checks
- Make WikipediaEditsSourceTest proper test

## Verifying this change

This change is a rework/code cleanup without any new test coverage.

## Does this pull request potentially affect one of the following parts:

- Dependencies (does it add or upgrade a dependency): yes, but only to 
`flink-test-utils-junit`
- The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
- The serializers: no
- The runtime per-record code paths (performance sensitive): no
- Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: no
- The S3 file system connector: no

## Documentation

- Does this pull request introduce a new feature? no
- If yes, how is the feature documented? not applicable


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/uce/flink 7762-8167-wikiedits

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/5102.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #5102


commit b2ab66f05ce545214a8132dc2d46b3143939b015
Author: Ufuk Celebi 
Date:   2017-11-29T15:28:18Z

[FLINK-8167] [connector-wikiedits] Harden WikipediaEditsSource

- Minor eager sanity checks
- Use UUID suffix for nickname. As reported in FLINK-8167, the current
  nickname suffix can result in nickname clashes which lead to test
  failures.

commit 06ec1542963bbe2afaf1ad1fd55a54d13f855304
Author: Ufuk Celebi 
Date:   2017-11-29T15:36:29Z

[FLINK-7762] [connector-wikiedits] Make WikipediaEditsSourceTest proper test

The WikipediaEditsSourceTest unnecessarily implements an integration
test that starts a FlinkMiniCluster and executes a small Flink program.

This simply creates a source and executes run in a separate thread until
a single WikipediaEditEvent is received.




> Make WikipediaEditsSourceTest a proper test
> ---
>
> Key: FLINK-7762
> URL: https://issues.apache.org/jira/browse/FLINK-7762
> Project: Flink
>  Issue Type: Improvement
>  Components: Streaming Connectors
>Reporter: Aljoscha Krettek
>Assignee: Ufuk Celebi
>Priority: Minor
>
> {{WikipediaEditsSourceTest}} is currently an ITCase even though it's called 
> test. Making it a test reduces runtime and also makes it more stable because 
> we don't run a whole Flink job.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-8148) Test instability in YarnFileStageTest

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271074#comment-16271074
 ] 

ASF GitHub Bot commented on FLINK-8148:
---

Github user NicoK commented on the issue:

https://github.com/apache/flink/pull/5066
  
oh, sorry for that


> Test instability in YarnFileStageTest
> -
>
> Key: FLINK-8148
> URL: https://issues.apache.org/jira/browse/FLINK-8148
> Project: Flink
>  Issue Type: Bug
>  Components: Tests, YARN
>Affects Versions: 1.4.0
>Reporter: Nico Kruber
>Assignee: Nico Kruber
>Priority: Critical
>  Labels: test-stability
>
> {code}
> Running org.apache.flink.yarn.YarnFileStageTestS3ITCase
> Tests run: 3, Failures: 1, Errors: 0, Skipped: 1, Time elapsed: 13.152 sec 
> <<< FAILURE! - in org.apache.flink.yarn.YarnFileStageTestS3ITCase
> testRecursiveUploadForYarnS3(org.apache.flink.yarn.YarnFileStageTestS3ITCase) 
>  Time elapsed: 8.515 sec  <<< FAILURE!
> java.lang.AssertionError: null
>   at org.junit.Assert.fail(Assert.java:86)
>   at org.junit.Assert.assertTrue(Assert.java:41)
>   at org.junit.Assert.assertFalse(Assert.java:64)
>   at org.junit.Assert.assertFalse(Assert.java:74)
>   at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarn(YarnFileStageTestS3ITCase.java:171)
>   at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarnS3(YarnFileStageTestS3ITCase.java:192)
> {code}
> from https://travis-ci.org/apache/flink/jobs/305861539
> {{YarnFileStageTestS3ITCase.testRecursiveUploadForYarn}} verifies that the 
> test directory used is cleaned up by 
> {{YarnFileStageTest.testCopyFromLocalRecursive}} which should clean up the 
> directory (in a {{finally}} block). However, for S3, we may not always see 
> our own deletes.
> Quoting from https://aws.amazon.com/s3/faqs/ here:
> {quote}Q: What data consistency model does Amazon S3 employ?
> Amazon S3 buckets in all Regions provide read-after-write consistency for 
> PUTS of new objects and eventual consistency for overwrite PUTS and 
> DELETES.{quote}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5066: [FLINK-8148][yarn/s3] fix test instability in YarnFileSta...

2017-11-29 Thread NicoK
Github user NicoK commented on the issue:

https://github.com/apache/flink/pull/5066
  
oh, sorry for that


---


[jira] [Commented] (FLINK-8149) Replace usages of deprecated SerializationSchema

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8149?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271050#comment-16271050
 ] 

ASF GitHub Bot commented on FLINK-8149:
---

Github user yew1eb commented on the issue:

https://github.com/apache/flink/pull/5069
  
@tzulitai, thanks for your review. I have updated the PR according to your 
comments.


> Replace usages of deprecated SerializationSchema
> 
>
> Key: FLINK-8149
> URL: https://issues.apache.org/jira/browse/FLINK-8149
> Project: Flink
>  Issue Type: Improvement
>  Components: Kinesis Connector
>Affects Versions: 1.4.0
>Reporter: Hai Zhou UTC+8
>Assignee: Hai Zhou UTC+8
> Fix For: 1.5.0
>
>
> The deprecated {{SerializationSchema}}  in {{flink-streaming-java}}, has been 
> moved to {{flink-core}}.
> But, the deprecate {{SerializationSchema}} is still used in 
> {{flink-connector-kinesis}}.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5069: [FLINK-8149][kinesis] Replace usages of deprecated Serial...

2017-11-29 Thread yew1eb
Github user yew1eb commented on the issue:

https://github.com/apache/flink/pull/5069
  
@tzulitai, thanks for your review. I have updated the PR according to your 
comments.


---


[jira] [Commented] (FLINK-8149) Replace usages of deprecated SerializationSchema

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8149?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271046#comment-16271046
 ] 

ASF GitHub Bot commented on FLINK-8149:
---

Github user yew1eb commented on a diff in the pull request:

https://github.com/apache/flink/pull/5069#discussion_r153841399
  
--- Diff: 
flink-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java
 ---
@@ -41,7 +42,6 @@
 import 
org.apache.flink.streaming.connectors.kinesis.serialization.KinesisDeserializationSchema;
 import 
org.apache.flink.streaming.connectors.kinesis.serialization.KinesisDeserializationSchemaWrapper;
 import 
org.apache.flink.streaming.connectors.kinesis.util.KinesisConfigUtil;
-import org.apache.flink.streaming.util.serialization.DeserializationSchema;
--- End diff --

make sense.


> Replace usages of deprecated SerializationSchema
> 
>
> Key: FLINK-8149
> URL: https://issues.apache.org/jira/browse/FLINK-8149
> Project: Flink
>  Issue Type: Improvement
>  Components: Kinesis Connector
>Affects Versions: 1.4.0
>Reporter: Hai Zhou UTC+8
>Assignee: Hai Zhou UTC+8
> Fix For: 1.5.0
>
>
> The deprecated {{SerializationSchema}}  in {{flink-streaming-java}}, has been 
> moved to {{flink-core}}.
> But, the deprecate {{SerializationSchema}} is still used in 
> {{flink-connector-kinesis}}.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink pull request #5069: [FLINK-8149][kinesis] Replace usages of deprecated...

2017-11-29 Thread yew1eb
Github user yew1eb commented on a diff in the pull request:

https://github.com/apache/flink/pull/5069#discussion_r153841399
  
--- Diff: 
flink-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java
 ---
@@ -41,7 +42,6 @@
 import 
org.apache.flink.streaming.connectors.kinesis.serialization.KinesisDeserializationSchema;
 import 
org.apache.flink.streaming.connectors.kinesis.serialization.KinesisDeserializationSchemaWrapper;
 import 
org.apache.flink.streaming.connectors.kinesis.util.KinesisConfigUtil;
-import org.apache.flink.streaming.util.serialization.DeserializationSchema;
--- End diff --

make sense.


---


[jira] [Commented] (FLINK-7907) Flink Metrics documentation missing Scala examples

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271043#comment-16271043
 ] 

ASF GitHub Bot commented on FLINK-7907:
---

Github user zentol commented on a diff in the pull request:

https://github.com/apache/flink/pull/5071#discussion_r153839062
  
--- Diff: docs/monitoring/metrics.md ---
@@ -61,9 +63,35 @@ public class MyMapper extends RichMapFunction {
 }
 
 {% endhighlight %}
+
+
+
+{% highlight scala %}
+
+public class MyMapper extends RichMapFunction[String,String] {
+  var counter: Counter
--- End diff --

reasonable sure this must be lazy (also applies to remaining examples)


> Flink Metrics documentation missing Scala examples
> --
>
> Key: FLINK-7907
> URL: https://issues.apache.org/jira/browse/FLINK-7907
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Colin Williams
>Assignee: Hai Zhou UTC+8
>Priority: Minor
>
> The Flink metrics documentation is missing Scala examples for many of the 
> metrics types. To be consistent there should be Scala examples for all the 
> types.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-7907) Flink Metrics documentation missing Scala examples

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271042#comment-16271042
 ] 

ASF GitHub Bot commented on FLINK-7907:
---

Github user zentol commented on a diff in the pull request:

https://github.com/apache/flink/pull/5071#discussion_r153838811
  
--- Diff: docs/monitoring/metrics.md ---
@@ -61,9 +63,35 @@ public class MyMapper extends RichMapFunction {
 }
 
 {% endhighlight %}
+
+
+
+{% highlight scala %}
+
+public class MyMapper extends RichMapFunction[String,String] {
--- End diff --

remove public keyword (also applies to remaining examples)


> Flink Metrics documentation missing Scala examples
> --
>
> Key: FLINK-7907
> URL: https://issues.apache.org/jira/browse/FLINK-7907
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Colin Williams
>Assignee: Hai Zhou UTC+8
>Priority: Minor
>
> The Flink metrics documentation is missing Scala examples for many of the 
> metrics types. To be consistent there should be Scala examples for all the 
> types.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink pull request #5071: [FLINK-7907][docs] The metrics documentation missi...

2017-11-29 Thread zentol
Github user zentol commented on a diff in the pull request:

https://github.com/apache/flink/pull/5071#discussion_r153839062
  
--- Diff: docs/monitoring/metrics.md ---
@@ -61,9 +63,35 @@ public class MyMapper extends RichMapFunction {
 }
 
 {% endhighlight %}
+
+
+
+{% highlight scala %}
+
+public class MyMapper extends RichMapFunction[String,String] {
+  var counter: Counter
--- End diff --

reasonable sure this must be lazy (also applies to remaining examples)


---


[GitHub] flink pull request #5071: [FLINK-7907][docs] The metrics documentation missi...

2017-11-29 Thread zentol
Github user zentol commented on a diff in the pull request:

https://github.com/apache/flink/pull/5071#discussion_r153838811
  
--- Diff: docs/monitoring/metrics.md ---
@@ -61,9 +63,35 @@ public class MyMapper extends RichMapFunction {
 }
 
 {% endhighlight %}
+
+
+
+{% highlight scala %}
+
+public class MyMapper extends RichMapFunction[String,String] {
--- End diff --

remove public keyword (also applies to remaining examples)


---


[jira] [Commented] (FLINK-8158) Rowtime window inner join emits late data

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8158?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271030#comment-16271030
 ] 

ASF GitHub Bot commented on FLINK-8158:
---

Github user xccui commented on the issue:

https://github.com/apache/flink/pull/5094
  
Hi @hequn8128, as @fhueske explained, watermarks should not define the 
semantics of the join. We just emit join results, in spite of whether they are 
late and the join function should not care about the lateness. You can take 
"holding back watermarks" as a special watermark assigner, based on which, 
there could be out-of-ordered data (join results) being taken as late. 
Currently, we just generates watermarks with a static delay that equals to the 
larger relative offset. In the future, we should dynamically decide whether a 
join result is late according to the real cache size.


> Rowtime window inner join emits late data
> -
>
> Key: FLINK-8158
> URL: https://issues.apache.org/jira/browse/FLINK-8158
> Project: Flink
>  Issue Type: Bug
>  Components: Table API & SQL
>Reporter: Hequn Cheng
>Assignee: Hequn Cheng
> Attachments: screenshot-1xxx.png
>
>
> When executing the join, the join operator needs to make sure that no late 
> data is emitted. Currently, this achieved by holding back watermarks. 
> However, the window border is not handled correctly. For the sql bellow: 
> {quote}
> val sqlQuery =
>   """
> SELECT t2.key, t2.id, t1.id
> FROM T1 as t1 join T2 as t2 ON
>   t1.key = t2.key AND
>   t1.rt BETWEEN t2.rt - INTERVAL '5' SECOND AND
> t2.rt + INTERVAL '1' SECOND
> """.stripMargin
> val data1 = new mutable.MutableList[(String, String, Long)]
> // for boundary test
> data1.+=(("A", "LEFT1", 6000L))
> val data2 = new mutable.MutableList[(String, String, Long)]
> data2.+=(("A", "RIGHT1", 6000L))
> {quote}
> Join will output a watermark with timestamp 1000, but if left comes with 
> another data ("A", "LEFT1", 1000L), join will output a record with timestamp 
> 1000 which equals previous watermark.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5094: [FLINK-8158] [table] Fix rowtime window inner join emits ...

2017-11-29 Thread xccui
Github user xccui commented on the issue:

https://github.com/apache/flink/pull/5094
  
Hi @hequn8128, as @fhueske explained, watermarks should not define the 
semantics of the join. We just emit join results, in spite of whether they are 
late and the join function should not care about the lateness. You can take 
"holding back watermarks" as a special watermark assigner, based on which, 
there could be out-of-ordered data (join results) being taken as late. 
Currently, we just generates watermarks with a static delay that equals to the 
larger relative offset. In the future, we should dynamically decide whether a 
join result is late according to the real cache size.


---


[jira] [Commented] (FLINK-8148) Test instability in YarnFileStageTest

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271028#comment-16271028
 ] 

ASF GitHub Bot commented on FLINK-8148:
---

Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5066
  
will fix the checkstyle violation on the way.


> Test instability in YarnFileStageTest
> -
>
> Key: FLINK-8148
> URL: https://issues.apache.org/jira/browse/FLINK-8148
> Project: Flink
>  Issue Type: Bug
>  Components: Tests, YARN
>Affects Versions: 1.4.0
>Reporter: Nico Kruber
>Assignee: Nico Kruber
>Priority: Critical
>  Labels: test-stability
>
> {code}
> Running org.apache.flink.yarn.YarnFileStageTestS3ITCase
> Tests run: 3, Failures: 1, Errors: 0, Skipped: 1, Time elapsed: 13.152 sec 
> <<< FAILURE! - in org.apache.flink.yarn.YarnFileStageTestS3ITCase
> testRecursiveUploadForYarnS3(org.apache.flink.yarn.YarnFileStageTestS3ITCase) 
>  Time elapsed: 8.515 sec  <<< FAILURE!
> java.lang.AssertionError: null
>   at org.junit.Assert.fail(Assert.java:86)
>   at org.junit.Assert.assertTrue(Assert.java:41)
>   at org.junit.Assert.assertFalse(Assert.java:64)
>   at org.junit.Assert.assertFalse(Assert.java:74)
>   at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarn(YarnFileStageTestS3ITCase.java:171)
>   at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarnS3(YarnFileStageTestS3ITCase.java:192)
> {code}
> from https://travis-ci.org/apache/flink/jobs/305861539
> {{YarnFileStageTestS3ITCase.testRecursiveUploadForYarn}} verifies that the 
> test directory used is cleaned up by 
> {{YarnFileStageTest.testCopyFromLocalRecursive}} which should clean up the 
> directory (in a {{finally}} block). However, for S3, we may not always see 
> our own deletes.
> Quoting from https://aws.amazon.com/s3/faqs/ here:
> {quote}Q: What data consistency model does Amazon S3 employ?
> Amazon S3 buckets in all Regions provide read-after-write consistency for 
> PUTS of new objects and eventual consistency for overwrite PUTS and 
> DELETES.{quote}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5066: [FLINK-8148][yarn/s3] fix test instability in YarnFileSta...

2017-11-29 Thread zentol
Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5066
  
will fix the checkstyle violation on the way.


---


[jira] [Commented] (FLINK-8148) Test instability in YarnFileStageTest

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271025#comment-16271025
 ] 

ASF GitHub Bot commented on FLINK-8148:
---

Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5066
  
merging.


> Test instability in YarnFileStageTest
> -
>
> Key: FLINK-8148
> URL: https://issues.apache.org/jira/browse/FLINK-8148
> Project: Flink
>  Issue Type: Bug
>  Components: Tests, YARN
>Affects Versions: 1.4.0
>Reporter: Nico Kruber
>Assignee: Nico Kruber
>Priority: Critical
>  Labels: test-stability
>
> {code}
> Running org.apache.flink.yarn.YarnFileStageTestS3ITCase
> Tests run: 3, Failures: 1, Errors: 0, Skipped: 1, Time elapsed: 13.152 sec 
> <<< FAILURE! - in org.apache.flink.yarn.YarnFileStageTestS3ITCase
> testRecursiveUploadForYarnS3(org.apache.flink.yarn.YarnFileStageTestS3ITCase) 
>  Time elapsed: 8.515 sec  <<< FAILURE!
> java.lang.AssertionError: null
>   at org.junit.Assert.fail(Assert.java:86)
>   at org.junit.Assert.assertTrue(Assert.java:41)
>   at org.junit.Assert.assertFalse(Assert.java:64)
>   at org.junit.Assert.assertFalse(Assert.java:74)
>   at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarn(YarnFileStageTestS3ITCase.java:171)
>   at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarnS3(YarnFileStageTestS3ITCase.java:192)
> {code}
> from https://travis-ci.org/apache/flink/jobs/305861539
> {{YarnFileStageTestS3ITCase.testRecursiveUploadForYarn}} verifies that the 
> test directory used is cleaned up by 
> {{YarnFileStageTest.testCopyFromLocalRecursive}} which should clean up the 
> directory (in a {{finally}} block). However, for S3, we may not always see 
> our own deletes.
> Quoting from https://aws.amazon.com/s3/faqs/ here:
> {quote}Q: What data consistency model does Amazon S3 employ?
> Amazon S3 buckets in all Regions provide read-after-write consistency for 
> PUTS of new objects and eventual consistency for overwrite PUTS and 
> DELETES.{quote}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5066: [FLINK-8148][yarn/s3] fix test instability in YarnFileSta...

2017-11-29 Thread zentol
Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5066
  
merging.


---


[jira] [Commented] (FLINK-8171) Remove work arounds in Flip6LocalStreamEnvironment

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8171?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16271021#comment-16271021
 ] 

ASF GitHub Bot commented on FLINK-8171:
---

Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5101
  
Can we also remove this bit 
```
// temporary hack for FLIP-6 since slot-sharing isn't implemented yet
jobGraph.setAllowQueuedScheduling(true);
```
from the `RestClusterClient`?


> Remove work arounds in Flip6LocalStreamEnvironment
> --
>
> Key: FLINK-8171
> URL: https://issues.apache.org/jira/browse/FLINK-8171
> Project: Flink
>  Issue Type: Improvement
>  Components: Tests
>Affects Versions: 1.5.0
>Reporter: Till Rohrmann
>Assignee: Till Rohrmann
>  Labels: flip-6
> Fix For: 1.5.0
>
>
> After adding FLINK-7956, it is no longer necessary that the 
> {{Flip6LocalStreamEnvironment}} waits for the registration of TaskManagers 
> before submitting a job. Moreover, it is also possible to use slot sharing 
> when submitting jobs.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5101: [FLINK-8171] [flip6] Remove work arounds from Flip6LocalS...

2017-11-29 Thread zentol
Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5101
  
Can we also remove this bit 
```
// temporary hack for FLIP-6 since slot-sharing isn't implemented yet
jobGraph.setAllowQueuedScheduling(true);
```
from the `RestClusterClient`?


---


[jira] [Closed] (FLINK-7989) flink-dist jar is deployed twice onto YARN

2017-11-29 Thread Chesnay Schepler (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-7989?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler closed FLINK-7989.
---
   Resolution: Fixed
Fix Version/s: (was: 1.4.1)
   1.4.0

1.4: 3d568f52070a7719803a2bea5346d0e6f2a72615

> flink-dist jar is deployed twice onto YARN
> --
>
> Key: FLINK-7989
> URL: https://issues.apache.org/jira/browse/FLINK-7989
> Project: Flink
>  Issue Type: Bug
>  Components: YARN
>Affects Versions: 1.4.0
>Reporter: Nico Kruber
>Assignee: Nico Kruber
> Fix For: 1.4.0, 1.5.0
>
>
> We always add the {{flink-dist*.jar}} ourselves, but it could also be inside 
> a shipped folder such as the "lib/" folder and is then deployed again.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-8171) Remove work arounds in Flip6LocalStreamEnvironment

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8171?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270996#comment-16270996
 ] 

ASF GitHub Bot commented on FLINK-8171:
---

GitHub user tillrohrmann opened a pull request:

https://github.com/apache/flink/pull/5101

[FLINK-8171] [flip6] Remove work arounds from Flip6LocalStreamEnvironment

## What is the purpose of the change

It is no longer needed to wait for the registration of task managers and
to not use slot sharing when submitting jobs to the Flip-6 MiniCluster.
Therefore, we can remove these work arounds from the
Flip6LocalStreamEnvironment.

This PR is based on #5091.

## Does this pull request potentially affect one of the following parts:

  - Dependencies (does it add or upgrade a dependency): (no)
  - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
  - The serializers: (no)
  - The runtime per-record code paths (performance sensitive): (no)
  - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
  - The S3 file system connector: (no)

## Documentation

  - Does this pull request introduce a new feature? (no)
  - If yes, how is the feature documented? (not applicable)


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/tillrohrmann/flink 
fixLocalStreamEnvironmentITCase

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/5101.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #5101


commit d30dde83548dbeff4249f3b57b67cdb6247af510
Author: Till Rohrmann 
Date:   2017-11-14T22:50:52Z

[FLINK-8078] Introduce LogicalSlot interface

The LogicalSlot interface decouples the task deployment from the actual
slot implementation which at the moment is Slot, SimpleSlot and SharedSlot.
This is a helpful step to introduce a different slot implementation for
Flip-6.

commit e5da9566a6fc8a36ac8b06bae911c0dff5554e5d
Author: Till Rohrmann 
Date:   2017-11-15T13:20:27Z

[FLINK-8085] Thin out LogicalSlot interface

Remove isCanceled, isReleased method and decouple logical slot from 
Execution by
introducing a Payload interface which is set for a LogicalSlot. The Payload 
interface
is implemented by the Execution and allows to fail an implementation and 
obtaining
a termination future.

Introduce proper Execution#releaseFuture which is completed once the 
Execution's
assigned resource has been released.

commit 84d86bebe2f9f8395430e7c71dd2393ba117b44f
Author: Till Rohrmann 
Date:   2017-11-24T17:03:49Z

[FLINK-8087] Decouple Slot from AllocatedSlot

This commit introduces the SlotContext which is an abstraction for the 
SimpleSlot
to obtain the relevant slot information to do the communication with the
TaskManager without relying on the AllocatedSlot which is now only used by 
the
SlotPool.

commit 80a3cc848a0c724a2bc09b1b967cc9e6ccec5942
Author: Till Rohrmann 
Date:   2017-11-24T17:06:10Z

[FLINK-8088] Associate logical slots with the slot request id

Before logical slots like the SimpleSlot and SharedSlot where associated to 
the
actually allocated slot via the AllocationID. This, however, was 
sub-optimal because
allocated slots can be re-used to fulfill also other slot requests (logical 
slots).
Therefore, we should bind the logical slots to the right id with the right 
lifecycle
which is the slot request id.

commit 3e4550c0607744b20893dc90c587b63e68e4de1e
Author: Till Rohrmann 
Date:   2017-11-13T14:42:07Z

[FLINK-8089] Also check for other pending slot requests in offerSlot

Not only check for a slot request with the right allocation id but also 
check
whether we can fulfill other pending slot requests with an unclaimed offered
slot before adding it to the list of available slots.

commit b04dda46aaf298d921929910574662970d9c5093
Author: Till Rohrmann 
Date:   2017-11-24T22:29:53Z

[hotfix] Speed up RecoveryITCase

commit e512558917f9bb5005024630b8a015cd624164b4
Author: Till Rohrmann 
Date:   2017-11-24T17:08:38Z

[FLINK-7956] [flip6] Add support for queued scheduling with slot sharing to 
SlotPool

This commit adds support for queued scheduling with slot sharing to the
SlotPool. The idea of slot sharing is that multiple tasks can run in the
same slot. Moreover, queued scheduling means that a slot request must not
be completed right away but at a later point in time. This allows to
start new TaskExecutors in case that there are no more slots left.

The main component responsible for the management o

[jira] [Commented] (FLINK-7574) Remove unused dependencies from flink-clients

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7574?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270997#comment-16270997
 ] 

ASF GitHub Bot commented on FLINK-7574:
---

Github user yew1eb commented on the issue:

https://github.com/apache/flink/pull/5076
  
Hi @greghogan,  You're quite right, Thanks for your advice.  
For a discussion of this improvement, you can see #4712.



> Remove unused dependencies from flink-clients
> -
>
> Key: FLINK-7574
> URL: https://issues.apache.org/jira/browse/FLINK-7574
> Project: Flink
>  Issue Type: Sub-task
>  Components: Build System
>Affects Versions: 1.3.2
> Environment: Apache Maven 3.3.9, Java version: 1.8.0_144
>Reporter: Hai Zhou UTC+8
>Assignee: Hai Zhou UTC+8
>
> [INFO] --- maven-dependency-plugin:2.10:analyze (default-cli) @ 
> flink-clients_2.11 ---
> [WARNING] Used undeclared dependencies found:
> [WARNING]org.scala-lang:scala-library:jar:2.11.11:compile
> [WARNING]com.data-artisans:flakka-actor_2.11:jar:2.3-custom:compile
> [WARNING] Unused declared dependencies found:
> [WARNING]org.hamcrest:hamcrest-all:jar:1.3:test
> [WARNING]org.apache.flink:force-shading:jar:1.4-SNAPSHOT:compile
> [WARNING]org.powermock:powermock-module-junit4:jar:1.6.5:test
> [WARNING]com.google.code.findbugs:jsr305:jar:1.3.9:compile
> [WARNING]log4j:log4j:jar:1.2.17:test
> [WARNING]org.powermock:powermock-api-mockito:jar:1.6.5:test
> [WARNING]org.slf4j:slf4j-log4j12:jar:1.7.7:test



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5076: [FLINK-7574][build] POM Cleanup flink-clients

2017-11-29 Thread yew1eb
Github user yew1eb commented on the issue:

https://github.com/apache/flink/pull/5076
  
Hi @greghogan,  You're quite right, Thanks for your advice.  
For a discussion of this improvement, you can see #4712.



---


[GitHub] flink pull request #5101: [FLINK-8171] [flip6] Remove work arounds from Flip...

2017-11-29 Thread tillrohrmann
GitHub user tillrohrmann opened a pull request:

https://github.com/apache/flink/pull/5101

[FLINK-8171] [flip6] Remove work arounds from Flip6LocalStreamEnvironment

## What is the purpose of the change

It is no longer needed to wait for the registration of task managers and
to not use slot sharing when submitting jobs to the Flip-6 MiniCluster.
Therefore, we can remove these work arounds from the
Flip6LocalStreamEnvironment.

This PR is based on #5091.

## Does this pull request potentially affect one of the following parts:

  - Dependencies (does it add or upgrade a dependency): (no)
  - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
  - The serializers: (no)
  - The runtime per-record code paths (performance sensitive): (no)
  - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
  - The S3 file system connector: (no)

## Documentation

  - Does this pull request introduce a new feature? (no)
  - If yes, how is the feature documented? (not applicable)


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/tillrohrmann/flink 
fixLocalStreamEnvironmentITCase

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/5101.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #5101


commit d30dde83548dbeff4249f3b57b67cdb6247af510
Author: Till Rohrmann 
Date:   2017-11-14T22:50:52Z

[FLINK-8078] Introduce LogicalSlot interface

The LogicalSlot interface decouples the task deployment from the actual
slot implementation which at the moment is Slot, SimpleSlot and SharedSlot.
This is a helpful step to introduce a different slot implementation for
Flip-6.

commit e5da9566a6fc8a36ac8b06bae911c0dff5554e5d
Author: Till Rohrmann 
Date:   2017-11-15T13:20:27Z

[FLINK-8085] Thin out LogicalSlot interface

Remove isCanceled, isReleased method and decouple logical slot from 
Execution by
introducing a Payload interface which is set for a LogicalSlot. The Payload 
interface
is implemented by the Execution and allows to fail an implementation and 
obtaining
a termination future.

Introduce proper Execution#releaseFuture which is completed once the 
Execution's
assigned resource has been released.

commit 84d86bebe2f9f8395430e7c71dd2393ba117b44f
Author: Till Rohrmann 
Date:   2017-11-24T17:03:49Z

[FLINK-8087] Decouple Slot from AllocatedSlot

This commit introduces the SlotContext which is an abstraction for the 
SimpleSlot
to obtain the relevant slot information to do the communication with the
TaskManager without relying on the AllocatedSlot which is now only used by 
the
SlotPool.

commit 80a3cc848a0c724a2bc09b1b967cc9e6ccec5942
Author: Till Rohrmann 
Date:   2017-11-24T17:06:10Z

[FLINK-8088] Associate logical slots with the slot request id

Before logical slots like the SimpleSlot and SharedSlot where associated to 
the
actually allocated slot via the AllocationID. This, however, was 
sub-optimal because
allocated slots can be re-used to fulfill also other slot requests (logical 
slots).
Therefore, we should bind the logical slots to the right id with the right 
lifecycle
which is the slot request id.

commit 3e4550c0607744b20893dc90c587b63e68e4de1e
Author: Till Rohrmann 
Date:   2017-11-13T14:42:07Z

[FLINK-8089] Also check for other pending slot requests in offerSlot

Not only check for a slot request with the right allocation id but also 
check
whether we can fulfill other pending slot requests with an unclaimed offered
slot before adding it to the list of available slots.

commit b04dda46aaf298d921929910574662970d9c5093
Author: Till Rohrmann 
Date:   2017-11-24T22:29:53Z

[hotfix] Speed up RecoveryITCase

commit e512558917f9bb5005024630b8a015cd624164b4
Author: Till Rohrmann 
Date:   2017-11-24T17:08:38Z

[FLINK-7956] [flip6] Add support for queued scheduling with slot sharing to 
SlotPool

This commit adds support for queued scheduling with slot sharing to the
SlotPool. The idea of slot sharing is that multiple tasks can run in the
same slot. Moreover, queued scheduling means that a slot request must not
be completed right away but at a later point in time. This allows to
start new TaskExecutors in case that there are no more slots left.

The main component responsible for the management of shared slots is the
SlotSharingManager. The SlotSharingManager maintains internally a tree-like
structure which stores the SlotContext future of the underlying
AllocatedSlot. Whenever this future is completed potentially pending
LogicalSl

[jira] [Created] (FLINK-8171) Remove work arounds in Flip6LocalStreamEnvironment

2017-11-29 Thread Till Rohrmann (JIRA)
Till Rohrmann created FLINK-8171:


 Summary: Remove work arounds in Flip6LocalStreamEnvironment
 Key: FLINK-8171
 URL: https://issues.apache.org/jira/browse/FLINK-8171
 Project: Flink
  Issue Type: Improvement
  Components: Tests
Affects Versions: 1.5.0
Reporter: Till Rohrmann
Assignee: Till Rohrmann
 Fix For: 1.5.0


After adding FLINK-7956, it is no longer necessary that the 
{{Flip6LocalStreamEnvironment}} waits for the registration of TaskManagers 
before submitting a job. Moreover, it is also possible to use slot sharing when 
submitting jobs.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-7573) Introduce Http protocol connector for Elasticsearch2

2017-11-29 Thread Dylan Meissner (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7573?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270979#comment-16270979
 ] 

Dylan Meissner commented on FLINK-7573:
---

Worth noting, as of Elasticsearch 5.6 there is an official [Java High Level 
REST 
Client](https://www.elastic.co/guide/en/elasticsearch/client/java-rest/5.6/java-rest-high.html).
 It _accepts the same request arguments as the TransportClient and returns the 
same response objects_.

> Introduce Http protocol connector for Elasticsearch2
> 
>
> Key: FLINK-7573
> URL: https://issues.apache.org/jira/browse/FLINK-7573
> Project: Flink
>  Issue Type: Improvement
>  Components: ElasticSearch Connector
>Reporter: mingleizhang
>
> Currently, all connectors as far as I have known that merely support the TCP 
> transport protocol of Elasticsearch, but some of company's ES cluster just 
> relies on the HTTP protocol, and close the TCP port on production 
> environment. So, I suggest add a new implemention for creating a HTTP 
> protocol by using {{JestClient}}, which is a Java HTTP Rest client for 
> ElasticSearch.
> FYI
> I used 9300 port to access ES cluster. It is really awful and scared. Because 
> program can not run on a production environment based on a 
> {{TransportClient}}. So, I have to access the ES cluster by 9200 port 
> instead. 9300 is for TCP. 9200 is for HTTP for accessing I guess here.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Comment Edited] (FLINK-7573) Introduce Http protocol connector for Elasticsearch2

2017-11-29 Thread Dylan Meissner (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7573?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270979#comment-16270979
 ] 

Dylan Meissner edited comment on FLINK-7573 at 11/29/17 3:52 PM:
-

Worth noting, as of Elasticsearch 5.6 there is an official "Java High Level 
REST Client". It _accepts the same request arguments as the TransportClient and 
returns the same response objects_.

https://www.elastic.co/guide/en/elasticsearch/client/java-rest/5.6/java-rest-high.html


was (Author: dylanmei):
Worth noting, as of Elasticsearch 5.6 there is an official [Java High Level 
REST 
Client](https://www.elastic.co/guide/en/elasticsearch/client/java-rest/5.6/java-rest-high.html).
 It _accepts the same request arguments as the TransportClient and returns the 
same response objects_.

> Introduce Http protocol connector for Elasticsearch2
> 
>
> Key: FLINK-7573
> URL: https://issues.apache.org/jira/browse/FLINK-7573
> Project: Flink
>  Issue Type: Improvement
>  Components: ElasticSearch Connector
>Reporter: mingleizhang
>
> Currently, all connectors as far as I have known that merely support the TCP 
> transport protocol of Elasticsearch, but some of company's ES cluster just 
> relies on the HTTP protocol, and close the TCP port on production 
> environment. So, I suggest add a new implemention for creating a HTTP 
> protocol by using {{JestClient}}, which is a Java HTTP Rest client for 
> ElasticSearch.
> FYI
> I used 9300 port to access ES cluster. It is really awful and scared. Because 
> program can not run on a production environment based on a 
> {{TransportClient}}. So, I have to access the ES cluster by 9200 port 
> instead. 9300 is for TCP. 9200 is for HTTP for accessing I guess here.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-4812) Report Watermark metrics in all operators

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270976#comment-16270976
 ] 

ASF GitHub Bot commented on FLINK-4812:
---

Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5100
  
the coupling of input to output gauge becomes rather tricky. The input 
gauge must be reachable in `Stream[Two]InputProcessor`, 
`OperatorChain#createChainedOperator`, and the output gauge in 
`AbstractStreamOperator#setup`.

For sources, sinks and watermark assigners we must refer to the same gauge 
in both places. I don't see how we can do this in a nice way.

The rather nice thing the current approach has is that everything is 
contained in `AbstractStreamOperator` and isn't scattered all over the place.

`Given that we allow people to completely re-define how operators handle 
watermarks, wouldn't it make sense to give them the same freedom when it comes 
to the input watermark metrics?




> Report Watermark metrics in all operators
> -
>
> Key: FLINK-4812
> URL: https://issues.apache.org/jira/browse/FLINK-4812
> Project: Flink
>  Issue Type: Improvement
>  Components: Metrics
>Reporter: Robert Metzger
>Assignee: Chesnay Schepler
>Priority: Critical
> Fix For: 1.5.0
>
>
> As reported by a user, Flink does currently not export the current low 
> watermark for sources 
> (http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/currentLowWatermark-metric-not-reported-for-all-tasks-td13770.html).
> This JIRA is for adding such a metric for the sources as well.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5100: [FLINK-4812][metrics] Expose currentLowWatermark for all ...

2017-11-29 Thread zentol
Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5100
  
the coupling of input to output gauge becomes rather tricky. The input 
gauge must be reachable in `Stream[Two]InputProcessor`, 
`OperatorChain#createChainedOperator`, and the output gauge in 
`AbstractStreamOperator#setup`.

For sources, sinks and watermark assigners we must refer to the same gauge 
in both places. I don't see how we can do this in a nice way.

The rather nice thing the current approach has is that everything is 
contained in `AbstractStreamOperator` and isn't scattered all over the place.

`Given that we allow people to completely re-define how operators handle 
watermarks, wouldn't it make sense to give them the same freedom when it comes 
to the input watermark metrics?




---


[jira] [Commented] (FLINK-7762) Make WikipediaEditsSourceTest a proper test

2017-11-29 Thread Hai Zhou UTC+8 (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7762?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270974#comment-16270974
 ] 

Hai Zhou UTC+8 commented on FLINK-7762:
---

Hi [~uce], yes. If you want, then you can work on it.

> Make WikipediaEditsSourceTest a proper test
> ---
>
> Key: FLINK-7762
> URL: https://issues.apache.org/jira/browse/FLINK-7762
> Project: Flink
>  Issue Type: Improvement
>  Components: Streaming Connectors
>Reporter: Aljoscha Krettek
>Assignee: Ufuk Celebi
>Priority: Minor
>
> {{WikipediaEditsSourceTest}} is currently an ITCase even though it's called 
> test. Making it a test reduces runtime and also makes it more stable because 
> we don't run a whole Flink job.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-4812) Report Watermark metrics in all operators

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270954#comment-16270954
 ] 

ASF GitHub Bot commented on FLINK-4812:
---

Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5100
  
ah, weÄ#re doing the same already for `numRecordsIn` maybe we can do 
this after all.


> Report Watermark metrics in all operators
> -
>
> Key: FLINK-4812
> URL: https://issues.apache.org/jira/browse/FLINK-4812
> Project: Flink
>  Issue Type: Improvement
>  Components: Metrics
>Reporter: Robert Metzger
>Assignee: Chesnay Schepler
>Priority: Critical
> Fix For: 1.5.0
>
>
> As reported by a user, Flink does currently not export the current low 
> watermark for sources 
> (http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/currentLowWatermark-metric-not-reported-for-all-tasks-td13770.html).
> This JIRA is for adding such a metric for the sources as well.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5100: [FLINK-4812][metrics] Expose currentLowWatermark for all ...

2017-11-29 Thread zentol
Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5100
  
ah, weÄ#re doing the same already for `numRecordsIn` maybe we can do 
this after all.


---


[jira] [Commented] (FLINK-7595) Removing stateless task from task chain breaks savepoint restore

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7595?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270939#comment-16270939
 ] 

ASF GitHub Bot commented on FLINK-7595:
---

Github user zentol commented on the issue:

https://github.com/apache/flink/pull/4651
  
Rebased on top of master, let's see what travis says. Didn't get any 
conflicts so 🤞 .


> Removing stateless task from task chain breaks savepoint restore
> 
>
> Key: FLINK-7595
> URL: https://issues.apache.org/jira/browse/FLINK-7595
> Project: Flink
>  Issue Type: Bug
>  Components: State Backends, Checkpointing
>Reporter: Ufuk Celebi
>Assignee: Chesnay Schepler
> Attachments: ChainedTaskRemoveTest.java
>
>
> When removing a stateless operator from a 2-task chain where the head 
> operator is stateful breaks savepoint restore with 
> {code}
> Caused by: java.lang.IllegalStateException: Failed to rollback to savepoint 
> /var/folders/py/s_1l8vln6f19ygc77m8c4zhrgn/T/junit1167397515334838028/junit8006766303945373008/savepoint-cb0bcf-3cfa67865ac0.
>  Cannot map savepoint state...
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #4651: [FLINK-7595] [Savepoints] Allow removing stateless operat...

2017-11-29 Thread zentol
Github user zentol commented on the issue:

https://github.com/apache/flink/pull/4651
  
Rebased on top of master, let's see what travis says. Didn't get any 
conflicts so 🤞 .


---


[jira] [Commented] (FLINK-4812) Report Watermark metrics in all operators

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270933#comment-16270933
 ] 

ASF GitHub Bot commented on FLINK-4812:
---

Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5100
  
You can't measure input metrics in the ChainingOutput (nicely) since for a 
chain A->B does B at no point have access to the output of A.

We would need to have the task setup all watermark metrics and either pass 
them on to operators for registration or have the task register them for 
operators, which is just backwards. This also clashes heavily with the idea of 
operators deciding whether the same metric is used for input&output.


> Report Watermark metrics in all operators
> -
>
> Key: FLINK-4812
> URL: https://issues.apache.org/jira/browse/FLINK-4812
> Project: Flink
>  Issue Type: Improvement
>  Components: Metrics
>Reporter: Robert Metzger
>Assignee: Chesnay Schepler
>Priority: Critical
> Fix For: 1.5.0
>
>
> As reported by a user, Flink does currently not export the current low 
> watermark for sources 
> (http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/currentLowWatermark-metric-not-reported-for-all-tasks-td13770.html).
> This JIRA is for adding such a metric for the sources as well.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (FLINK-8167) Failing WikipediaEditsSourceTest on Travis

2017-11-29 Thread Ufuk Celebi (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8167?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270934#comment-16270934
 ] 

Ufuk Celebi commented on FLINK-8167:


I reproduced this locally. A nickname clash as indicated by the Travis run log 
results in a connection timeout which subsequently results in the test to 
timeout.

> Failing WikipediaEditsSourceTest on Travis
> --
>
> Key: FLINK-8167
> URL: https://issues.apache.org/jira/browse/FLINK-8167
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.5.0
>Reporter: Till Rohrmann
>Assignee: Ufuk Celebi
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.5.0
>
>
> I've observed the {{WikipediaEditsSourceTest}} failing on Travis.
> https://travis-ci.org/tillrohrmann/flink/jobs/308351298



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5100: [FLINK-4812][metrics] Expose currentLowWatermark for all ...

2017-11-29 Thread zentol
Github user zentol commented on the issue:

https://github.com/apache/flink/pull/5100
  
You can't measure input metrics in the ChainingOutput (nicely) since for a 
chain A->B does B at no point have access to the output of A.

We would need to have the task setup all watermark metrics and either pass 
them on to operators for registration or have the task register them for 
operators, which is just backwards. This also clashes heavily with the idea of 
operators deciding whether the same metric is used for input&output.


---


[jira] [Commented] (FLINK-8158) Rowtime window inner join emits late data

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-8158?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270924#comment-16270924
 ] 

ASF GitHub Bot commented on FLINK-8158:
---

Github user hequn8128 commented on the issue:

https://github.com/apache/flink/pull/5094
  
Thanks for your replies.
@fhueske : The watermark must be aligned with timestamps, and it is the 
main reason why watermarks are hold back(right?).  Current window join may 
output a record with timestamp equals to the earlier **output** watermark(see 
`testRowTimeJoinWithCommonBounds2 ` in `JoinHarnessTest `).  

@xccui : I am considering that if we can cache more late records to have a 
more complete join result, and this can be achieved by caching both left and 
right data that is later than the holding back watermark.


> Rowtime window inner join emits late data
> -
>
> Key: FLINK-8158
> URL: https://issues.apache.org/jira/browse/FLINK-8158
> Project: Flink
>  Issue Type: Bug
>  Components: Table API & SQL
>Reporter: Hequn Cheng
>Assignee: Hequn Cheng
> Attachments: screenshot-1xxx.png
>
>
> When executing the join, the join operator needs to make sure that no late 
> data is emitted. Currently, this achieved by holding back watermarks. 
> However, the window border is not handled correctly. For the sql bellow: 
> {quote}
> val sqlQuery =
>   """
> SELECT t2.key, t2.id, t1.id
> FROM T1 as t1 join T2 as t2 ON
>   t1.key = t2.key AND
>   t1.rt BETWEEN t2.rt - INTERVAL '5' SECOND AND
> t2.rt + INTERVAL '1' SECOND
> """.stripMargin
> val data1 = new mutable.MutableList[(String, String, Long)]
> // for boundary test
> data1.+=(("A", "LEFT1", 6000L))
> val data2 = new mutable.MutableList[(String, String, Long)]
> data2.+=(("A", "RIGHT1", 6000L))
> {quote}
> Join will output a watermark with timestamp 1000, but if left comes with 
> another data ("A", "LEFT1", 1000L), join will output a record with timestamp 
> 1000 which equals previous watermark.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5094: [FLINK-8158] [table] Fix rowtime window inner join emits ...

2017-11-29 Thread hequn8128
Github user hequn8128 commented on the issue:

https://github.com/apache/flink/pull/5094
  
Thanks for your replies.
@fhueske : The watermark must be aligned with timestamps, and it is the 
main reason why watermarks are hold back(right?).  Current window join may 
output a record with timestamp equals to the earlier **output** watermark(see 
`testRowTimeJoinWithCommonBounds2 ` in `JoinHarnessTest `).  

@xccui : I am considering that if we can cache more late records to have a 
more complete join result, and this can be achieved by caching both left and 
right data that is later than the holding back watermark.


---


[jira] [Commented] (FLINK-7595) Removing stateless task from task chain breaks savepoint restore

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-7595?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270923#comment-16270923
 ] 

ASF GitHub Bot commented on FLINK-7595:
---

Github user aljoscha commented on the issue:

https://github.com/apache/flink/pull/4651
  
This change looks good! Does it still build when rebased on 
master/release-1.4?


> Removing stateless task from task chain breaks savepoint restore
> 
>
> Key: FLINK-7595
> URL: https://issues.apache.org/jira/browse/FLINK-7595
> Project: Flink
>  Issue Type: Bug
>  Components: State Backends, Checkpointing
>Reporter: Ufuk Celebi
>Assignee: Chesnay Schepler
> Attachments: ChainedTaskRemoveTest.java
>
>
> When removing a stateless operator from a 2-task chain where the head 
> operator is stateful breaks savepoint restore with 
> {code}
> Caused by: java.lang.IllegalStateException: Failed to rollback to savepoint 
> /var/folders/py/s_1l8vln6f19ygc77m8c4zhrgn/T/junit1167397515334838028/junit8006766303945373008/savepoint-cb0bcf-3cfa67865ac0.
>  Cannot map savepoint state...
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #4651: [FLINK-7595] [Savepoints] Allow removing stateless operat...

2017-11-29 Thread aljoscha
Github user aljoscha commented on the issue:

https://github.com/apache/flink/pull/4651
  
This change looks good! Does it still build when rebased on 
master/release-1.4?


---


[jira] [Commented] (FLINK-4812) Report Watermark metrics in all operators

2017-11-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16270898#comment-16270898
 ] 

ASF GitHub Bot commented on FLINK-4812:
---

Github user aljoscha commented on the issue:

https://github.com/apache/flink/pull/5100
  
Ah stupid, so we have to do it in the chain and in the input processors. 
Because the problem of operators overriding `processWatermark()` is real. 😅 


> Report Watermark metrics in all operators
> -
>
> Key: FLINK-4812
> URL: https://issues.apache.org/jira/browse/FLINK-4812
> Project: Flink
>  Issue Type: Improvement
>  Components: Metrics
>Reporter: Robert Metzger
>Assignee: Chesnay Schepler
>Priority: Critical
> Fix For: 1.5.0
>
>
> As reported by a user, Flink does currently not export the current low 
> watermark for sources 
> (http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/currentLowWatermark-metric-not-reported-for-all-tasks-td13770.html).
> This JIRA is for adding such a metric for the sources as well.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] flink issue #5100: [FLINK-4812][metrics] Expose currentLowWatermark for all ...

2017-11-29 Thread aljoscha
Github user aljoscha commented on the issue:

https://github.com/apache/flink/pull/5100
  
Ah stupid, so we have to do it in the chain and in the input processors. 
Because the problem of operators overriding `processWatermark()` is real. 😅 


---


  1   2   3   4   >