[jira] [Updated] (FLINK-4574) Strengthen fetch interval implementation in Kinesis consumer

2016-09-03 Thread Tzu-Li (Gordon) Tai (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-4574?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tzu-Li (Gordon) Tai updated FLINK-4574:
---
Description: 
As pointed out by [~rmetzger], right now the fetch interval implementation in 
the {{ShardConsumer}} class of the Kinesis consumer can lead to much longer 
interval times than specified by the user, ex. say the specified fetch interval 
is {{f}}, it takes {{x}} to complete a {{getRecords()}} call, and {{y}} to 
complete processing the fetched records for emitting, than the actual interval 
between each fetch is actually {{f+x+y}}.

The main problem with this is that we can never guarantee how much time has 
past since the last {{getRecords}} call, thus can not guarantee that returned 
shard iterators will not have expired the next time we use them, even if we 
limit the user-given value for {{f}} to not be longer than the iterator expire 
time.

I propose to improve this by, per {{ShardConsumer}}, use a 
{{ScheduledExecutorService}} / {{Timer}} to do the fixed-interval fetching, and 
a separate blocking queue that collects the fetched records for emitting.

  was:
As pointed out by [~rmetzger], right now the fetch interval implementation in 
the {{ShardConsumer}} class of the Kinesis consumer can lead to much longer 
interval times than specified by the user, ex. say the specified fetch interval 
is {{f}}, it takes {{x}} to complete a {{getRecords()}} call, and {{y}} to 
complete processing the fetched records for emitting, than the actual interval 
between each fetch is actually {{f+x+y}}.

The main problem with this is that we can never guarantee how much time has 
past since the last {{getRecords}} call, thus can not guarantee that returned 
shard iterators will not have expired the next time we use them, even if we 
limit the user-given value for {{f}} to not be longer than the iterator expire 
time.

I propose to improve this by, per {{ShardConsumer}}, use a 
{{ScheduledExecutorService}} to do the fixed-interval fetching, and a separate 
blocking queue that collects the fetched records for emitting.


> Strengthen fetch interval implementation in Kinesis consumer
> 
>
> Key: FLINK-4574
> URL: https://issues.apache.org/jira/browse/FLINK-4574
> Project: Flink
>  Issue Type: Improvement
>  Components: Kinesis Connector
>Affects Versions: 1.1.0
>Reporter: Tzu-Li (Gordon) Tai
> Fix For: 1.2.0, 1.1.3
>
>
> As pointed out by [~rmetzger], right now the fetch interval implementation in 
> the {{ShardConsumer}} class of the Kinesis consumer can lead to much longer 
> interval times than specified by the user, ex. say the specified fetch 
> interval is {{f}}, it takes {{x}} to complete a {{getRecords()}} call, and 
> {{y}} to complete processing the fetched records for emitting, than the 
> actual interval between each fetch is actually {{f+x+y}}.
> The main problem with this is that we can never guarantee how much time has 
> past since the last {{getRecords}} call, thus can not guarantee that returned 
> shard iterators will not have expired the next time we use them, even if we 
> limit the user-given value for {{f}} to not be longer than the iterator 
> expire time.
> I propose to improve this by, per {{ShardConsumer}}, use a 
> {{ScheduledExecutorService}} / {{Timer}} to do the fixed-interval fetching, 
> and a separate blocking queue that collects the fetched records for emitting.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-4559) Kinesis Producer not setting credentials provider properly when AWS_CREDENTIALS_PROVIDER is "AUTO"

2016-09-03 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15462090#comment-15462090
 ] 

ASF GitHub Bot commented on FLINK-4559:
---

Github user tzulitai commented on the issue:

https://github.com/apache/flink/pull/2470
  
Thanks for the review @greghogan.
I think it makes sense to let `AUTO` be the default, as the usual use case 
for most AWS users is to auto-detect the credentials delivered to instances by 
AWS's metadata service. Changing this should not break user code either. Would 
you like to open the ticket for that?


> Kinesis Producer not setting credentials provider properly when 
> AWS_CREDENTIALS_PROVIDER is "AUTO"
> --
>
> Key: FLINK-4559
> URL: https://issues.apache.org/jira/browse/FLINK-4559
> Project: Flink
>  Issue Type: Bug
>  Components: Kinesis Connector
>Affects Versions: 1.1.0, 1.1.1, 1.1.2
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Tzu-Li (Gordon) Tai
>Priority: Minor
> Fix For: 1.1.3
>
>
> If the {{AWS_CREDENTIAL_PROVIDER}} option is set to {{AUTO}}, 
> {{AWSUtils.getCredentialsProvider}} will return {{null}}, so 
> {{setCredentialsProvider}} should not be explicitly called on the internally 
> built {{KinesisProducerConfiguration}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink issue #2470: [FLINK-4559][kinesis-connector] Fix AWSUtil.getCredential...

2016-09-03 Thread tzulitai
Github user tzulitai commented on the issue:

https://github.com/apache/flink/pull/2470
  
Thanks for the review @greghogan.
I think it makes sense to let `AUTO` be the default, as the usual use case 
for most AWS users is to auto-detect the credentials delivered to instances by 
AWS's metadata service. Changing this should not break user code either. Would 
you like to open the ticket for that?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-4513) Kafka connector documentation refers to Flink 1.1-SNAPSHOT

2016-09-03 Thread Neelesh Srinivas Salian (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4513?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15461833#comment-15461833
 ] 

Neelesh Srinivas Salian commented on FLINK-4513:


Shall I work on this? Need to change all the connector pages to appropriate 
this.
Will check if this is true for other versions too.


> Kafka connector documentation refers to Flink 1.1-SNAPSHOT
> --
>
> Key: FLINK-4513
> URL: https://issues.apache.org/jira/browse/FLINK-4513
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.1.1
>Reporter: Fabian Hueske
>Priority: Trivial
> Fix For: 1.1.2
>
>
> The Kafka connector documentation: 
> https://ci.apache.org/projects/flink/flink-docs-release-1.1/apis/streaming/connectors/kafka.html
>  of Flink 1.1 refers to a Flink 1.1-SNAPSHOT Maven version. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request #2471: Broken links on website

2016-09-03 Thread apivovarov
GitHub user apivovarov opened a pull request:

https://github.com/apache/flink/pull/2471

Broken links on website

Thanks for contributing to Apache Flink. Before you open your pull request, 
please take the following check list into consideration.
If your changes take all of the items into account, feel free to open your 
pull request. For more information and/or questions please refer to the [How To 
Contribute guide](http://flink.apache.org/how-to-contribute.html).
In addition to going through the list, please provide a meaningful 
description of your changes.

- [ ] General
  - The pull request references the related JIRA issue ("[FLINK-XXX] Jira 
title text")
  - The pull request addresses only one issue
  - Each commit in the PR has a meaningful commit message (including the 
JIRA id)

- [ ] Documentation
  - Documentation has been added for new functionality
  - Old documentation affected by the pull request has been updated
  - JavaDoc for public methods has been added

- [ ] Tests & Build
  - Functionality added by the pull request is covered by tests
  - `mvn clean verify` has been executed successfully locally or a Travis 
build has passed

The website has the following broken links

DataSet API for static data embedded in Java, Scala, and Python,
http://ci.apache.org/projects/flink/flink-docs-master/apis/batch/index.html

Table API with a SQL-like expression language embedded in Java and Scala.

http://ci.apache.org/projects/flink/flink-docs-master/apis/batch/libs/table.html

Gelly, a graph processing API and library.

http://ci.apache.org/projects/flink/flink-docs-master/apis/batch/libs/gelly.html

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/apivovarov/flink patch-1

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/2471.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2471


commit b6d56682a55ae46e70cba33326ec58eb753fa73a
Author: Alexander Pivovarov 
Date:   2016-09-03T21:14:58Z

Broken links on website

The website has the following broken links

DataSet API for static data embedded in Java, Scala, and Python,
http://ci.apache.org/projects/flink/flink-docs-master/apis/batch/index.html

Table API with a SQL-like expression language embedded in Java and Scala.

http://ci.apache.org/projects/flink/flink-docs-master/apis/batch/libs/table.html

Gelly, a graph processing API and library.

http://ci.apache.org/projects/flink/flink-docs-master/apis/batch/libs/gelly.html




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-4559) Kinesis Producer not setting credentials provider properly when AWS_CREDENTIALS_PROVIDER is "AUTO"

2016-09-03 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15461654#comment-15461654
 ] 

ASF GitHub Bot commented on FLINK-4559:
---

Github user greghogan commented on the issue:

https://github.com/apache/flink/pull/2470
  
Looks good to me.

Should we also create a ticket to make AUTO the default as this checks for 
credentials in multiple places?
  
http://docs.aws.amazon.com/java-sdk/latest/developer-guide/credentials.html


> Kinesis Producer not setting credentials provider properly when 
> AWS_CREDENTIALS_PROVIDER is "AUTO"
> --
>
> Key: FLINK-4559
> URL: https://issues.apache.org/jira/browse/FLINK-4559
> Project: Flink
>  Issue Type: Bug
>  Components: Kinesis Connector
>Affects Versions: 1.1.0, 1.1.1, 1.1.2
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Tzu-Li (Gordon) Tai
>Priority: Minor
> Fix For: 1.1.3
>
>
> If the {{AWS_CREDENTIAL_PROVIDER}} option is set to {{AUTO}}, 
> {{AWSUtils.getCredentialsProvider}} will return {{null}}, so 
> {{setCredentialsProvider}} should not be explicitly called on the internally 
> built {{KinesisProducerConfiguration}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink issue #2470: [FLINK-4559][kinesis-connector] Fix AWSUtil.getCredential...

2016-09-03 Thread greghogan
Github user greghogan commented on the issue:

https://github.com/apache/flink/pull/2470
  
Looks good to me.

Should we also create a ticket to make AUTO the default as this checks for 
credentials in multiple places?
  
http://docs.aws.amazon.com/java-sdk/latest/developer-guide/credentials.html


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-4559) Kinesis Producer not setting credentials provider properly when AWS_CREDENTIALS_PROVIDER is "AUTO"

2016-09-03 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15461528#comment-15461528
 ] 

ASF GitHub Bot commented on FLINK-4559:
---

GitHub user tzulitai opened a pull request:

https://github.com/apache/flink/pull/2470

[FLINK-4559][kinesis-connector] Fix AWSUtil.getCredentialsProvider() to not 
return null

Let `AWSUtil#getCredentialsProvider` return a 
`DefaultAWSCredentialsProviderChain` instead of `null` when
`AWS_CREDENTIALS_PROVIDER` config is set to `AUTO`.

This mainly fixes the Kinesis producer to correctly get credentials when 
using the `AUTO` option.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/tzulitai/flink FLINK-4559

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/2470.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2470


commit 7b86137672a4bdec7d67108bf423ef10269ddfa7
Author: Tzu-Li (Gordon) Tai 
Date:   2016-09-03T17:37:39Z

[FLINK-4559][kinesis-connector] Fix AWSUtil.getCredentialsProvider() to not 
return null

Return a DefaultAWSCredentialsProviderChain instead of null when
AWS_CREDENTIALS_PROVIDER config is set to "AUTO"




> Kinesis Producer not setting credentials provider properly when 
> AWS_CREDENTIALS_PROVIDER is "AUTO"
> --
>
> Key: FLINK-4559
> URL: https://issues.apache.org/jira/browse/FLINK-4559
> Project: Flink
>  Issue Type: Bug
>  Components: Kinesis Connector
>Affects Versions: 1.1.0, 1.1.1, 1.1.2
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Tzu-Li (Gordon) Tai
>Priority: Minor
> Fix For: 1.1.3
>
>
> If the {{AWS_CREDENTIAL_PROVIDER}} option is set to {{AUTO}}, 
> {{AWSUtils.getCredentialsProvider}} will return {{null}}, so 
> {{setCredentialsProvider}} should not be explicitly called on the internally 
> built {{KinesisProducerConfiguration}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request #2470: [FLINK-4559][kinesis-connector] Fix AWSUtil.getCre...

2016-09-03 Thread tzulitai
GitHub user tzulitai opened a pull request:

https://github.com/apache/flink/pull/2470

[FLINK-4559][kinesis-connector] Fix AWSUtil.getCredentialsProvider() to not 
return null

Let `AWSUtil#getCredentialsProvider` return a 
`DefaultAWSCredentialsProviderChain` instead of `null` when
`AWS_CREDENTIALS_PROVIDER` config is set to `AUTO`.

This mainly fixes the Kinesis producer to correctly get credentials when 
using the `AUTO` option.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/tzulitai/flink FLINK-4559

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/2470.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2470


commit 7b86137672a4bdec7d67108bf423ef10269ddfa7
Author: Tzu-Li (Gordon) Tai 
Date:   2016-09-03T17:37:39Z

[FLINK-4559][kinesis-connector] Fix AWSUtil.getCredentialsProvider() to not 
return null

Return a DefaultAWSCredentialsProviderChain instead of null when
AWS_CREDENTIALS_PROVIDER config is set to "AUTO"




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Assigned] (FLINK-4559) Kinesis Producer not setting credentials provider properly when AWS_CREDENTIALS_PROVIDER is "AUTO"

2016-09-03 Thread Tzu-Li (Gordon) Tai (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-4559?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tzu-Li (Gordon) Tai reassigned FLINK-4559:
--

Assignee: Tzu-Li (Gordon) Tai

> Kinesis Producer not setting credentials provider properly when 
> AWS_CREDENTIALS_PROVIDER is "AUTO"
> --
>
> Key: FLINK-4559
> URL: https://issues.apache.org/jira/browse/FLINK-4559
> Project: Flink
>  Issue Type: Bug
>  Components: Kinesis Connector
>Affects Versions: 1.1.0, 1.1.1, 1.1.2
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Tzu-Li (Gordon) Tai
>Priority: Minor
> Fix For: 1.1.3
>
>
> If the {{AWS_CREDENTIAL_PROVIDER}} option is set to {{AUTO}}, 
> {{AWSUtils.getCredentialsProvider}} will return {{null}}, so 
> {{setCredentialsProvider}} should not be explicitly called on the internally 
> built {{KinesisProducerConfiguration}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-4194) Implement isEndOfStream() for KinesisDeserializationSchema

2016-09-03 Thread Tzu-Li (Gordon) Tai (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15461269#comment-15461269
 ] 

Tzu-Li (Gordon) Tai commented on FLINK-4194:


On second thought, I think we ought to discuss whether or not adding the 
function back to {{KinesisDeserializationSchema}} and breaking user code is a 
good idea now, since the Kinesis connector is now in a released version and 
would need a migration plan. Its not a highly requested feature at the moment, 
so I would say it is currently not worth the effort. Perhaps a good time to add 
this back is when it is actually possible to "stop" parallel consumption across 
the subtasks with an atomic checkpoint-and-stop, so an {{isEndOfStream()}} 
behaviour can be truly defined.

What do you think, [~rmetzger]?

> Implement isEndOfStream() for KinesisDeserializationSchema
> --
>
> Key: FLINK-4194
> URL: https://issues.apache.org/jira/browse/FLINK-4194
> Project: Flink
>  Issue Type: New Feature
>  Components: Kinesis Connector
>Reporter: Robert Metzger
>Assignee: Tzu-Li (Gordon) Tai
> Fix For: 1.2.0
>
>
> **Original JIRA title: KinesisDeserializationSchema.isEndOfStream() is never 
> called. The corresponding part of the code has been commented out with 
> reference to this JIRA.**
> The Kinesis connector does not respect the 
> {{KinesisDeserializationSchema.isEndOfStream()}} method.
> The purpose of this method is to stop consuming from a source, based on input 
> data.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (FLINK-4574) Strengthen fetch interval implementation in Kinesis consumer

2016-09-03 Thread Tzu-Li (Gordon) Tai (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-4574?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tzu-Li (Gordon) Tai updated FLINK-4574:
---
Description: 
As pointed out by [~rmetzger], right now the fetch interval implementation in 
the {{ShardConsumer}} class of the Kinesis consumer can lead to much longer 
interval times than specified by the user, ex. say the specified fetch interval 
is {{f}}, it takes {{x}} to complete a {{getRecords()}} call, and {{y}} to 
complete processing the fetched records for emitting, than the actual interval 
between each fetch is actually {{f+x+y}}.

The main problem with this is that we can never guarantee how much time has 
past since the last {{getRecords}} call, thus can not guarantee that returned 
shard iterators will not have expired the next time we use them, even if we 
limit the user-given value for {{f}} to not be longer than the iterator expire 
time.

I propose to improve this by, per {{ShardConsumer}}, use a 
{{ScheduledExecutorService}} to do the fixed-interval fetching, and a separate 
blocking queue that collects the fetched records for emitting.

  was:
As pointed out by [~rmetzger], right now the fetch interval implementation in 
the {{ShardConsumer}} class of the Kinesis consumer can lead to much longer 
interval times than specified by the user, ex. say the specified fetch interval 
is {{f}}, it takes {{x}} to complete a {{getRecords()}} call, and {{y}} to 
complete processing the fetched records for emitting, than the actual interval 
between each fetch is actually {{f+x+y}}.

The main problem with this is that we can never guarantee how much time has 
past since the last {{getRecords}} call, thus can not guarantee that returned 
shard iterators will not have expired the next time we use them, even if we 
limit the user-given value for {{f}} to not be longer than the iterator expire 
time.

I propose to change to, per {{ShardConsumer}}, use a 
{{ScheduledExecutorService}} to do the fixed-interval fetching, and a separate 
blocking queue that collects the fetched records for emitting.


> Strengthen fetch interval implementation in Kinesis consumer
> 
>
> Key: FLINK-4574
> URL: https://issues.apache.org/jira/browse/FLINK-4574
> Project: Flink
>  Issue Type: Improvement
>  Components: Kinesis Connector
>Affects Versions: 1.1.0
>Reporter: Tzu-Li (Gordon) Tai
> Fix For: 1.2.0, 1.1.3
>
>
> As pointed out by [~rmetzger], right now the fetch interval implementation in 
> the {{ShardConsumer}} class of the Kinesis consumer can lead to much longer 
> interval times than specified by the user, ex. say the specified fetch 
> interval is {{f}}, it takes {{x}} to complete a {{getRecords()}} call, and 
> {{y}} to complete processing the fetched records for emitting, than the 
> actual interval between each fetch is actually {{f+x+y}}.
> The main problem with this is that we can never guarantee how much time has 
> past since the last {{getRecords}} call, thus can not guarantee that returned 
> shard iterators will not have expired the next time we use them, even if we 
> limit the user-given value for {{f}} to not be longer than the iterator 
> expire time.
> I propose to improve this by, per {{ShardConsumer}}, use a 
> {{ScheduledExecutorService}} to do the fixed-interval fetching, and a 
> separate blocking queue that collects the fetched records for emitting.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (FLINK-4574) Strengthen fetch interval implementation in Kinesis consumer

2016-09-03 Thread Tzu-Li (Gordon) Tai (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-4574?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tzu-Li (Gordon) Tai updated FLINK-4574:
---
Description: 
As pointed out by [~rmetzger], right now the fetch interval implementation in 
the {{ShardConsumer}} class of the Kinesis consumer can lead to much longer 
interval times than specified by the user, ex. say the specified fetch interval 
is {{f}}, it takes {{x}} to complete a {{getRecords()}} call, and {{y}} to 
complete processing the fetched records for emitting, than the actual interval 
between each fetch is actually {{f+x+y}}.

The main problem with this is that we can never guarantee how much time has 
past since the last {{getRecords}} call, thus can not guarantee that returned 
shard iterators will not have expired the next time we use them, even if we 
limit the user-given value for {{f}} to not be longer than the iterator expire 
time.

I propose to change to, per {{ShardConsumer}}, use a 
{{ScheduledExecutorService}} to do the fixed-interval fetching, and a separate 
blocking queue that collects the fetched records for emitting.

  was:
As pointed out by [~rmetzger], right now the fetch interval implementation in 
the {{ShardConsumer}} class of the Kinesis consumer can lead to much longer 
interval times than specified by the user, ex. say the specified fetch interval 
is {{f}}, it takes {{x}} to complete a {{getRecords()}} call, and {{y}} to 
complete processing the fetched records for emitting, than the actual interval 
between each fetch is actually {{f+x+y}}.

The main problem with this is that we can never guarantee how much time has 
past since the last {{getRecords}} call, thus can not guarantee that returned 
shard iterators will not have expired the next time we use them.

I propose to change to, per {{ShardConsumer}}, use a 
{{ScheduledExecutorService}} to do the fixed-interval fetching, and a separate 
blocking queue that collects the fetched records for emitting.


> Strengthen fetch interval implementation in Kinesis consumer
> 
>
> Key: FLINK-4574
> URL: https://issues.apache.org/jira/browse/FLINK-4574
> Project: Flink
>  Issue Type: Improvement
>  Components: Kinesis Connector
>Affects Versions: 1.1.0
>Reporter: Tzu-Li (Gordon) Tai
> Fix For: 1.2.0, 1.1.3
>
>
> As pointed out by [~rmetzger], right now the fetch interval implementation in 
> the {{ShardConsumer}} class of the Kinesis consumer can lead to much longer 
> interval times than specified by the user, ex. say the specified fetch 
> interval is {{f}}, it takes {{x}} to complete a {{getRecords()}} call, and 
> {{y}} to complete processing the fetched records for emitting, than the 
> actual interval between each fetch is actually {{f+x+y}}.
> The main problem with this is that we can never guarantee how much time has 
> past since the last {{getRecords}} call, thus can not guarantee that returned 
> shard iterators will not have expired the next time we use them, even if we 
> limit the user-given value for {{f}} to not be longer than the iterator 
> expire time.
> I propose to change to, per {{ShardConsumer}}, use a 
> {{ScheduledExecutorService}} to do the fixed-interval fetching, and a 
> separate blocking queue that collects the fetched records for emitting.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (FLINK-4574) Strengthen fetch interval implementation in Kinesis consumer

2016-09-03 Thread Tzu-Li (Gordon) Tai (JIRA)
Tzu-Li (Gordon) Tai created FLINK-4574:
--

 Summary: Strengthen fetch interval implementation in Kinesis 
consumer
 Key: FLINK-4574
 URL: https://issues.apache.org/jira/browse/FLINK-4574
 Project: Flink
  Issue Type: Improvement
  Components: Kinesis Connector
Affects Versions: 1.1.0
Reporter: Tzu-Li (Gordon) Tai
 Fix For: 1.2.0, 1.1.3


As pointed out by [~rmetzger], right now the fetch interval implementation in 
the {{ShardConsumer}} class of the Kinesis consumer can lead to much longer 
interval times than specified by the user, ex. say the specified fetch interval 
is {{f}}, it takes {{x}} to complete a {{getRecords()}} call, and {{y}} to 
complete processing the fetched records for emitting, than the actual interval 
between each fetch is actually {{f+x+y}}.

The main problem with this is that we can never guarantee how much time has 
past since the last {{getRecords}} call, thus can not guarantee that returned 
shard iterators will not have expired the next time we use them.

I propose to change to, per {{ShardConsumer}}, use a 
{{ScheduledExecutorService}} to do the fixed-interval fetching, and a separate 
blocking queue that collects the fetched records for emitting.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-4572) Convert to negative in LongValueToIntValue

2016-09-03 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4572?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15461121#comment-15461121
 ] 

ASF GitHub Bot commented on FLINK-4572:
---

GitHub user greghogan opened a pull request:

https://github.com/apache/flink/pull/2469

[FLINK-4572] [gelly] Convert to negative in LongValueToIntValue

The Gelly drivers expect that scale 32 edges, represented by the lower 32 
bits of long values, can be converted to int values. Values between 2^31 and 
2^32 - 1 should be converted to negative integers.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/greghogan/flink 
4572_convert_to_negative_in_longvaluetointvalue

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/2469.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2469


commit 8a9fad5bb72e2f399b2afd0d095b6e29c9e5c9d0
Author: Greg Hogan 
Date:   2016-09-02T16:01:29Z

[FLINK-4572] [gelly] Convert to negative in LongValueToIntValue

The Gelly drivers expect that scale 32 edges, represented by the lower
32 bits of long values, can be converted to int values. Values between
2^31 and 2^32 - 1 should be converted to negative integers.




> Convert to negative in LongValueToIntValue
> --
>
> Key: FLINK-4572
> URL: https://issues.apache.org/jira/browse/FLINK-4572
> Project: Flink
>  Issue Type: Bug
>  Components: Gelly
>Affects Versions: 1.2.0
>Reporter: Greg Hogan
>Assignee: Greg Hogan
>Priority: Minor
>
> The Gelly drivers expect that scale 32 edges, represented by the lower 32 
> bits of {{long}} values, can be converted to {{int}} values. Values between 
> 2^31 and 2^32 - 1 should be converted to negative integers, which is not 
> supported by {{MathUtils.checkedDownCast}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request #2469: [FLINK-4572] [gelly] Convert to negative in LongVa...

2016-09-03 Thread greghogan
GitHub user greghogan opened a pull request:

https://github.com/apache/flink/pull/2469

[FLINK-4572] [gelly] Convert to negative in LongValueToIntValue

The Gelly drivers expect that scale 32 edges, represented by the lower 32 
bits of long values, can be converted to int values. Values between 2^31 and 
2^32 - 1 should be converted to negative integers.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/greghogan/flink 
4572_convert_to_negative_in_longvaluetointvalue

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/2469.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2469


commit 8a9fad5bb72e2f399b2afd0d095b6e29c9e5c9d0
Author: Greg Hogan 
Date:   2016-09-02T16:01:29Z

[FLINK-4572] [gelly] Convert to negative in LongValueToIntValue

The Gelly drivers expect that scale 32 edges, represented by the lower
32 bits of long values, can be converted to int values. Values between
2^31 and 2^32 - 1 should be converted to negative integers.




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-4561) replace all the scala version as a `scala.binary.version` property

2016-09-03 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-4561?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15460619#comment-15460619
 ] 

ASF GitHub Bot commented on FLINK-4561:
---

Github user shijinkui commented on the issue:

https://github.com/apache/flink/pull/2459
  
@chiwanpark @StephanEwen thanks for the reply.  dependency's version which 
contain scala 2.10 replace as property?


> replace all the scala version as a `scala.binary.version` property
> --
>
> Key: FLINK-4561
> URL: https://issues.apache.org/jira/browse/FLINK-4561
> Project: Flink
>  Issue Type: Improvement
>Reporter: shijinkui
>
> replace all the scala version(2.10) as a property `scala.binary.version` 
> defined in root pom properties. default scala version property is 2.10.
> modify:
> 1. dependency include scala version 
> 2. module defining include scala version
> 3. scala version upgrade to 2.11.8 from 2.11.7



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink issue #2459: [FLINK-4561] replace all the scala version as a `scala.bi...

2016-09-03 Thread shijinkui
Github user shijinkui commented on the issue:

https://github.com/apache/flink/pull/2459
  
@chiwanpark @StephanEwen thanks for the reply.  dependency's version which 
contain scala 2.10 replace as property?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-3580) Reintroduce Date/Time and implement scalar functions for it

2016-09-03 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-3580?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15460594#comment-15460594
 ] 

ASF GitHub Bot commented on FLINK-3580:
---

GitHub user twalthr opened a pull request:

https://github.com/apache/flink/pull/2468

[FLINK-3580] [table] Add OVERLAPS function

Thanks for contributing to Apache Flink. Before you open your pull request, 
please take the following check list into consideration.
If your changes take all of the items into account, feel free to open your 
pull request. For more information and/or questions please refer to the [How To 
Contribute guide](http://flink.apache.org/how-to-contribute.html).
In addition to going through the list, please provide a meaningful 
description of your changes.

- [x] General
  - The pull request references the related JIRA issue ("[FLINK-XXX] Jira 
title text")
  - The pull request addresses only one issue
  - Each commit in the PR has a meaningful commit message (including the 
JIRA id)

- [x] Documentation
  - Documentation has been added for new functionality
  - Old documentation affected by the pull request has been updated
  - JavaDoc for public methods has been added

- [x] Tests & Build
  - Functionality added by the pull request is covered by tests
  - `mvn clean verify` has been executed successfully locally or a Travis 
build has passed

This PR implements the SQL OVERLAPS function for Table API and SQL. It 
allows for checking if two anchored intervals overlap. The SQL documentation is 
missing as it will be reworked anyway shortly.



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/twalthr/flink FLINK-3580_OVERLAPS

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/2468.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2468


commit edc61d6f85005cad3904a319289669a1ade9c46d
Author: twalthr 
Date:   2016-09-03T06:00:58Z

[FLINK-3580] [table] Add OVERLAPS function




> Reintroduce Date/Time and implement scalar functions for it
> ---
>
> Key: FLINK-3580
> URL: https://issues.apache.org/jira/browse/FLINK-3580
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table API & SQL
>Reporter: Timo Walther
>Assignee: Timo Walther
>
> This task includes:
> {code}
> DATETIME_PLUS
> EXTRACT_DATE
> FLOOR
> CEIL
> CURRENT_TIME
> CURRENT_TIMESTAMP
> LOCALTIME
> LOCALTIMESTAMP
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request #2468: [FLINK-3580] [table] Add OVERLAPS function

2016-09-03 Thread twalthr
GitHub user twalthr opened a pull request:

https://github.com/apache/flink/pull/2468

[FLINK-3580] [table] Add OVERLAPS function

Thanks for contributing to Apache Flink. Before you open your pull request, 
please take the following check list into consideration.
If your changes take all of the items into account, feel free to open your 
pull request. For more information and/or questions please refer to the [How To 
Contribute guide](http://flink.apache.org/how-to-contribute.html).
In addition to going through the list, please provide a meaningful 
description of your changes.

- [x] General
  - The pull request references the related JIRA issue ("[FLINK-XXX] Jira 
title text")
  - The pull request addresses only one issue
  - Each commit in the PR has a meaningful commit message (including the 
JIRA id)

- [x] Documentation
  - Documentation has been added for new functionality
  - Old documentation affected by the pull request has been updated
  - JavaDoc for public methods has been added

- [x] Tests & Build
  - Functionality added by the pull request is covered by tests
  - `mvn clean verify` has been executed successfully locally or a Travis 
build has passed

This PR implements the SQL OVERLAPS function for Table API and SQL. It 
allows for checking if two anchored intervals overlap. The SQL documentation is 
missing as it will be reworked anyway shortly.



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/twalthr/flink FLINK-3580_OVERLAPS

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/2468.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2468


commit edc61d6f85005cad3904a319289669a1ade9c46d
Author: twalthr 
Date:   2016-09-03T06:00:58Z

[FLINK-3580] [table] Add OVERLAPS function




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---