[jira] [Resolved] (FLINK-8985) End-to-end test: CLI

2018-11-06 Thread Timo Walther (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-8985?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Timo Walther resolved FLINK-8985.
-
   Resolution: Fixed
Fix Version/s: 1.7.0

Fixed in 1.8: 44f217555f6ca72d1aa38d8626c1673a76c12b22, 
8106e7236edc0b8035adc92a53ddba79456bc05a
Fixed in 1.7: a73f979ea56d03a80d620a76bfd9078818d8c6d5, 
89a19ec21eb0bd2fc0d3f8ef23c1f0388b512905

> End-to-end test: CLI
> 
>
> Key: FLINK-8985
> URL: https://issues.apache.org/jira/browse/FLINK-8985
> Project: Flink
>  Issue Type: Sub-task
>  Components: Client, Tests
>Affects Versions: 1.5.0
>Reporter: Till Rohrmann
>Assignee: Rong Rong
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>
> We should add end-to-end test which verifies that all client commands are 
> working correctly.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-10807) KafkaConsumer still consume removed topic after changing topics list

2018-11-06 Thread Jiayi Liao (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10807?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jiayi Liao updated FLINK-10807:
---
Description: subscribedPartitionsToStartOffsets in KafkaConsumerBase is 
getting values from restoredState, which is initialized in initializeState and 
discovering partitions. However, if we remove a topic in topics list and 
restore the Flink program, the restoredState still keeps removed topic, and the 
fetcher still fetches the data of the removed topic.  (was: 
subscribedPartitionsToStartOffsets in KafkaConsumerBase is getting values from 
restoredState, which is initialized in initializeState and discovering 
partitions. However, if we remove a topic in topics list and restore the Flink 
program, the restoredState still keeps removed topic, and the fetcher will 
still fetch the data of the topic.)

> KafkaConsumer still consume removed topic after changing topics list
> 
>
> Key: FLINK-10807
> URL: https://issues.apache.org/jira/browse/FLINK-10807
> Project: Flink
>  Issue Type: Bug
>  Components: Kafka Connector
>Affects Versions: 1.6.2
>Reporter: Jiayi Liao
>Assignee: Jiayi Liao
>Priority: Major
>
> subscribedPartitionsToStartOffsets in KafkaConsumerBase is getting values 
> from restoredState, which is initialized in initializeState and discovering 
> partitions. However, if we remove a topic in topics list and restore the 
> Flink program, the restoredState still keeps removed topic, and the fetcher 
> still fetches the data of the removed topic.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-10807) KafkaConsumer still consume removed topic after changing topics list

2018-11-06 Thread Jiayi Liao (JIRA)
Jiayi Liao created FLINK-10807:
--

 Summary: KafkaConsumer still consume removed topic after changing 
topics list
 Key: FLINK-10807
 URL: https://issues.apache.org/jira/browse/FLINK-10807
 Project: Flink
  Issue Type: Bug
  Components: Kafka Connector
Affects Versions: 1.6.2
Reporter: Jiayi Liao
Assignee: Jiayi Liao


subscribedPartitionsToStartOffsets in KafkaConsumerBase is getting values from 
restoredState, which is initialized in initializeState and discovering 
partitions. However, if we remove a topic in topics list and restore the Flink 
program, the restoredState still keeps removed topic, and the fetcher will 
still fetch the data of the topic.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10790) Refactor all the StreamPartitioner implementations into runtime module

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10790?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=1660#comment-1660
 ] 

ASF GitHub Bot commented on FLINK-10790:


zhijiangW commented on issue #7037: [FLINK-10790][network] Refactor 
StreamPartitioner implementations into runtime module
URL: https://github.com/apache/flink/pull/7037#issuecomment-436525702
 
 
   @pnowojski , there are some problems for this issue now. I will update the 
codes later. Sorry for interrupting. I will cc you after ready.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Refactor all the StreamPartitioner implementations into runtime module
> --
>
> Key: FLINK-10790
> URL: https://issues.apache.org/jira/browse/FLINK-10790
> Project: Flink
>  Issue Type: Sub-task
>  Components: Network
>Affects Versions: 1.8.0
>Reporter: zhijiang
>Assignee: zhijiang
>Priority: Minor
>  Labels: pull-request-available
>
> The motivation is for refactoring {{ChannelSelector}} interface mentioned in 
> [FLINK-10662|https://issues.apache.org/jira/browse/FLINK-10662].
> The {{ChannelSelector}} implementations are separate for stream and batch 
> currently, resulting in many duplicated logics for selecting channels. The 
> root cause is that they are locating in different modules to not reuse the 
> existing ones.
> In general, the core runtime codes should be unified for both stream and 
> batch jobs as much as possible, then we can reduce the work overhead to 
> maintain the duplicated codes.
> This jira would migrate the existing partitioner package from 
> {{flink-streaming-jave}} to {{flink-runtime}}. Next we want to de-duplicate 
> implementations for {{RoundRobinChannelSelector}} and {{OutputEmitter}}.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] zhijiangW commented on issue #7037: [FLINK-10790][network] Refactor StreamPartitioner implementations into runtime module

2018-11-06 Thread GitBox
zhijiangW commented on issue #7037: [FLINK-10790][network] Refactor 
StreamPartitioner implementations into runtime module
URL: https://github.com/apache/flink/pull/7037#issuecomment-436525702
 
 
   @pnowojski , there are some problems for this issue now. I will update the 
codes later. Sorry for interrupting. I will cc you after ready.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Issue Comment Deleted] (FLINK-10384) Add Sinh math function supported in Table API and SQL

2018-11-06 Thread vinoyang (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

vinoyang updated FLINK-10384:
-
Comment: was deleted

(was: [~pnovi...@gmail.com] I think this issue could be closed.)

> Add Sinh math function supported in Table API and SQL
> -
>
> Key: FLINK-10384
> URL: https://issues.apache.org/jira/browse/FLINK-10384
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table API  SQL
>Reporter: vinoyang
>Assignee: vinoyang
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>
> like FLINK-10340 for adding Cosh math function



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (FLINK-10384) Add Sinh math function supported in Table API and SQL

2018-11-06 Thread vinoyang (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

vinoyang closed FLINK-10384.

   Resolution: Fixed
Fix Version/s: 1.7.0

> Add Sinh math function supported in Table API and SQL
> -
>
> Key: FLINK-10384
> URL: https://issues.apache.org/jira/browse/FLINK-10384
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table API  SQL
>Reporter: vinoyang
>Assignee: vinoyang
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>
> like FLINK-10340 for adding Cosh math function



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] tzulitai commented on issue #7028: [FLINK-10789] New serializer snapshots after 1.6 should implement TypeSerializerSnapshot

2018-11-06 Thread GitBox
tzulitai commented on issue #7028: [FLINK-10789] New serializer snapshots after 
1.6 should implement TypeSerializerSnapshot
URL: https://github.com/apache/flink/pull/7028#issuecomment-436501351
 
 
   Thanks for the review @StefanRRichter!
   Will proceed to merge this and also the 1.7 backport.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10789) Some new serializer snapshots added after 1.6 are not implementing the new TypeSerializerSnapshot interface

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16677657#comment-16677657
 ] 

ASF GitHub Bot commented on FLINK-10789:


tzulitai commented on issue #7028: [FLINK-10789] New serializer snapshots after 
1.6 should implement TypeSerializerSnapshot
URL: https://github.com/apache/flink/pull/7028#issuecomment-436501351
 
 
   Thanks for the review @StefanRRichter!
   Will proceed to merge this and also the 1.7 backport.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Some new serializer snapshots added after 1.6 are not implementing the new 
> TypeSerializerSnapshot interface
> ---
>
> Key: FLINK-10789
> URL: https://issues.apache.org/jira/browse/FLINK-10789
> Project: Flink
>  Issue Type: Improvement
>  Components: Type Serialization System
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Tzu-Li (Gordon) Tai
>Priority: Critical
>  Labels: pull-request-available
>
> In commit 
> [3787b89105c6a005ddca999b0b8a0356e812ede6|[https://github.com/apache/flink/commit/3787b89105c6a005ddca999b0b8a0356e812ede6],]
>  there were a few serializer snapshot implementations that were newly added, 
> namely:
> - {{LockableSerializerConfigSnapshot}}
> - {{ListViewSerializerConfigSnapshot}}
> - {{MapViewSerializerConfigSnapshot}}
> - {{ScalaEitherSerializerConfigSnapshot}}
> The problem is that instead of implementing the newer 
> {{TypeSerializerSnapshot}}, they are extending the legacy deprecated 
> {{TypeSerializerConfigSnapshot}}. This doesn't break things, but we will be 
> redundantly writing Java-serialized with the new snapshots, as well as 
> causing us to must introduce migration paths code later on after releasing 
> them.
> It would be more ideal to implement them properly as 
> {{TypeSerializerSnapshot}} for the release.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10791) Provide end-to-end test for Kafka 0.11 connector

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10791?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16677655#comment-16677655
 ] 

ASF GitHub Bot commented on FLINK-10791:


yanghua opened a new pull request #7038: [FLINK-10791] Provide end-to-end test 
for Kafka 0.11 connector
URL: https://github.com/apache/flink/pull/7038
 
 
   
   ## What is the purpose of the change
   
   *Provide end-to-end test for Kafka 0.11 connector*
   
   
   ## Brief change log
   
 - *Provide end-to-end test for Kafka 0.11 connector*
   
   ## Verifying this change
   
   This change is already covered by existing tests, such as *(please describe 
tests)*.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (**yes** / no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / **no**)
 - The serializers: (yes / **no** / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / **no** 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (yes / **no** / don't know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / **no**)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ **not documented**)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Provide end-to-end test for Kafka 0.11 connector
> 
>
> Key: FLINK-10791
> URL: https://issues.apache.org/jira/browse/FLINK-10791
> Project: Flink
>  Issue Type: Test
>  Components: E2E Tests
>Reporter: vinoyang
>Assignee: vinoyang
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-10791) Provide end-to-end test for Kafka 0.11 connector

2018-11-06 Thread ASF GitHub Bot (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10791?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-10791:
---
Labels: pull-request-available  (was: )

> Provide end-to-end test for Kafka 0.11 connector
> 
>
> Key: FLINK-10791
> URL: https://issues.apache.org/jira/browse/FLINK-10791
> Project: Flink
>  Issue Type: Test
>  Components: E2E Tests
>Reporter: vinoyang
>Assignee: vinoyang
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] yanghua opened a new pull request #7038: [FLINK-10791] Provide end-to-end test for Kafka 0.11 connector

2018-11-06 Thread GitBox
yanghua opened a new pull request #7038: [FLINK-10791] Provide end-to-end test 
for Kafka 0.11 connector
URL: https://github.com/apache/flink/pull/7038
 
 
   
   ## What is the purpose of the change
   
   *Provide end-to-end test for Kafka 0.11 connector*
   
   
   ## Brief change log
   
 - *Provide end-to-end test for Kafka 0.11 connector*
   
   ## Verifying this change
   
   This change is already covered by existing tests, such as *(please describe 
tests)*.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (**yes** / no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / **no**)
 - The serializers: (yes / **no** / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / **no** 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (yes / **no** / don't know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / **no**)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ **not documented**)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10790) Refactor all the StreamPartitioner implementations into runtime module

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10790?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16677646#comment-16677646
 ] 

ASF GitHub Bot commented on FLINK-10790:


zhijiangW commented on issue #7037: [FLINK-10790][network] Refactor 
StreamPartitioner implementations into runtime module
URL: https://github.com/apache/flink/pull/7037#issuecomment-436496303
 
 
   @pnowojski , this is the preparation for unifying implementations in both 
stream and batch modes. As I described in above purpose, I would submit the 
following PRs for de-duplication if you think this module refactor make sense.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Refactor all the StreamPartitioner implementations into runtime module
> --
>
> Key: FLINK-10790
> URL: https://issues.apache.org/jira/browse/FLINK-10790
> Project: Flink
>  Issue Type: Sub-task
>  Components: Network
>Affects Versions: 1.8.0
>Reporter: zhijiang
>Assignee: zhijiang
>Priority: Minor
>  Labels: pull-request-available
>
> The motivation is for refactoring {{ChannelSelector}} interface mentioned in 
> [FLINK-10662|https://issues.apache.org/jira/browse/FLINK-10662].
> The {{ChannelSelector}} implementations are separate for stream and batch 
> currently, resulting in many duplicated logics for selecting channels. The 
> root cause is that they are locating in different modules to not reuse the 
> existing ones.
> In general, the core runtime codes should be unified for both stream and 
> batch jobs as much as possible, then we can reduce the work overhead to 
> maintain the duplicated codes.
> This jira would migrate the existing partitioner package from 
> {{flink-streaming-jave}} to {{flink-runtime}}. Next we want to de-duplicate 
> implementations for {{RoundRobinChannelSelector}} and {{OutputEmitter}}.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] zhijiangW commented on issue #7037: [FLINK-10790][network] Refactor StreamPartitioner implementations into runtime module

2018-11-06 Thread GitBox
zhijiangW commented on issue #7037: [FLINK-10790][network] Refactor 
StreamPartitioner implementations into runtime module
URL: https://github.com/apache/flink/pull/7037#issuecomment-436496303
 
 
   @pnowojski , this is the preparation for unifying implementations in both 
stream and batch modes. As I described in above purpose, I would submit the 
following PRs for de-duplication if you think this module refactor make sense.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-10790) Refactor all the StreamPartitioner implementations into runtime module

2018-11-06 Thread ASF GitHub Bot (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10790?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-10790:
---
Labels: pull-request-available  (was: )

> Refactor all the StreamPartitioner implementations into runtime module
> --
>
> Key: FLINK-10790
> URL: https://issues.apache.org/jira/browse/FLINK-10790
> Project: Flink
>  Issue Type: Sub-task
>  Components: Network
>Affects Versions: 1.8.0
>Reporter: zhijiang
>Assignee: zhijiang
>Priority: Minor
>  Labels: pull-request-available
>
> The motivation is for refactoring {{ChannelSelector}} interface mentioned in 
> [FLINK-10662|https://issues.apache.org/jira/browse/FLINK-10662].
> The {{ChannelSelector}} implementations are separate for stream and batch 
> currently, resulting in many duplicated logics for selecting channels. The 
> root cause is that they are locating in different modules to not reuse the 
> existing ones.
> In general, the core runtime codes should be unified for both stream and 
> batch jobs as much as possible, then we can reduce the work overhead to 
> maintain the duplicated codes.
> This jira would migrate the existing partitioner package from 
> {{flink-streaming-jave}} to {{flink-runtime}}. Next we want to de-duplicate 
> implementations for {{RoundRobinChannelSelector}} and {{OutputEmitter}}.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10790) Refactor all the StreamPartitioner implementations into runtime module

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10790?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16677643#comment-16677643
 ] 

ASF GitHub Bot commented on FLINK-10790:


zhijiangW opened a new pull request #7037: [FLINK-10790][network] Refactor 
StreamPartitioner implementations into runtime module
URL: https://github.com/apache/flink/pull/7037
 
 
   ## What is the purpose of the change
   
   *Before refactoring `ChannelSelector` interface mentioned in 
[FLINK-10662](https://issues.apache.org/jira/browse/FLINK-10622), we found 
there are many duplicated implementations for `ChannelSelector` in stream and 
batch modes separately. The root cause is that they are locating in different 
modules, so can not reference each other.*
   
   *The core runtime codes should be unified for both stream and batch modes as 
much as possible, then we can reduce the work overhead to maintain the 
duplicated codes, especially for the following refactor.*
   
   *Next we would remove duplicated implementations for `OutputEmitter` and 
`RoundRobinChannelSelector`.*
   
   ## Brief change log
   
 - *Refactor the existing partitioner package from `flink-streaming-jave` 
to `flink-runtime`.*
 - *Remove `StreamRecord` type explicitly from `SerializationDelegate`.*
 - *Modify related tests based on refactor changes.*
   
   
   ## Verifying this change
   
   *This change is already covered by existing tests.*
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (yes)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Refactor all the StreamPartitioner implementations into runtime module
> --
>
> Key: FLINK-10790
> URL: https://issues.apache.org/jira/browse/FLINK-10790
> Project: Flink
>  Issue Type: Sub-task
>  Components: Network
>Affects Versions: 1.8.0
>Reporter: zhijiang
>Assignee: zhijiang
>Priority: Minor
>  Labels: pull-request-available
>
> The motivation is for refactoring {{ChannelSelector}} interface mentioned in 
> [FLINK-10662|https://issues.apache.org/jira/browse/FLINK-10662].
> The {{ChannelSelector}} implementations are separate for stream and batch 
> currently, resulting in many duplicated logics for selecting channels. The 
> root cause is that they are locating in different modules to not reuse the 
> existing ones.
> In general, the core runtime codes should be unified for both stream and 
> batch jobs as much as possible, then we can reduce the work overhead to 
> maintain the duplicated codes.
> This jira would migrate the existing partitioner package from 
> {{flink-streaming-jave}} to {{flink-runtime}}. Next we want to de-duplicate 
> implementations for {{RoundRobinChannelSelector}} and {{OutputEmitter}}.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] zhijiangW opened a new pull request #7037: [FLINK-10790][network] Refactor StreamPartitioner implementations into runtime module

2018-11-06 Thread GitBox
zhijiangW opened a new pull request #7037: [FLINK-10790][network] Refactor 
StreamPartitioner implementations into runtime module
URL: https://github.com/apache/flink/pull/7037
 
 
   ## What is the purpose of the change
   
   *Before refactoring `ChannelSelector` interface mentioned in 
[FLINK-10662](https://issues.apache.org/jira/browse/FLINK-10622), we found 
there are many duplicated implementations for `ChannelSelector` in stream and 
batch modes separately. The root cause is that they are locating in different 
modules, so can not reference each other.*
   
   *The core runtime codes should be unified for both stream and batch modes as 
much as possible, then we can reduce the work overhead to maintain the 
duplicated codes, especially for the following refactor.*
   
   *Next we would remove duplicated implementations for `OutputEmitter` and 
`RoundRobinChannelSelector`.*
   
   ## Brief change log
   
 - *Refactor the existing partitioner package from `flink-streaming-jave` 
to `flink-runtime`.*
 - *Remove `StreamRecord` type explicitly from `SerializationDelegate`.*
 - *Modify related tests based on refactor changes.*
   
   
   ## Verifying this change
   
   *This change is already covered by existing tests.*
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (yes)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-10806) Support multiple consuming offsets when discovering a new topic

2018-11-06 Thread Jiayi Liao (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10806?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jiayi Liao updated FLINK-10806:
---
Summary: Support multiple consuming offsets when discovering a new topic  
(was: Add optional consuming parameters when discovering a new topic)

> Support multiple consuming offsets when discovering a new topic
> ---
>
> Key: FLINK-10806
> URL: https://issues.apache.org/jira/browse/FLINK-10806
> Project: Flink
>  Issue Type: Improvement
>  Components: Kafka Connector
>Affects Versions: 1.6.2
>Reporter: Jiayi Liao
>Assignee: Jiayi Liao
>Priority: Major
>
> In KafkaConsumerBase, we discover the TopicPartitions and compare them with 
> the restoredState. It's reasonable when a topic's partitions scaled. However, 
> if we add a new topic which has too much data and restore the Flink program, 
> the data of the new topic will be consumed from the start, which may not be 
> what we want.  I think this should be an option for developers.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-10806) Add optional consuming parameters when discovering a new topic

2018-11-06 Thread Jiayi Liao (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10806?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jiayi Liao updated FLINK-10806:
---
Summary: Add optional consuming parameters when discovering a new topic  
(was: Support consume from earliest offset with a new topic)

> Add optional consuming parameters when discovering a new topic
> --
>
> Key: FLINK-10806
> URL: https://issues.apache.org/jira/browse/FLINK-10806
> Project: Flink
>  Issue Type: Improvement
>  Components: Kafka Connector
>Affects Versions: 1.6.2
>Reporter: Jiayi Liao
>Assignee: Jiayi Liao
>Priority: Major
>
> In KafkaConsumerBase, we discover the TopicPartitions and compare them with 
> the restoredState. It's reasonable when a topic's partitions scaled. However, 
> if we add a new topic which has too much data and restore the Flink program, 
> the data of the new topic will be consumed from the start, which may not be 
> what we want.  I think this should be an option for developers.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-10806) Support consume from earliest offset with a new topic

2018-11-06 Thread Jiayi Liao (JIRA)
Jiayi Liao created FLINK-10806:
--

 Summary: Support consume from earliest offset with a new topic
 Key: FLINK-10806
 URL: https://issues.apache.org/jira/browse/FLINK-10806
 Project: Flink
  Issue Type: Improvement
  Components: Kafka Connector
Affects Versions: 1.6.2
Reporter: Jiayi Liao
Assignee: Jiayi Liao


In KafkaConsumerBase, we discover the TopicPartitions and compare them with the 
restoredState. It's reasonable when a topic's partitions scaled. However, if we 
add a new topic which has too much data and restore the Flink program, the data 
of the new topic will be consumed from the start, which may not be what we 
want.  I think this should be an option for developers.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10706) Update the Flink Dashboard and its components

2018-11-06 Thread Yadong Xie (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10706?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16677570#comment-16677570
 ] 

Yadong Xie commented on FLINK-10706:


Hi [~fwollert], would you please assign this issue to me?

In order to make this change more easier to be reviewed, I plan to add more 
comments on the codes and submit them step by step according to different 
functional modules.

Thanks a lot for your efforts

> Update the Flink Dashboard and its components
> -
>
> Key: FLINK-10706
> URL: https://issues.apache.org/jira/browse/FLINK-10706
> Project: Flink
>  Issue Type: Sub-task
>  Components: Webfrontend
>Affects Versions: 1.6.2
>Reporter: Fabian Wollert
>Assignee: Fabian Wollert
>Priority: Major
> Fix For: 1.8.0
>
>
> The Flink Dashboard uses currently Angular 1, which had its successor coming 
> out two years ago. Its expected that Angular 1 (or Angular.js how it is 
> called now) will ceise to exist in the future, due to its successor Angular 
> 2-7 and React being the more actively developed platforms.
> We should move to Angular 7 or React.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-9699) Add api to replace registered table

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-9699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16677502#comment-16677502
 ] 

ASF GitHub Bot commented on FLINK-9699:
---

zjffdu commented on issue #6236: [FLINK-9699] [table] Add api to replace 
registered table
URL: https://github.com/apache/flink/pull/6236#issuecomment-436471468
 
 
   Thanks @hequn8128 @yanghua `registerOrReplaceXXX` make sense for me. I will 
update the PR.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Add api to replace registered table
> ---
>
> Key: FLINK-9699
> URL: https://issues.apache.org/jira/browse/FLINK-9699
> Project: Flink
>  Issue Type: Improvement
>Reporter: Jeff Zhang
>Assignee: Jeff Zhang
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] zjffdu commented on issue #6236: [FLINK-9699] [table] Add api to replace registered table

2018-11-06 Thread GitBox
zjffdu commented on issue #6236: [FLINK-9699] [table] Add api to replace 
registered table
URL: https://github.com/apache/flink/pull/6236#issuecomment-436471468
 
 
   Thanks @hequn8128 @yanghua `registerOrReplaceXXX` make sense for me. I will 
update the PR.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10802) Add metrics on vertex in job graph picture show

2018-11-06 Thread lining (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16677495#comment-16677495
 ] 

lining commented on FLINK-10802:


Hi [~till.rohrmann], now user see something need click many page. If we can add 
something import on the vertex will efficiency. The problem about overload, we 
can control the refresh rate. Or add job dashbord for user can see overview of 
job. Such as back pressure, checkpoints, failover.

> Add metrics on vertex in job graph picture show
> ---
>
> Key: FLINK-10802
> URL: https://issues.apache.org/jira/browse/FLINK-10802
> Project: Flink
>  Issue Type: Improvement
>  Components: Webfrontend
>Reporter: lining
>Assignee: lining
>Priority: Major
> Attachments: image-2018-11-06-18-12-16-914.png
>
>
> Hi. Now, if someone want see  job metrics, have to click job overview. if we 
> can add some import metrics on graph, can impore it.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-10644) Batch Job: Speculative execution

2018-11-06 Thread JIN SUN (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10644?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

JIN SUN updated FLINK-10644:

Description: 
Strugglers/outlier are tasks that run slower than most of the all tasks in a 
Batch Job, this somehow impact job latency, as pretty much this straggler will 
be in the critical path of the job and become as the bottleneck.

Tasks may be slow for various reasons, including hardware degradation, or 
software mis-configuration, or noise neighboring. It's hard for JM to predict 
the runtime.

To reduce the overhead of strugglers, other system such as Hadoop/Tez, Spark 
has *_speculative execution_*. Speculative execution is a health-check 
procedure that checks for tasks to be speculated, i.e. running slower in a 
ExecutionJobVertex than the median of all successfully completed tasks in that 
EJV, Such slow tasks will be re-submitted to another TM. It will not stop the 
slow tasks, but run a new copy in parallel. And will kill the others if one of 
them complete.

This JIRA is an umbrella to apply this kind of idea in FLINK. Details will be 
append later.

  was:
Strugglers/outlier are tasks that run slower than most of the all tasks in a 
Batch Job, this somehow impact job latency, as pretty much this straggler will 
be in the critical path of the job and become as the bottleneck.

Tasks may be slow for various reasons, including hardware degradation, or 
software mis-configuration, or noise neighboring. It's hard for JM to predict 
the runtime.

To reduce the overhead of strugglers, other system such as Hadoop/Tez, Spark 
has *_speculative execution_*. Speculative execution is a health-check 
procedure that checks for tasks to be speculated, i.e. running slower in a 
ExecutionJobVertex than the median of all successfully completed tasks in that 
EJV, Such slow tasks will be re-submitted to another TM. It will not stop the 
slow tasks, but run a new copy in parallel. And will kill the others if one of 
them complete.

This JIRA is an umbrella to apply this kind of idea in FLINK. Details will be 
append later.

 

the document contribute by is here: 
[https://docs.google.com/document/d/1X_Pfo4WcO-TEZmmVTTYNn44LQg5gnFeeaeqM7ZNLQ7M/edit]
 


> Batch Job: Speculative execution
> 
>
> Key: FLINK-10644
> URL: https://issues.apache.org/jira/browse/FLINK-10644
> Project: Flink
>  Issue Type: New Feature
>  Components: JobManager
>Reporter: JIN SUN
>Assignee: JIN SUN
>Priority: Major
> Fix For: 1.8.0
>
>
> Strugglers/outlier are tasks that run slower than most of the all tasks in a 
> Batch Job, this somehow impact job latency, as pretty much this straggler 
> will be in the critical path of the job and become as the bottleneck.
> Tasks may be slow for various reasons, including hardware degradation, or 
> software mis-configuration, or noise neighboring. It's hard for JM to predict 
> the runtime.
> To reduce the overhead of strugglers, other system such as Hadoop/Tez, Spark 
> has *_speculative execution_*. Speculative execution is a health-check 
> procedure that checks for tasks to be speculated, i.e. running slower in a 
> ExecutionJobVertex than the median of all successfully completed tasks in 
> that EJV, Such slow tasks will be re-submitted to another TM. It will not 
> stop the slow tasks, but run a new copy in parallel. And will kill the others 
> if one of them complete.
> This JIRA is an umbrella to apply this kind of idea in FLINK. Details will be 
> append later.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-10644) Batch Job: Speculative execution

2018-11-06 Thread JIN SUN (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10644?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

JIN SUN updated FLINK-10644:

Description: 
Strugglers/outlier are tasks that run slower than most of the all tasks in a 
Batch Job, this somehow impact job latency, as pretty much this straggler will 
be in the critical path of the job and become as the bottleneck.

Tasks may be slow for various reasons, including hardware degradation, or 
software mis-configuration, or noise neighboring. It's hard for JM to predict 
the runtime.

To reduce the overhead of strugglers, other system such as Hadoop/Tez, Spark 
has *_speculative execution_*. Speculative execution is a health-check 
procedure that checks for tasks to be speculated, i.e. running slower in a 
ExecutionJobVertex than the median of all successfully completed tasks in that 
EJV, Such slow tasks will be re-submitted to another TM. It will not stop the 
slow tasks, but run a new copy in parallel. And will kill the others if one of 
them complete.

This JIRA is an umbrella to apply this kind of idea in FLINK. Details will be 
append later.

 

the document contribute by is here: 
[https://docs.google.com/document/d/1X_Pfo4WcO-TEZmmVTTYNn44LQg5gnFeeaeqM7ZNLQ7M/edit]
 

  was:
Strugglers/outlier are tasks that run slower than most of the all tasks in a 
Batch Job, this somehow impact job latency, as pretty much this straggler will 
be in the critical path of the job and become as the bottleneck. 

Tasks may be slow for various reasons, including hardware degradation, or 
software mis-configuration, or noise neighboring. It's hard for JM to predict 
the runtime. 

To reduce the overhead of strugglers, other system such as Hadoop/Tez, Spark 
has *_speculative execution_*. Speculative execution is a health-check 
procedure that checks for tasks to be speculated, i.e. running slower in a 
ExecutionJobVertex than the median of all successfully completed tasks in that 
EJV, Such slow tasks will be re-submitted to another TM. It will not stop the 
slow tasks, but run a new copy in parallel. And will kill the others if one of 
them complete. 

This JIRA is an umbrella to apply this kind of idea in FLINK. Details will be 
append later.


> Batch Job: Speculative execution
> 
>
> Key: FLINK-10644
> URL: https://issues.apache.org/jira/browse/FLINK-10644
> Project: Flink
>  Issue Type: New Feature
>  Components: JobManager
>Reporter: JIN SUN
>Assignee: JIN SUN
>Priority: Major
> Fix For: 1.8.0
>
>
> Strugglers/outlier are tasks that run slower than most of the all tasks in a 
> Batch Job, this somehow impact job latency, as pretty much this straggler 
> will be in the critical path of the job and become as the bottleneck.
> Tasks may be slow for various reasons, including hardware degradation, or 
> software mis-configuration, or noise neighboring. It's hard for JM to predict 
> the runtime.
> To reduce the overhead of strugglers, other system such as Hadoop/Tez, Spark 
> has *_speculative execution_*. Speculative execution is a health-check 
> procedure that checks for tasks to be speculated, i.e. running slower in a 
> ExecutionJobVertex than the median of all successfully completed tasks in 
> that EJV, Such slow tasks will be re-submitted to another TM. It will not 
> stop the slow tasks, but run a new copy in parallel. And will kill the others 
> if one of them complete.
> This JIRA is an umbrella to apply this kind of idea in FLINK. Details will be 
> append later.
>  
> the document contribute by is here: 
> [https://docs.google.com/document/d/1X_Pfo4WcO-TEZmmVTTYNn44LQg5gnFeeaeqM7ZNLQ7M/edit]
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (FLINK-10671) rest monitoring api Savepoint status call fails if akka.ask.timeout < checkpoint duration

2018-11-06 Thread Gary Yao (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10671?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao closed FLINK-10671.

Resolution: Not A Problem

Please reopen if necessary. 

> rest monitoring api Savepoint status call fails if akka.ask.timeout < 
> checkpoint duration
> -
>
> Key: FLINK-10671
> URL: https://issues.apache.org/jira/browse/FLINK-10671
> Project: Flink
>  Issue Type: Bug
>  Components: REST
>Affects Versions: 1.6.1
>Reporter: Cliff Resnick
>Assignee: Gary Yao
>Priority: Minor
>
> Hi,
>  
> There seems to be a problem with REST monitoring API:
> |/jobs/:jobid/savepoints/:triggerid|
>  
> The problem is that when the Savepoint represented by {{:triggerid}} is 
> called with {{cancel=true}} the above status call seems to fail if the 
> savepoint duration exceeds {{akka.ask.timeout}} value.
>  
> Below is a log in which I invoke "cancel with savepoint" then poll the above 
> endpoint for status at 2 second intervals. {{akka.ask.timout}} is set for 
> twenty seconds. The error is repeatable at various values of 
> {{akka.ask.timeout}}.
>  
> {noformat}
> 2018/10/24 19:42:25 savepoint id 925964b35b2d501f4a45b714eca0a2ca is 
> IN_PROGRESS
> 2018/10/24 19:42:27 savepoint id 925964b35b2d501f4a45b714eca0a2ca is 
> IN_PROGRESS
> 2018/10/24 19:42:29 savepoint id 925964b35b2d501f4a45b714eca0a2ca is 
> IN_PROGRESS
> 2018/10/24 19:42:31 savepoint id 925964b35b2d501f4a45b714eca0a2ca is 
> IN_PROGRESS
> 2018/10/24 19:42:33 savepoint id 925964b35b2d501f4a45b714eca0a2ca is 
> IN_PROGRESS
> 2018/10/24 19:42:35 savepoint id 925964b35b2d501f4a45b714eca0a2ca is 
> IN_PROGRESS
> 2018/10/24 19:42:37 savepoint id 925964b35b2d501f4a45b714eca0a2ca is 
> IN_PROGRESS
> 2018/10/24 19:42:39 savepoint id 925964b35b2d501f4a45b714eca0a2ca is 
> IN_PROGRESS
> 2018/10/24 19:42:41 savepoint id 925964b35b2d501f4a45b714eca0a2ca is 
> IN_PROGRESS
> 2018/10/24 19:42:43 savepoint id 925964b35b2d501f4a45b714eca0a2ca is 
> IN_PROGRESS
> 2018/10/24 19:42:45 Cancel with Savepoint may have failed: 
> java.util.concurrent.CompletionException: akka.pattern.AskTimeoutException: 
> Ask timed out on [Actor[akka://flink/user/jobmanager_0#-234856817]] after 
> [2 ms]. Sender[null] sent message of type 
> "org.apache.flink.runtime.rpc.messages.LocalFencedMessage".
> at 
> java.util.concurrent.CompletableFuture.encodeRelay(CompletableFuture.java:326)
> at 
> java.util.concurrent.CompletableFuture.completeRelay(CompletableFuture.java:338)
> at java.util.concurrent.CompletableFuture.uniRelay(CompletableFuture.java:911)
> at 
> java.util.concurrent.CompletableFuture$UniRelay.tryFire(CompletableFuture.java:899)
> at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
> at 
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
> at 
> org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:770)
> at akka.dispatch.OnComplete.internal(Future.scala:258)
> at akka.dispatch.OnComplete.internal(Future.scala:256)
> at akka.dispatch.japi$CallbackBridge.apply(Future.scala:186)
> at akka.dispatch.japi$CallbackBridge.apply(Future.scala:183)
> at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
> at 
> org.apache.flink.runtime.concurrent.Executors$DirectExecutionContext.execute(Executors.java:83)
> at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44)
> at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252)
> at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:603)
> at akka.actor.Scheduler$$anon$4.run(Scheduler.scala:126)
> at 
> scala.concurrent.Future$InternalCallbackExecutor$.unbatchedExecute(Future.scala:601)
> at scala.concurrent.BatchingExecutor$class.execute(BatchingExecutor.scala:109)
> at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:599)
> at 
> akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(LightArrayRevolverScheduler.scala:329)
> at 
> akka.actor.LightArrayRevolverScheduler$$anon$4.executeBucket$1(LightArrayRevolverScheduler.scala:280)
> at 
> akka.actor.LightArrayRevolverScheduler$$anon$4.nextTick(LightArrayRevolverScheduler.scala:284)
> at 
> akka.actor.LightArrayRevolverScheduler$$anon$4.run(LightArrayRevolverScheduler.scala:236)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: akka.pattern.AskTimeoutException: Ask timed out on 
> [Actor[akka://flink/user/jobmanager_0#-234856817]] after [2 ms]. 
> Sender[null] sent message of type 
> "org.apache.flink.runtime.rpc.messages.LocalFencedMessage".
> at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:604)
> ... 9 more
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10356) Add sanity checks to SpillingAdaptiveSpanningRecordDeserializer

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10356?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16677127#comment-16677127
 ] 

ASF GitHub Bot commented on FLINK-10356:


NicoK commented on issue #6705: [FLINK-10356][network] add sanity checks to 
SpillingAdaptiveSpanningRecordDeserializer
URL: https://github.com/apache/flink/pull/6705#issuecomment-436356439
 
 
   Thanks @zhijiangW and @pnowojski for the reviews. I have addressed the 
comments and committed some changes. Please have a look whether these look 
better to you.
   
   I'm honestly not sure the additional code in the tests is worth the 
de-duplication...but maybe it is better readable now.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Add sanity checks to SpillingAdaptiveSpanningRecordDeserializer
> ---
>
> Key: FLINK-10356
> URL: https://issues.apache.org/jira/browse/FLINK-10356
> Project: Flink
>  Issue Type: Sub-task
>  Components: Network
>Affects Versions: 1.5.0, 1.5.1, 1.5.2, 1.5.3, 1.5.4, 1.6.0, 1.6.1, 1.7.0
>Reporter: Nico Kruber
>Assignee: Nico Kruber
>Priority: Major
>  Labels: pull-request-available
>
> {{SpillingAdaptiveSpanningRecordDeserializer}} doesn't have any consistency 
> checks for usage calls or serializers behaving properly, e.g. to read only as 
> many bytes as available/promised for that record. At least these checks 
> should be added:
>  # Check that buffers have not been read from yet before adding them (this is 
> an invariant {{SpillingAdaptiveSpanningRecordDeserializer}} works with and 
> from what I can see, it is followed now.
>  # Check that after deserialization, we actually consumed {{recordLength}} 
> bytes
>  ** If not, in the spanning deserializer, we currently simply skip the 
> remaining bytes.
>  ** But in the non-spanning deserializer, we currently continue from the 
> wrong offset.
>  # Protect against {{setNextBuffer}} being called before draining all 
> available records



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] NicoK commented on issue #6705: [FLINK-10356][network] add sanity checks to SpillingAdaptiveSpanningRecordDeserializer

2018-11-06 Thread GitBox
NicoK commented on issue #6705: [FLINK-10356][network] add sanity checks to 
SpillingAdaptiveSpanningRecordDeserializer
URL: https://github.com/apache/flink/pull/6705#issuecomment-436356439
 
 
   Thanks @zhijiangW and @pnowojski for the reviews. I have addressed the 
comments and committed some changes. Please have a look whether these look 
better to you.
   
   I'm honestly not sure the additional code in the tests is worth the 
de-duplication...but maybe it is better readable now.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10768) Move external catalog related code from TableEnvironment to CatalogManager

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10768?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16677099#comment-16677099
 ] 

ASF GitHub Bot commented on FLINK-10768:


bowenli86 commented on issue #7011: [FLINK-10768][Table & SQL] Move external 
catalog related code from TableEnvironment to CatalogManager
URL: https://github.com/apache/flink/pull/7011#issuecomment-436350350
 
 
   holding on this PR as there are some related discussion going on at[ the 
design 
doc](https://docs.google.com/document/d/1SkppRD_rE3uOKSN-LuZCqn4f7dz0zW5aa6T_hBZq5_o/edit)


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Move external catalog related code from TableEnvironment to CatalogManager
> --
>
> Key: FLINK-10768
> URL: https://issues.apache.org/jira/browse/FLINK-10768
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table API  SQL
>Reporter: Bowen Li
>Assignee: Bowen Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.8.0
>
>
> Add a new CatalogManager class and port existing Calcite-directly-related 
> code from TableEnvironment into CatalogManager.
> Background: there are two parallel efforts going on right now - FLINK-10686, 
> driven by Timo, includes moving external catalogs APIs from flink-table to 
> flink-table-common, also from Scala to Java; FLINK-10744 I'm working on right 
> now to integrate Flink with Hive and enhance external catalog functionality.
> As discussed with @twalthr in FLINK-10689, we'd better parallelize these 
> efforts while introducing minimal overhead for integrating them later. Our 
> agreed way is to writing new code/feature related to external catalogs/hive 
> in Java in flink-table first then move to other module like 
> flink-table-common, this way we can minimize migration efforts. If existing 
> classes are modified for a feature we can start migrating them to Java in a 
> separate commit first and then perform the actual feature changes, and 
> migrated classes can be placed in flink-table/src/main/java until we find a 
> better module structure.
> Thus, this is NOT a feature, but purely refactor, thus no new functions 
> should be introduced. It acts the pre-requisite for FLINK-10698



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] bowenli86 commented on issue #7011: [FLINK-10768][Table & SQL] Move external catalog related code from TableEnvironment to CatalogManager

2018-11-06 Thread GitBox
bowenli86 commented on issue #7011: [FLINK-10768][Table & SQL] Move external 
catalog related code from TableEnvironment to CatalogManager
URL: https://github.com/apache/flink/pull/7011#issuecomment-436350350
 
 
   holding on this PR as there are some related discussion going on at[ the 
design 
doc](https://docs.google.com/document/d/1SkppRD_rE3uOKSN-LuZCqn4f7dz0zW5aa6T_hBZq5_o/edit)


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-10699) Create FlinkHmsCatalog for persistent Flink meta objects using Hive metastore as a registry

2018-11-06 Thread Bowen Li (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bowen Li updated FLINK-10699:
-
Summary: Create FlinkHmsCatalog for persistent Flink meta objects using 
Hive metastore as a registry  (was: Create a catalog implementation for 
persistent Flink meta objects using Hive metastore as a registry)

> Create FlinkHmsCatalog for persistent Flink meta objects using Hive metastore 
> as a registry
> ---
>
> Key: FLINK-10699
> URL: https://issues.apache.org/jira/browse/FLINK-10699
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table API  SQL
>Affects Versions: 1.6.1
>Reporter: Xuefu Zhang
>Assignee: Bowen Li
>Priority: Major
>
> Similar to FLINK-10697, but using Hive metastore as persistent storage.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-10531) State TTL RocksDb backend end-to-end test failed on Travis

2018-11-06 Thread Kostas Kloudas (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas updated FLINK-10531:
---
Fix Version/s: 1.7.0

> State TTL RocksDb backend end-to-end test failed on Travis
> --
>
> Key: FLINK-10531
> URL: https://issues.apache.org/jira/browse/FLINK-10531
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.6.1
>Reporter: Till Rohrmann
>Assignee: Kostas Kloudas
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Fix For: 1.7.0
>
>
> The {{State TTL RocksDb backend end-to-end test}} end-to-end test failed on 
> Travis.
> https://travis-ci.org/apache/flink/jobs/438226190
> https://api.travis-ci.org/v3/job/438226190/log.txt



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10531) State TTL RocksDb backend end-to-end test failed on Travis

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16677023#comment-16677023
 ] 

ASF GitHub Bot commented on FLINK-10531:


kl0u commented on issue #7036: [FLINK-10531][e2e] Fix unstable TTL end-to-end 
test.
URL: https://github.com/apache/flink/pull/7036#issuecomment-436331588
 
 
   R @azagrebin 


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> State TTL RocksDb backend end-to-end test failed on Travis
> --
>
> Key: FLINK-10531
> URL: https://issues.apache.org/jira/browse/FLINK-10531
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.6.1
>Reporter: Till Rohrmann
>Assignee: Kostas Kloudas
>Priority: Critical
>  Labels: pull-request-available, test-stability
>
> The {{State TTL RocksDb backend end-to-end test}} end-to-end test failed on 
> Travis.
> https://travis-ci.org/apache/flink/jobs/438226190
> https://api.travis-ci.org/v3/job/438226190/log.txt



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] kl0u commented on issue #7036: [FLINK-10531][e2e] Fix unstable TTL end-to-end test.

2018-11-06 Thread GitBox
kl0u commented on issue #7036: [FLINK-10531][e2e] Fix unstable TTL end-to-end 
test.
URL: https://github.com/apache/flink/pull/7036#issuecomment-436331588
 
 
   R @azagrebin 


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10531) State TTL RocksDb backend end-to-end test failed on Travis

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16677018#comment-16677018
 ] 

ASF GitHub Bot commented on FLINK-10531:


kl0u opened a new pull request #7036: [FLINK-10531][e2e] Fix unstable TTL 
end-to-end test.
URL: https://github.com/apache/flink/pull/7036
 
 
   As describe on the JIRA, the problem is that the clock on the machine on 
Travis seems to have jumped backwards.
   
   This resulted in a mismatch between the elements discarded as expired by 
Flink's internal TTL mechanism, and the "user-code" in the test that computes 
the expired elements based on the timestamp of the latest update.
   
   I repeat the explanation from the JIRA here for reference:
   
   
---
   
   The way the test works is the following. For a given state:
   1) it keeps the state with TTL
   2) and keeps all the updates to that state irrespective of their timestamp 
(e.g. all the elements added in a ListState)
   
   To verify correctness:
   1) it fetches the state from Flink, as cleaned up by the internal TTL 
mechanism
   2) it reconstructs the expected state from the stored updates by taking the 
timestamp 
   of the latest update (ts) discarding elements with timestamp ts-ttl
   
   As you can see from the stacktrace in the error from Travis, the latest 
update has timestamp ts=1538918066021
   while there are elements in the list with timestamps greater than ts (e.g. 
1538918066136). This means that the internal
   clock on that machine went backwards, so Flink's TTL may have removed 
elements that appear in the expected state
   of the test, as it takes as current timestamp the ts=1538918066021.
   
   The fix is simply to assume that (for the test), processing time increases 
monotonically and ignore "updates from the past".


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> State TTL RocksDb backend end-to-end test failed on Travis
> --
>
> Key: FLINK-10531
> URL: https://issues.apache.org/jira/browse/FLINK-10531
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.6.1
>Reporter: Till Rohrmann
>Assignee: Kostas Kloudas
>Priority: Critical
>  Labels: pull-request-available, test-stability
>
> The {{State TTL RocksDb backend end-to-end test}} end-to-end test failed on 
> Travis.
> https://travis-ci.org/apache/flink/jobs/438226190
> https://api.travis-ci.org/v3/job/438226190/log.txt



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-10531) State TTL RocksDb backend end-to-end test failed on Travis

2018-11-06 Thread ASF GitHub Bot (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-10531:
---
Labels: pull-request-available test-stability  (was: test-stability)

> State TTL RocksDb backend end-to-end test failed on Travis
> --
>
> Key: FLINK-10531
> URL: https://issues.apache.org/jira/browse/FLINK-10531
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.6.1
>Reporter: Till Rohrmann
>Assignee: Kostas Kloudas
>Priority: Critical
>  Labels: pull-request-available, test-stability
>
> The {{State TTL RocksDb backend end-to-end test}} end-to-end test failed on 
> Travis.
> https://travis-ci.org/apache/flink/jobs/438226190
> https://api.travis-ci.org/v3/job/438226190/log.txt



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] kl0u opened a new pull request #7036: [FLINK-10531][e2e] Fix unstable TTL end-to-end test.

2018-11-06 Thread GitBox
kl0u opened a new pull request #7036: [FLINK-10531][e2e] Fix unstable TTL 
end-to-end test.
URL: https://github.com/apache/flink/pull/7036
 
 
   As describe on the JIRA, the problem is that the clock on the machine on 
Travis seems to have jumped backwards.
   
   This resulted in a mismatch between the elements discarded as expired by 
Flink's internal TTL mechanism, and the "user-code" in the test that computes 
the expired elements based on the timestamp of the latest update.
   
   I repeat the explanation from the JIRA here for reference:
   
   
---
   
   The way the test works is the following. For a given state:
   1) it keeps the state with TTL
   2) and keeps all the updates to that state irrespective of their timestamp 
(e.g. all the elements added in a ListState)
   
   To verify correctness:
   1) it fetches the state from Flink, as cleaned up by the internal TTL 
mechanism
   2) it reconstructs the expected state from the stored updates by taking the 
timestamp 
   of the latest update (ts) discarding elements with timestamp ts-ttl
   
   As you can see from the stacktrace in the error from Travis, the latest 
update has timestamp ts=1538918066021
   while there are elements in the list with timestamps greater than ts (e.g. 
1538918066136). This means that the internal
   clock on that machine went backwards, so Flink's TTL may have removed 
elements that appear in the expected state
   of the test, as it takes as current timestamp the ts=1538918066021.
   
   The fix is simply to assume that (for the test), processing time increases 
monotonically and ignore "updates from the past".


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10797) "IntelliJ Setup" link is broken in Readme.md

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10797?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16677010#comment-16677010
 ] 

ASF GitHub Bot commented on FLINK-10797:


xndai closed pull request #7025: [FLINK-10797] "IntelliJ Setup" link is broken 
in README.md
URL: https://github.com/apache/flink/pull/7025
 
 
   

This is a PR merged from a forked repository.
As GitHub hides the original diff on merge, it is displayed below for
the sake of provenance:

As this is a foreign pull request (from a fork), the diff is supplied
below (as it won't show otherwise due to GitHub magic):

diff --git a/README.md b/README.md
index 7966ecd3d76..b7a96cdee46 100644
--- a/README.md
+++ b/README.md
@@ -100,7 +100,7 @@ The IntelliJ IDE supports Maven out of the box and offers a 
plugin for Scala dev
 * IntelliJ download: 
[https://www.jetbrains.com/idea/](https://www.jetbrains.com/idea/)
 * IntelliJ Scala Plugin: 
[http://plugins.jetbrains.com/plugin/?id=1347](http://plugins.jetbrains.com/plugin/?id=1347)
 
-Check out our [Setting up 
IntelliJ](https://github.com/apache/flink/blob/master/docs/internals/ide_setup.md#intellij-idea)
 guide for details.
+Check out our [Setting up 
IntelliJ](https://github.com/apache/flink/blob/master/docs/flinkDev/ide_setup.md#intellij-idea)
 guide for details.
 
 ### Eclipse Scala IDE
 


 


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> "IntelliJ Setup" link is broken in Readme.md
> 
>
> Key: FLINK-10797
> URL: https://issues.apache.org/jira/browse/FLINK-10797
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.7.0
>Reporter: Xiening Dai
>Assignee: TisonKun
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>
> The link points to 
> https://github.com/apache/flink/blob/master/docs/internals/ide_setup.md#intellij-idea
>  which is a 404 not found.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] xndai closed pull request #7025: [FLINK-10797] "IntelliJ Setup" link is broken in README.md

2018-11-06 Thread GitBox
xndai closed pull request #7025: [FLINK-10797] "IntelliJ Setup" link is broken 
in README.md
URL: https://github.com/apache/flink/pull/7025
 
 
   

This is a PR merged from a forked repository.
As GitHub hides the original diff on merge, it is displayed below for
the sake of provenance:

As this is a foreign pull request (from a fork), the diff is supplied
below (as it won't show otherwise due to GitHub magic):

diff --git a/README.md b/README.md
index 7966ecd3d76..b7a96cdee46 100644
--- a/README.md
+++ b/README.md
@@ -100,7 +100,7 @@ The IntelliJ IDE supports Maven out of the box and offers a 
plugin for Scala dev
 * IntelliJ download: 
[https://www.jetbrains.com/idea/](https://www.jetbrains.com/idea/)
 * IntelliJ Scala Plugin: 
[http://plugins.jetbrains.com/plugin/?id=1347](http://plugins.jetbrains.com/plugin/?id=1347)
 
-Check out our [Setting up 
IntelliJ](https://github.com/apache/flink/blob/master/docs/internals/ide_setup.md#intellij-idea)
 guide for details.
+Check out our [Setting up 
IntelliJ](https://github.com/apache/flink/blob/master/docs/flinkDev/ide_setup.md#intellij-idea)
 guide for details.
 
 ### Eclipse Scala IDE
 


 


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10796) Add a default external catalog (as FlinkInMemoryCatalog) to CatalogManager

2018-11-06 Thread Bowen Li (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10796?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676991#comment-16676991
 ] 

Bowen Li commented on FLINK-10796:
--

[~xiaow...@gmail.com] That's right. I've added the requirement in description

> Add a default external catalog (as FlinkInMemoryCatalog) to CatalogManager
> --
>
> Key: FLINK-10796
> URL: https://issues.apache.org/jira/browse/FLINK-10796
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table API  SQL
>Reporter: Bowen Li
>Assignee: Bowen Li
>Priority: Major
> Fix For: 1.8.0
>
>
> CatalogManager has a default catalog (as FlinkInMemoryCatalog), which will be 
> selected when a catalog name isn’t given in a request.
> Users can also specify an external catalog as their preferred default one.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-10796) Add a default external catalog (as FlinkInMemoryCatalog) to CatalogManager

2018-11-06 Thread Bowen Li (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10796?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bowen Li updated FLINK-10796:
-
Description: 
CatalogManager has a default catalog (as FlinkInMemoryCatalog), which will be 
selected when a catalog name isn’t given in a request.

Users can also specify an external catalog as their preferred default one.

  was:
CatalogManager has a default catalog, which will be selected when a catalog 
name isn’t given in a request.

Users can also specify an external catalog as their preferred default one.


> Add a default external catalog (as FlinkInMemoryCatalog) to CatalogManager
> --
>
> Key: FLINK-10796
> URL: https://issues.apache.org/jira/browse/FLINK-10796
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table API  SQL
>Reporter: Bowen Li
>Assignee: Bowen Li
>Priority: Major
> Fix For: 1.8.0
>
>
> CatalogManager has a default catalog (as FlinkInMemoryCatalog), which will be 
> selected when a catalog name isn’t given in a request.
> Users can also specify an external catalog as their preferred default one.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10789) Some new serializer snapshots added after 1.6 are not implementing the new TypeSerializerSnapshot interface

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676988#comment-16676988
 ] 

ASF GitHub Bot commented on FLINK-10789:


StefanRRichter commented on a change in pull request #7028: [FLINK-10789] New 
serializer snapshots after 1.6 should implement TypeSerializerSnapshot
URL: https://github.com/apache/flink/pull/7028#discussion_r231202779
 
 

 ##
 File path: 
flink-scala/src/main/java/org/apache/flink/api/scala/typeutils/ScalaEitherSerializerSnapshot.java
 ##
 @@ -0,0 +1,99 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.api.scala.typeutils;
+
+import org.apache.flink.api.common.typeutils.CompositeSerializerSnapshot;
+import org.apache.flink.api.common.typeutils.TypeSerializer;
+import org.apache.flink.api.common.typeutils.TypeSerializerSchemaCompatibility;
+import org.apache.flink.api.common.typeutils.TypeSerializerSnapshot;
+import org.apache.flink.core.memory.DataInputView;
+import org.apache.flink.core.memory.DataOutputView;
+import org.apache.flink.util.Preconditions;
+
+import java.io.IOException;
+
+import scala.util.Either;
+
+import static org.apache.flink.util.Preconditions.checkState;
+
+/**
+ * Configuration snapshot for serializers of Scala's {@link Either} type,
+ * containing configuration snapshots of the Left and Right serializers.
+ */
+public class ScalaEitherSerializerSnapshot implements 
TypeSerializerSnapshot> {
+
+
 
 Review comment:
   2 newlines.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Some new serializer snapshots added after 1.6 are not implementing the new 
> TypeSerializerSnapshot interface
> ---
>
> Key: FLINK-10789
> URL: https://issues.apache.org/jira/browse/FLINK-10789
> Project: Flink
>  Issue Type: Improvement
>  Components: Type Serialization System
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Tzu-Li (Gordon) Tai
>Priority: Critical
>  Labels: pull-request-available
>
> In commit 
> [3787b89105c6a005ddca999b0b8a0356e812ede6|[https://github.com/apache/flink/commit/3787b89105c6a005ddca999b0b8a0356e812ede6],]
>  there were a few serializer snapshot implementations that were newly added, 
> namely:
> - {{LockableSerializerConfigSnapshot}}
> - {{ListViewSerializerConfigSnapshot}}
> - {{MapViewSerializerConfigSnapshot}}
> - {{ScalaEitherSerializerConfigSnapshot}}
> The problem is that instead of implementing the newer 
> {{TypeSerializerSnapshot}}, they are extending the legacy deprecated 
> {{TypeSerializerConfigSnapshot}}. This doesn't break things, but we will be 
> redundantly writing Java-serialized with the new snapshots, as well as 
> causing us to must introduce migration paths code later on after releasing 
> them.
> It would be more ideal to implement them properly as 
> {{TypeSerializerSnapshot}} for the release.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] StefanRRichter commented on a change in pull request #7028: [FLINK-10789] New serializer snapshots after 1.6 should implement TypeSerializerSnapshot

2018-11-06 Thread GitBox
StefanRRichter commented on a change in pull request #7028: [FLINK-10789] New 
serializer snapshots after 1.6 should implement TypeSerializerSnapshot
URL: https://github.com/apache/flink/pull/7028#discussion_r231202779
 
 

 ##
 File path: 
flink-scala/src/main/java/org/apache/flink/api/scala/typeutils/ScalaEitherSerializerSnapshot.java
 ##
 @@ -0,0 +1,99 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.api.scala.typeutils;
+
+import org.apache.flink.api.common.typeutils.CompositeSerializerSnapshot;
+import org.apache.flink.api.common.typeutils.TypeSerializer;
+import org.apache.flink.api.common.typeutils.TypeSerializerSchemaCompatibility;
+import org.apache.flink.api.common.typeutils.TypeSerializerSnapshot;
+import org.apache.flink.core.memory.DataInputView;
+import org.apache.flink.core.memory.DataOutputView;
+import org.apache.flink.util.Preconditions;
+
+import java.io.IOException;
+
+import scala.util.Either;
+
+import static org.apache.flink.util.Preconditions.checkState;
+
+/**
+ * Configuration snapshot for serializers of Scala's {@link Either} type,
+ * containing configuration snapshots of the Left and Right serializers.
+ */
+public class ScalaEitherSerializerSnapshot implements 
TypeSerializerSnapshot> {
+
+
 
 Review comment:
   2 newlines.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-10796) Add a default external catalog (as FlinkInMemoryCatalog) to CatalogManager

2018-11-06 Thread Bowen Li (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10796?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bowen Li updated FLINK-10796:
-
Description: 
CatalogManager has a default catalog, which will be selected when a catalog 
name isn’t given in a request.

Users can also specify an external catalog as their preferred default one.

  was:CatalogManager has a default catalog, which will be selected when a 
catalog name isn’t given in a request.


> Add a default external catalog (as FlinkInMemoryCatalog) to CatalogManager
> --
>
> Key: FLINK-10796
> URL: https://issues.apache.org/jira/browse/FLINK-10796
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table API  SQL
>Reporter: Bowen Li
>Assignee: Bowen Li
>Priority: Major
> Fix For: 1.8.0
>
>
> CatalogManager has a default catalog, which will be selected when a catalog 
> name isn’t given in a request.
> Users can also specify an external catalog as their preferred default one.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] asfgit closed pull request #5863: [FLINK-8985][e2etest] initial support for End-to-end CLI test, excluding YARN test

2018-11-06 Thread GitBox
asfgit closed pull request #5863: [FLINK-8985][e2etest] initial support for 
End-to-end CLI test, excluding YARN test
URL: https://github.com/apache/flink/pull/5863
 
 
   

This is a PR merged from a forked repository.
As GitHub hides the original diff on merge, it is displayed below for
the sake of provenance:

As this is a foreign pull request (from a fork), the diff is supplied
below (as it won't show otherwise due to GitHub magic):

diff --git a/flink-end-to-end-tests/flink-api-test/pom.xml 
b/flink-end-to-end-tests/flink-api-test/pom.xml
new file mode 100644
index 000..24a85e02bef
--- /dev/null
+++ b/flink-end-to-end-tests/flink-api-test/pom.xml
@@ -0,0 +1,105 @@
+
+
+http://maven.apache.org/POM/4.0.0;
+xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
+xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/xsd/maven-4.0.0.xsd;>
+
+   
+   flink-end-to-end-tests
+   org.apache.flink
+   1.6-SNAPSHOT
+   ..
+   
+
+   4.0.0
+
+   flink-api-test
+   flink-api-test
+   jar
+
+   
+   
+   org.apache.flink
+   flink-core
+   ${project.version}
+   
+   
+   org.apache.flink
+   
flink-streaming-java_${scala.binary.version}
+   ${project.version}
+   
+   
+   org.apache.flink
+   
flink-connector-filesystem_${scala.binary.version}
+   ${project.version}
+   
+   
+   org.apache.flink
+   
flink-runtime-web_${scala.binary.version}
+   ${project.version}
+   test-jar
+   
+   
+
+   
+   
+   
+   org.apache.maven.plugins
+   maven-shade-plugin
+   3.0.0
+   
+   
+   package
+   
+   shade
+   
+   
+   
PeriodicStreamingJob
+   
+   
+   
com.google.code.findbugs:jsr305
+   
org.slf4j:*
+   
log4j:*
+   
+   
+   
+   
+   
*:*
+   

+   
META-INF/*.SF
+   
META-INF/*.DSA
+   
META-INF/*.RSA
+   

+   
+   
+   
+   
+   
org.apache.flink.runtime.tests.PeriodicStreamingJob
+   
+   
+   
+   
+   
+   
+   
+   
+
+
+
diff --git 
a/flink-end-to-end-tests/flink-api-test/src/main/java/org/apache/flink/runtime/tests/PeriodicStreamingJob.java
 
b/flink-end-to-end-tests/flink-api-test/src/main/java/org/apache/flink/runtime/tests/PeriodicStreamingJob.java
new file mode 100644
index 000..ee8a49b8583
--- /dev/null
+++ 
b/flink-end-to-end-tests/flink-api-test/src/main/java/org/apache/flink/runtime/tests/PeriodicStreamingJob.java
@@ -0,0 +1,127 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright 

[jira] [Commented] (FLINK-8985) End-to-end test: CLI

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-8985?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676976#comment-16676976
 ] 

ASF GitHub Bot commented on FLINK-8985:
---

asfgit closed pull request #5863: [FLINK-8985][e2etest] initial support for 
End-to-end CLI test, excluding YARN test
URL: https://github.com/apache/flink/pull/5863
 
 
   

This is a PR merged from a forked repository.
As GitHub hides the original diff on merge, it is displayed below for
the sake of provenance:

As this is a foreign pull request (from a fork), the diff is supplied
below (as it won't show otherwise due to GitHub magic):

diff --git a/flink-end-to-end-tests/flink-api-test/pom.xml 
b/flink-end-to-end-tests/flink-api-test/pom.xml
new file mode 100644
index 000..24a85e02bef
--- /dev/null
+++ b/flink-end-to-end-tests/flink-api-test/pom.xml
@@ -0,0 +1,105 @@
+
+
+http://maven.apache.org/POM/4.0.0;
+xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
+xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/xsd/maven-4.0.0.xsd;>
+
+   
+   flink-end-to-end-tests
+   org.apache.flink
+   1.6-SNAPSHOT
+   ..
+   
+
+   4.0.0
+
+   flink-api-test
+   flink-api-test
+   jar
+
+   
+   
+   org.apache.flink
+   flink-core
+   ${project.version}
+   
+   
+   org.apache.flink
+   
flink-streaming-java_${scala.binary.version}
+   ${project.version}
+   
+   
+   org.apache.flink
+   
flink-connector-filesystem_${scala.binary.version}
+   ${project.version}
+   
+   
+   org.apache.flink
+   
flink-runtime-web_${scala.binary.version}
+   ${project.version}
+   test-jar
+   
+   
+
+   
+   
+   
+   org.apache.maven.plugins
+   maven-shade-plugin
+   3.0.0
+   
+   
+   package
+   
+   shade
+   
+   
+   
PeriodicStreamingJob
+   
+   
+   
com.google.code.findbugs:jsr305
+   
org.slf4j:*
+   
log4j:*
+   
+   
+   
+   
+   
*:*
+   

+   
META-INF/*.SF
+   
META-INF/*.DSA
+   
META-INF/*.RSA
+   

+   
+   
+   
+   
+   
org.apache.flink.runtime.tests.PeriodicStreamingJob
+   
+   
+   
+   
+   
+   
+   
+   
+
+
+
diff --git 
a/flink-end-to-end-tests/flink-api-test/src/main/java/org/apache/flink/runtime/tests/PeriodicStreamingJob.java
 
b/flink-end-to-end-tests/flink-api-test/src/main/java/org/apache/flink/runtime/tests/PeriodicStreamingJob.java
new file mode 100644
index 000..ee8a49b8583
--- /dev/null
+++ 
b/flink-end-to-end-tests/flink-api-test/src/main/java/org/apache/flink/runtime/tests/PeriodicStreamingJob.java

[jira] [Commented] (FLINK-10463) Null literal cannot be properly parsed in Java Table API function call

2018-11-06 Thread Xingcan Cui (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10463?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676968#comment-16676968
 ] 

Xingcan Cui commented on FLINK-10463:
-

[~twalthr], sure. I'll cherry-pick the commit to versions 1.6.3 and 1.7.1.

> Null literal cannot be properly parsed in Java Table API function call
> --
>
> Key: FLINK-10463
> URL: https://issues.apache.org/jira/browse/FLINK-10463
> Project: Flink
>  Issue Type: Bug
>  Components: Table API  SQL
>Reporter: Xingcan Cui
>Assignee: Hequn Cheng
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.6.3, 1.8.0, 1.7.1
>
>
> For example, the expression `Null(STRING).regexpReplace('oo|ar', '')` throws 
> the following exception.
> {code:java}
> org.apache.flink.table.api.ExpressionParserException: Could not parse 
> expression at column 13: string matching regex 
> `(?i)\Qas\E(?![_$\p{javaJavaIdentifierPart}])' expected but `.' found
> Null(STRING).regexpReplace('oo|ar', '')
> ^
>   at 
> org.apache.flink.table.expressions.ExpressionParser$.throwError(ExpressionParser.scala:576)
>   at 
> org.apache.flink.table.expressions.ExpressionParser$.parseExpression(ExpressionParser.scala:569)
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-10463) Null literal cannot be properly parsed in Java Table API function call

2018-11-06 Thread Xingcan Cui (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xingcan Cui updated FLINK-10463:

Fix Version/s: 1.7.1
   1.6.3

> Null literal cannot be properly parsed in Java Table API function call
> --
>
> Key: FLINK-10463
> URL: https://issues.apache.org/jira/browse/FLINK-10463
> Project: Flink
>  Issue Type: Bug
>  Components: Table API  SQL
>Reporter: Xingcan Cui
>Assignee: Hequn Cheng
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.6.3, 1.8.0, 1.7.1
>
>
> For example, the expression `Null(STRING).regexpReplace('oo|ar', '')` throws 
> the following exception.
> {code:java}
> org.apache.flink.table.api.ExpressionParserException: Could not parse 
> expression at column 13: string matching regex 
> `(?i)\Qas\E(?![_$\p{javaJavaIdentifierPart}])' expected but `.' found
> Null(STRING).regexpReplace('oo|ar', '')
> ^
>   at 
> org.apache.flink.table.expressions.ExpressionParser$.throwError(ExpressionParser.scala:576)
>   at 
> org.apache.flink.table.expressions.ExpressionParser$.parseExpression(ExpressionParser.scala:569)
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (FLINK-10720) Add stress deployment end-to-end test

2018-11-06 Thread Stefan Richter (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10720?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stefan Richter closed FLINK-10720.
--
Resolution: Implemented

Merged in:
master: 660bb687c1
release-1.7: 380df0559b

> Add stress deployment end-to-end test
> -
>
> Key: FLINK-10720
> URL: https://issues.apache.org/jira/browse/FLINK-10720
> Project: Flink
>  Issue Type: Sub-task
>  Components: E2E Tests
>Affects Versions: 1.7.0
>Reporter: Till Rohrmann
>Assignee: Stefan Richter
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>
> In order to test Flink's scalability, I suggest to add an end-to-end test 
> which tests the deployment of a job which is very demanding. The job should 
> have large {{TaskDeploymentDescriptors}} (e.g. a job using union state or 
> having a high degree of parallelism). That way we can test that the 
> serialization overhead of the TDDs does not affect the health of the cluster 
> (e.g. heartbeats are not affected because the serialization does not happen 
> in the main thread).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10531) State TTL RocksDb backend end-to-end test failed on Travis

2018-11-06 Thread Kostas Kloudas (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676957#comment-16676957
 ] 

Kostas Kloudas commented on FLINK-10531:


It is a test instability. 

The way the test works is the following. For a given state:
1) it keeps the state with TTL
2) and keeps all the updates to that state irrespective of their timestamp 
(e.g. all the elements added in a ListState)

To verify correctness:
1) it fetches the state from Flink, as cleaned up by the internal TTL mechanism
2) it reconstructs the expected state from the stored updates by taking the 
timestamp 
   of the latest update (ts) discarding elements with timestamp ts-ttl

As you can see from the stacktrace in the error from Travis, the latest update 
has timestamp ts=1538918066021
while there are elements in the list with timestamps greater than ts (e.g. 
1538918066136). This means that the internal
clock on that machine went backwards, so Flink's TTL may have removed elements 
that appear in the expected state
of the test, as it takes as current timestamp the ts=1538918066021.

The fix is simply to assume that (for the test), processing time increases 
monotonically and ignore "updates from the past".

> State TTL RocksDb backend end-to-end test failed on Travis
> --
>
> Key: FLINK-10531
> URL: https://issues.apache.org/jira/browse/FLINK-10531
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.6.1
>Reporter: Till Rohrmann
>Assignee: Kostas Kloudas
>Priority: Critical
>  Labels: test-stability
>
> The {{State TTL RocksDb backend end-to-end test}} end-to-end test failed on 
> Travis.
> https://travis-ci.org/apache/flink/jobs/438226190
> https://api.travis-ci.org/v3/job/438226190/log.txt



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10463) Null literal cannot be properly parsed in Java Table API function call

2018-11-06 Thread Timo Walther (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10463?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676955#comment-16676955
 ] 

Timo Walther commented on FLINK-10463:
--

[~xccui] can you also cherry-pick bug fixes to the last two release branches if 
backporting is not a big effort? thanks.

> Null literal cannot be properly parsed in Java Table API function call
> --
>
> Key: FLINK-10463
> URL: https://issues.apache.org/jira/browse/FLINK-10463
> Project: Flink
>  Issue Type: Bug
>  Components: Table API  SQL
>Reporter: Xingcan Cui
>Assignee: Hequn Cheng
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.8.0
>
>
> For example, the expression `Null(STRING).regexpReplace('oo|ar', '')` throws 
> the following exception.
> {code:java}
> org.apache.flink.table.api.ExpressionParserException: Could not parse 
> expression at column 13: string matching regex 
> `(?i)\Qas\E(?![_$\p{javaJavaIdentifierPart}])' expected but `.' found
> Null(STRING).regexpReplace('oo|ar', '')
> ^
>   at 
> org.apache.flink.table.expressions.ExpressionParser$.throwError(ExpressionParser.scala:576)
>   at 
> org.apache.flink.table.expressions.ExpressionParser$.parseExpression(ExpressionParser.scala:569)
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (FLINK-10531) State TTL RocksDb backend end-to-end test failed on Travis

2018-11-06 Thread Kostas Kloudas (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas reassigned FLINK-10531:
--

Assignee: Kostas Kloudas  (was: Renjie Liu)

> State TTL RocksDb backend end-to-end test failed on Travis
> --
>
> Key: FLINK-10531
> URL: https://issues.apache.org/jira/browse/FLINK-10531
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.6.1
>Reporter: Till Rohrmann
>Assignee: Kostas Kloudas
>Priority: Critical
>  Labels: test-stability
>
> The {{State TTL RocksDb backend end-to-end test}} end-to-end test failed on 
> Travis.
> https://travis-ci.org/apache/flink/jobs/438226190
> https://api.travis-ci.org/v3/job/438226190/log.txt



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10720) Add stress deployment end-to-end test

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10720?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676948#comment-16676948
 ] 

ASF GitHub Bot commented on FLINK-10720:


asfgit closed pull request #6994: [FLINK-10720][tests] Add deployment 
end-to-end stress test with many …
URL: https://github.com/apache/flink/pull/6994
 
 
   

This is a PR merged from a forked repository.
As GitHub hides the original diff on merge, it is displayed below for
the sake of provenance:

As this is a foreign pull request (from a fork), the diff is supplied
below (as it won't show otherwise due to GitHub magic):

diff --git 
a/flink-end-to-end-tests/flink-datastream-allround-test/src/main/java/org/apache/flink/streaming/tests/DataStreamAllroundTestJobFactory.java
 
b/flink-end-to-end-tests/flink-datastream-allround-test/src/main/java/org/apache/flink/streaming/tests/DataStreamAllroundTestJobFactory.java
index 3c8d0ad537f..8bd649cc5eb 100644
--- 
a/flink-end-to-end-tests/flink-datastream-allround-test/src/main/java/org/apache/flink/streaming/tests/DataStreamAllroundTestJobFactory.java
+++ 
b/flink-end-to-end-tests/flink-datastream-allround-test/src/main/java/org/apache/flink/streaming/tests/DataStreamAllroundTestJobFactory.java
@@ -88,7 +88,7 @@
  * Total duration is (sliding_window_operator.num_events) * 
(sequence_generator_source.event_time.clock_progress).
  * 
  */
-class DataStreamAllroundTestJobFactory {
+public class DataStreamAllroundTestJobFactory {
private static final ConfigOption TEST_SEMANTICS = ConfigOptions
.key("test.semantics")
.defaultValue("exactly-once")
@@ -201,7 +201,7 @@
.key("tumbling_window_operator.num_events")
.defaultValue(20L);
 
-   static void setupEnvironment(StreamExecutionEnvironment env, 
ParameterTool pt) throws Exception {
+   public static void setupEnvironment(StreamExecutionEnvironment env, 
ParameterTool pt) throws Exception {
 
// set checkpointing semantics
String semantics = pt.get(TEST_SEMANTICS.key(), 
TEST_SEMANTICS.defaultValue());
diff --git a/flink-end-to-end-tests/flink-heavy-deployment-stress-test/pom.xml 
b/flink-end-to-end-tests/flink-heavy-deployment-stress-test/pom.xml
new file mode 100644
index 000..fa9444dec1b
--- /dev/null
+++ b/flink-end-to-end-tests/flink-heavy-deployment-stress-test/pom.xml
@@ -0,0 +1,81 @@
+
+
+http://maven.apache.org/POM/4.0.0; 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
+xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/maven-v4_0_0.xsd;>
+
+   4.0.0
+   
+   
+   org.apache.flink
+   flink-end-to-end-tests
+   1.8-SNAPSHOT
+   ..
+   
+
+   flink-heavy-deployment-stress-test
+   flink-heavy-deployment-stress-test
+   jar
+
+   
+   
+   org.apache.flink
+   flink-streaming-java_2.11
+   ${project.version}
+   
+   
+   org.apache.flink
+   flink-datastream-allround-test
+   ${project.version}
+   
+   
+   org.apache.flink
+   flink-test-utils_2.11
+   ${project.version}
+   compile
+   
+   
+
+   
+   
+   
+   org.apache.maven.plugins
+   maven-shade-plugin
+   
+   
+   
HeavyDeploymentStressTestProgram
+   package
+   
+   shade
+   
+   
+   
HeavyDeploymentStressTestProgram
+   
+   
+   
org.apache.flink.deployment.HeavyDeploymentStressTestProgram
+   
+   
+   
+   
+   
+   
+   
+   
+
+
diff --git 
a/flink-end-to-end-tests/flink-heavy-deployment-stress-test/src/main/java/org/apache/flink/deployment/HeavyDeploymentStressTestProgram.java
 

[GitHub] asfgit closed pull request #6994: [FLINK-10720][tests] Add deployment end-to-end stress test with many …

2018-11-06 Thread GitBox
asfgit closed pull request #6994: [FLINK-10720][tests] Add deployment 
end-to-end stress test with many …
URL: https://github.com/apache/flink/pull/6994
 
 
   

This is a PR merged from a forked repository.
As GitHub hides the original diff on merge, it is displayed below for
the sake of provenance:

As this is a foreign pull request (from a fork), the diff is supplied
below (as it won't show otherwise due to GitHub magic):

diff --git 
a/flink-end-to-end-tests/flink-datastream-allround-test/src/main/java/org/apache/flink/streaming/tests/DataStreamAllroundTestJobFactory.java
 
b/flink-end-to-end-tests/flink-datastream-allround-test/src/main/java/org/apache/flink/streaming/tests/DataStreamAllroundTestJobFactory.java
index 3c8d0ad537f..8bd649cc5eb 100644
--- 
a/flink-end-to-end-tests/flink-datastream-allround-test/src/main/java/org/apache/flink/streaming/tests/DataStreamAllroundTestJobFactory.java
+++ 
b/flink-end-to-end-tests/flink-datastream-allround-test/src/main/java/org/apache/flink/streaming/tests/DataStreamAllroundTestJobFactory.java
@@ -88,7 +88,7 @@
  * Total duration is (sliding_window_operator.num_events) * 
(sequence_generator_source.event_time.clock_progress).
  * 
  */
-class DataStreamAllroundTestJobFactory {
+public class DataStreamAllroundTestJobFactory {
private static final ConfigOption TEST_SEMANTICS = ConfigOptions
.key("test.semantics")
.defaultValue("exactly-once")
@@ -201,7 +201,7 @@
.key("tumbling_window_operator.num_events")
.defaultValue(20L);
 
-   static void setupEnvironment(StreamExecutionEnvironment env, 
ParameterTool pt) throws Exception {
+   public static void setupEnvironment(StreamExecutionEnvironment env, 
ParameterTool pt) throws Exception {
 
// set checkpointing semantics
String semantics = pt.get(TEST_SEMANTICS.key(), 
TEST_SEMANTICS.defaultValue());
diff --git a/flink-end-to-end-tests/flink-heavy-deployment-stress-test/pom.xml 
b/flink-end-to-end-tests/flink-heavy-deployment-stress-test/pom.xml
new file mode 100644
index 000..fa9444dec1b
--- /dev/null
+++ b/flink-end-to-end-tests/flink-heavy-deployment-stress-test/pom.xml
@@ -0,0 +1,81 @@
+
+
+http://maven.apache.org/POM/4.0.0; 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
+xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/maven-v4_0_0.xsd;>
+
+   4.0.0
+   
+   
+   org.apache.flink
+   flink-end-to-end-tests
+   1.8-SNAPSHOT
+   ..
+   
+
+   flink-heavy-deployment-stress-test
+   flink-heavy-deployment-stress-test
+   jar
+
+   
+   
+   org.apache.flink
+   flink-streaming-java_2.11
+   ${project.version}
+   
+   
+   org.apache.flink
+   flink-datastream-allround-test
+   ${project.version}
+   
+   
+   org.apache.flink
+   flink-test-utils_2.11
+   ${project.version}
+   compile
+   
+   
+
+   
+   
+   
+   org.apache.maven.plugins
+   maven-shade-plugin
+   
+   
+   
HeavyDeploymentStressTestProgram
+   package
+   
+   shade
+   
+   
+   
HeavyDeploymentStressTestProgram
+   
+   
+   
org.apache.flink.deployment.HeavyDeploymentStressTestProgram
+   
+   
+   
+   
+   
+   
+   
+   
+
+
diff --git 
a/flink-end-to-end-tests/flink-heavy-deployment-stress-test/src/main/java/org/apache/flink/deployment/HeavyDeploymentStressTestProgram.java
 
b/flink-end-to-end-tests/flink-heavy-deployment-stress-test/src/main/java/org/apache/flink/deployment/HeavyDeploymentStressTestProgram.java
new file mode 100644
index 000..d65583f98af
--- /dev/null
+++ 

[jira] [Commented] (FLINK-10720) Add stress deployment end-to-end test

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10720?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676943#comment-16676943
 ] 

ASF GitHub Bot commented on FLINK-10720:


StefanRRichter commented on issue #6994: [FLINK-10720][tests] Add deployment 
end-to-end stress test with many …
URL: https://github.com/apache/flink/pull/6994#issuecomment-436309544
 
 
   Thanks @tillrohrmann! Merging.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Add stress deployment end-to-end test
> -
>
> Key: FLINK-10720
> URL: https://issues.apache.org/jira/browse/FLINK-10720
> Project: Flink
>  Issue Type: Sub-task
>  Components: E2E Tests
>Affects Versions: 1.7.0
>Reporter: Till Rohrmann
>Assignee: Stefan Richter
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>
> In order to test Flink's scalability, I suggest to add an end-to-end test 
> which tests the deployment of a job which is very demanding. The job should 
> have large {{TaskDeploymentDescriptors}} (e.g. a job using union state or 
> having a high degree of parallelism). That way we can test that the 
> serialization overhead of the TDDs does not affect the health of the cluster 
> (e.g. heartbeats are not affected because the serialization does not happen 
> in the main thread).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] StefanRRichter commented on issue #6994: [FLINK-10720][tests] Add deployment end-to-end stress test with many …

2018-11-06 Thread GitBox
StefanRRichter commented on issue #6994: [FLINK-10720][tests] Add deployment 
end-to-end stress test with many …
URL: https://github.com/apache/flink/pull/6994#issuecomment-436309544
 
 
   Thanks @tillrohrmann! Merging.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-10805) test_confluent_schema_registry.sh and test_sql_client.sh end-to-end test broken

2018-11-06 Thread ASF GitHub Bot (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10805?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-10805:
---
Labels: pull-request-available test-stability  (was: test-stability)

> test_confluent_schema_registry.sh and test_sql_client.sh end-to-end test 
> broken
> ---
>
> Key: FLINK-10805
> URL: https://issues.apache.org/jira/browse/FLINK-10805
> Project: Flink
>  Issue Type: Bug
>  Components: E2E Tests
>Affects Versions: 1.7.0
>Reporter: Till Rohrmann
>Assignee: Till Rohrmann
>Priority: Major
>  Labels: pull-request-available, test-stability
>
> The end-to-end test {{test_confluent_schema_registry.sh}} and 
> {{test_sql_client.sh}} seem to be broken after fixing FLINK-10600. 
> I suspect that the kafka-common.sh import has not been adapted.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] tillrohrmann opened a new pull request #7035: [BP-1.7][FLINK-10805] Fix failing end-to-end tests

2018-11-06 Thread GitBox
tillrohrmann opened a new pull request #7035: [BP-1.7][FLINK-10805] Fix failing 
end-to-end tests
URL: https://github.com/apache/flink/pull/7035
 
 
   Backport of #7034 for `release-1.7`.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] tillrohrmann opened a new pull request #7034: [FLINK-10805] Fix failing end-to-end tests

2018-11-06 Thread GitBox
tillrohrmann opened a new pull request #7034: [FLINK-10805] Fix failing 
end-to-end tests
URL: https://github.com/apache/flink/pull/7034
 
 
   ## What is the purpose of the change
   
   Fix test_confluent_schema_registry.sh and test_sql_client.sh end-to-end 
tests which failed
   because of missing arguments for kafka-common.sh and the newly introduce set 
-e flag.
   
   ## Verifying this change
   
   - Manually tested
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10805) test_confluent_schema_registry.sh and test_sql_client.sh end-to-end test broken

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10805?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676937#comment-16676937
 ] 

ASF GitHub Bot commented on FLINK-10805:


tillrohrmann opened a new pull request #7034: [FLINK-10805] Fix failing 
end-to-end tests
URL: https://github.com/apache/flink/pull/7034
 
 
   ## What is the purpose of the change
   
   Fix test_confluent_schema_registry.sh and test_sql_client.sh end-to-end 
tests which failed
   because of missing arguments for kafka-common.sh and the newly introduce set 
-e flag.
   
   ## Verifying this change
   
   - Manually tested
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> test_confluent_schema_registry.sh and test_sql_client.sh end-to-end test 
> broken
> ---
>
> Key: FLINK-10805
> URL: https://issues.apache.org/jira/browse/FLINK-10805
> Project: Flink
>  Issue Type: Bug
>  Components: E2E Tests
>Affects Versions: 1.7.0
>Reporter: Till Rohrmann
>Assignee: Till Rohrmann
>Priority: Major
>  Labels: pull-request-available, test-stability
>
> The end-to-end test {{test_confluent_schema_registry.sh}} and 
> {{test_sql_client.sh}} seem to be broken after fixing FLINK-10600. 
> I suspect that the kafka-common.sh import has not been adapted.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10771) Replace hard code of job graph file path with config option for FileJobGraphRetriever

2018-11-06 Thread vinoyang (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676916#comment-16676916
 ] 

vinoyang commented on FLINK-10771:
--

Yes, In AbstractYarnClusterDescriptor, we can explicitly set the value 
"job.graph" for "internal.jobgraph-path" and then uniformly retrieve it from 
the configuration object.

> Replace hard code of job graph file path with config option for 
> FileJobGraphRetriever
> -
>
> Key: FLINK-10771
> URL: https://issues.apache.org/jira/browse/FLINK-10771
> Project: Flink
>  Issue Type: Improvement
>  Components: YARN
>Reporter: vinoyang
>Assignee: vinoyang
>Priority: Major
>
> There is a config option to configure the value :
> {code:java}
> public static final ConfigOption JOB_GRAPH_FILE_PATH = ConfigOptions
>.key("internal.jobgraph-path")
>.defaultValue("job.graph");
> {code}
> However, we used the default hard code in AbstractYarnClusterDescriptor. 
> This is just a preliminary refactoring, and I finally recommend that we use 
> Zookeeper-based storage. It has been implemented in FLINK-10292.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-8985) End-to-end test: CLI

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-8985?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676906#comment-16676906
 ] 

ASF GitHub Bot commented on FLINK-8985:
---

twalthr commented on issue #5863: [FLINK-8985][e2etest] initial support for 
End-to-end CLI test, excluding YARN test
URL: https://github.com/apache/flink/pull/5863#issuecomment-436296096
 
 
   Thanks for the PR @walterddr. I think the test looks good to merge. I will 
add some additional safety checks and rebase this code to the current master.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> End-to-end test: CLI
> 
>
> Key: FLINK-8985
> URL: https://issues.apache.org/jira/browse/FLINK-8985
> Project: Flink
>  Issue Type: Sub-task
>  Components: Client, Tests
>Affects Versions: 1.5.0
>Reporter: Till Rohrmann
>Assignee: Rong Rong
>Priority: Major
>  Labels: pull-request-available
>
> We should add end-to-end test which verifies that all client commands are 
> working correctly.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-8985) End-to-end test: CLI

2018-11-06 Thread ASF GitHub Bot (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-8985?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-8985:
--
Labels: pull-request-available  (was: )

> End-to-end test: CLI
> 
>
> Key: FLINK-8985
> URL: https://issues.apache.org/jira/browse/FLINK-8985
> Project: Flink
>  Issue Type: Sub-task
>  Components: Client, Tests
>Affects Versions: 1.5.0
>Reporter: Till Rohrmann
>Assignee: Rong Rong
>Priority: Major
>  Labels: pull-request-available
>
> We should add end-to-end test which verifies that all client commands are 
> working correctly.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] twalthr commented on issue #5863: [FLINK-8985][e2etest] initial support for End-to-end CLI test, excluding YARN test

2018-11-06 Thread GitBox
twalthr commented on issue #5863: [FLINK-8985][e2etest] initial support for 
End-to-end CLI test, excluding YARN test
URL: https://github.com/apache/flink/pull/5863#issuecomment-436296096
 
 
   Thanks for the PR @walterddr. I think the test looks good to merge. I will 
add some additional safety checks and rebase this code to the current master.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Resolved] (FLINK-10463) Null literal cannot be properly parsed in Java Table API function call

2018-11-06 Thread Xingcan Cui (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xingcan Cui resolved FLINK-10463.
-
Resolution: Fixed

Fixed in 1.8.0 5a4e5e9dcaa83368862d27bd80429235c02ed66e

> Null literal cannot be properly parsed in Java Table API function call
> --
>
> Key: FLINK-10463
> URL: https://issues.apache.org/jira/browse/FLINK-10463
> Project: Flink
>  Issue Type: Bug
>  Components: Table API  SQL
>Reporter: Xingcan Cui
>Assignee: Hequn Cheng
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.8.0
>
>
> For example, the expression `Null(STRING).regexpReplace('oo|ar', '')` throws 
> the following exception.
> {code:java}
> org.apache.flink.table.api.ExpressionParserException: Could not parse 
> expression at column 13: string matching regex 
> `(?i)\Qas\E(?![_$\p{javaJavaIdentifierPart}])' expected but `.' found
> Null(STRING).regexpReplace('oo|ar', '')
> ^
>   at 
> org.apache.flink.table.expressions.ExpressionParser$.throwError(ExpressionParser.scala:576)
>   at 
> org.apache.flink.table.expressions.ExpressionParser$.parseExpression(ExpressionParser.scala:569)
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-10463) Null literal cannot be properly parsed in Java Table API function call

2018-11-06 Thread Xingcan Cui (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xingcan Cui updated FLINK-10463:

Fix Version/s: 1.8.0

> Null literal cannot be properly parsed in Java Table API function call
> --
>
> Key: FLINK-10463
> URL: https://issues.apache.org/jira/browse/FLINK-10463
> Project: Flink
>  Issue Type: Bug
>  Components: Table API  SQL
>Reporter: Xingcan Cui
>Assignee: Hequn Cheng
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.8.0
>
>
> For example, the expression `Null(STRING).regexpReplace('oo|ar', '')` throws 
> the following exception.
> {code:java}
> org.apache.flink.table.api.ExpressionParserException: Could not parse 
> expression at column 13: string matching regex 
> `(?i)\Qas\E(?![_$\p{javaJavaIdentifierPart}])' expected but `.' found
> Null(STRING).regexpReplace('oo|ar', '')
> ^
>   at 
> org.apache.flink.table.expressions.ExpressionParser$.throwError(ExpressionParser.scala:576)
>   at 
> org.apache.flink.table.expressions.ExpressionParser$.parseExpression(ExpressionParser.scala:569)
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10463) Null literal cannot be properly parsed in Java Table API function call

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10463?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676893#comment-16676893
 ] 

ASF GitHub Bot commented on FLINK-10463:


asfgit closed pull request #6888: [FLINK-10463][table] Null literal cannot be 
properly parsed in Java Table API function call
URL: https://github.com/apache/flink/pull/6888
 
 
   

This is a PR merged from a forked repository.
As GitHub hides the original diff on merge, it is displayed below for
the sake of provenance:

As this is a foreign pull request (from a fork), the diff is supplied
below (as it won't show otherwise due to GitHub magic):

diff --git a/docs/dev/table/functions.md b/docs/dev/table/functions.md
index e47b31772b3..e652c304e93 100644
--- a/docs/dev/table/functions.md
+++ b/docs/dev/table/functions.md
@@ -4186,7 +4186,7 @@ CAST(value AS type)
   
   
 Returns a new value being cast to type type. See the 
supported types here.
-E.g., CAST('42' AS INT) returns 42.
+E.g., CAST('42' AS INT) returns 42; CAST(NULL AS 
VARCHAR) returns NULL of type VARCHAR.
   
 
   
@@ -4211,7 +4211,7 @@ ANY.cast(TYPE)
   
   
 Returns a new ANY being cast to type TYPE. See the 
supported types here.
-E.g., '42'.cast(INT) returns 42.
+E.g., '42'.cast(INT) returns 42; 
Null(STRING) returns NULL of type STRING.
   
 
 
@@ -4236,7 +4236,7 @@ ANY.cast(TYPE)
   
   
 Returns a new ANY being cast to type TYPE. See the 
supported types here.
-E.g., "42".cast(Types.INT) returns 42.
+E.g., "42".cast(Types.INT) returns 42; 
Null(Types.STRING) returns NULL of type STRING.
   
 
   
diff --git a/docs/dev/table/tableApi.md b/docs/dev/table/tableApi.md
index 6c4e1be0cf7..3f599a8f5a2 100644
--- a/docs/dev/table/tableApi.md
+++ b/docs/dev/table/tableApi.md
@@ -1738,11 +1738,13 @@ term = product , [ ( "+" | "-" ) , product ] ;
 
 product = unary , [ ( "*" | "/" | "%") , unary ] ;
 
-unary = [ "!" | "-" ] , composite ;
+unary = [ "!" | "-" | "+" ] , composite ;
 
-composite = over | nullLiteral | suffixed | atom ;
+composite = over | suffixed | nullLiteral | prefixed | atom ;
 
-suffixed = interval | cast | as | if | functionCall ;
+suffixed = interval | suffixAs | suffixCast | suffixIf | suffixDistinct | 
suffixFunctionCall ;
+
+prefixed = prefixAs | prefixCast | prefixIf | prefixDistinct | 
prefixFunctionCall ;
 
 interval = timeInterval | rowInterval ;
 
@@ -1750,15 +1752,27 @@ timeInterval = composite , "." , ("year" | "years" | 
"quarter" | "quarters" | "m
 
 rowInterval = composite , "." , "rows" ;
 
-cast = composite , ".cast(" , dataType , ")" ;
+suffixCast = composite , ".cast(" , dataType , ")" ;
+
+prefixCast = "cast(" , expression , dataType , ")" ;
 
 dataType = "BYTE" | "SHORT" | "INT" | "LONG" | "FLOAT" | "DOUBLE" | "BOOLEAN" 
| "STRING" | "DECIMAL" | "SQL_DATE" | "SQL_TIME" | "SQL_TIMESTAMP" | 
"INTERVAL_MONTHS" | "INTERVAL_MILLIS" | ( "MAP" , "(" , dataType , "," , 
dataType , ")" ) | ( "PRIMITIVE_ARRAY" , "(" , dataType , ")" ) | ( 
"OBJECT_ARRAY" , "(" , dataType , ")" ) ;
 
-as = composite , ".as(" , fieldReference , ")" ;
+suffixAs = composite , ".as(" , fieldReference , ")" ;
+
+prefixAs = "as(" , expression, fieldReference , ")" ;
+
+suffixIf = composite , ".?(" , expression , "," , expression , ")" ;
+
+prefixIf = "?(" , expression , "," , expression , "," , expression , ")" ;
+
+suffixDistinct = composite , "distinct.()" ;
+
+prefixDistinct = functionIdentifier , ".distinct" , [ "(" , [ expression , { 
"," , expression } ] , ")" ] ;
 
-if = composite , ".?(" , expression , "," , expression , ")" ;
+suffixFunctionCall = composite , "." , functionIdentifier , [ "(" , [ 
expression , { "," , expression } ] , ")" ] ;
 
-functionCall = composite , "." , functionIdentifier , [ "(" , [ expression , { 
"," , expression } ] , ")" ] ;
+prefixFunctionCall = functionIdentifier , [ "(" , [ expression , { "," , 
expression } ] , ")" ] ;
 
 atom = ( "(" , expression , ")" ) | literal | fieldReference ;
 
diff --git 
a/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/expressions/ExpressionParser.scala
 
b/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/expressions/ExpressionParser.scala
index 0b84f9490a1..7fd9309b5db 100644
--- 
a/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/expressions/ExpressionParser.scala
+++ 
b/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/expressions/ExpressionParser.scala
@@ -455,7 +455,7 @@ object ExpressionParser extends JavaTokenParsers with 
PackratParsers {
 
   // suffix/prefix composite
 
-  lazy val composite: PackratParser[Expression] = over | nullLiteral | 
suffixed | prefixed | atom |
+  lazy val composite: PackratParser[Expression] = over | suffixed | 
nullLiteral | prefixed | atom |
 failure("Composite expression expected.")
 
   // unary ops

[GitHub] asfgit closed pull request #6888: [FLINK-10463][table] Null literal cannot be properly parsed in Java Table API function call

2018-11-06 Thread GitBox
asfgit closed pull request #6888: [FLINK-10463][table] Null literal cannot be 
properly parsed in Java Table API function call
URL: https://github.com/apache/flink/pull/6888
 
 
   

This is a PR merged from a forked repository.
As GitHub hides the original diff on merge, it is displayed below for
the sake of provenance:

As this is a foreign pull request (from a fork), the diff is supplied
below (as it won't show otherwise due to GitHub magic):

diff --git a/docs/dev/table/functions.md b/docs/dev/table/functions.md
index e47b31772b3..e652c304e93 100644
--- a/docs/dev/table/functions.md
+++ b/docs/dev/table/functions.md
@@ -4186,7 +4186,7 @@ CAST(value AS type)
   
   
 Returns a new value being cast to type type. See the 
supported types here.
-E.g., CAST('42' AS INT) returns 42.
+E.g., CAST('42' AS INT) returns 42; CAST(NULL AS 
VARCHAR) returns NULL of type VARCHAR.
   
 
   
@@ -4211,7 +4211,7 @@ ANY.cast(TYPE)
   
   
 Returns a new ANY being cast to type TYPE. See the 
supported types here.
-E.g., '42'.cast(INT) returns 42.
+E.g., '42'.cast(INT) returns 42; 
Null(STRING) returns NULL of type STRING.
   
 
 
@@ -4236,7 +4236,7 @@ ANY.cast(TYPE)
   
   
 Returns a new ANY being cast to type TYPE. See the 
supported types here.
-E.g., "42".cast(Types.INT) returns 42.
+E.g., "42".cast(Types.INT) returns 42; 
Null(Types.STRING) returns NULL of type STRING.
   
 
   
diff --git a/docs/dev/table/tableApi.md b/docs/dev/table/tableApi.md
index 6c4e1be0cf7..3f599a8f5a2 100644
--- a/docs/dev/table/tableApi.md
+++ b/docs/dev/table/tableApi.md
@@ -1738,11 +1738,13 @@ term = product , [ ( "+" | "-" ) , product ] ;
 
 product = unary , [ ( "*" | "/" | "%") , unary ] ;
 
-unary = [ "!" | "-" ] , composite ;
+unary = [ "!" | "-" | "+" ] , composite ;
 
-composite = over | nullLiteral | suffixed | atom ;
+composite = over | suffixed | nullLiteral | prefixed | atom ;
 
-suffixed = interval | cast | as | if | functionCall ;
+suffixed = interval | suffixAs | suffixCast | suffixIf | suffixDistinct | 
suffixFunctionCall ;
+
+prefixed = prefixAs | prefixCast | prefixIf | prefixDistinct | 
prefixFunctionCall ;
 
 interval = timeInterval | rowInterval ;
 
@@ -1750,15 +1752,27 @@ timeInterval = composite , "." , ("year" | "years" | 
"quarter" | "quarters" | "m
 
 rowInterval = composite , "." , "rows" ;
 
-cast = composite , ".cast(" , dataType , ")" ;
+suffixCast = composite , ".cast(" , dataType , ")" ;
+
+prefixCast = "cast(" , expression , dataType , ")" ;
 
 dataType = "BYTE" | "SHORT" | "INT" | "LONG" | "FLOAT" | "DOUBLE" | "BOOLEAN" 
| "STRING" | "DECIMAL" | "SQL_DATE" | "SQL_TIME" | "SQL_TIMESTAMP" | 
"INTERVAL_MONTHS" | "INTERVAL_MILLIS" | ( "MAP" , "(" , dataType , "," , 
dataType , ")" ) | ( "PRIMITIVE_ARRAY" , "(" , dataType , ")" ) | ( 
"OBJECT_ARRAY" , "(" , dataType , ")" ) ;
 
-as = composite , ".as(" , fieldReference , ")" ;
+suffixAs = composite , ".as(" , fieldReference , ")" ;
+
+prefixAs = "as(" , expression, fieldReference , ")" ;
+
+suffixIf = composite , ".?(" , expression , "," , expression , ")" ;
+
+prefixIf = "?(" , expression , "," , expression , "," , expression , ")" ;
+
+suffixDistinct = composite , "distinct.()" ;
+
+prefixDistinct = functionIdentifier , ".distinct" , [ "(" , [ expression , { 
"," , expression } ] , ")" ] ;
 
-if = composite , ".?(" , expression , "," , expression , ")" ;
+suffixFunctionCall = composite , "." , functionIdentifier , [ "(" , [ 
expression , { "," , expression } ] , ")" ] ;
 
-functionCall = composite , "." , functionIdentifier , [ "(" , [ expression , { 
"," , expression } ] , ")" ] ;
+prefixFunctionCall = functionIdentifier , [ "(" , [ expression , { "," , 
expression } ] , ")" ] ;
 
 atom = ( "(" , expression , ")" ) | literal | fieldReference ;
 
diff --git 
a/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/expressions/ExpressionParser.scala
 
b/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/expressions/ExpressionParser.scala
index 0b84f9490a1..7fd9309b5db 100644
--- 
a/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/expressions/ExpressionParser.scala
+++ 
b/flink-libraries/flink-table/src/main/scala/org/apache/flink/table/expressions/ExpressionParser.scala
@@ -455,7 +455,7 @@ object ExpressionParser extends JavaTokenParsers with 
PackratParsers {
 
   // suffix/prefix composite
 
-  lazy val composite: PackratParser[Expression] = over | nullLiteral | 
suffixed | prefixed | atom |
+  lazy val composite: PackratParser[Expression] = over | suffixed | 
nullLiteral | prefixed | atom |
 failure("Composite expression expected.")
 
   // unary ops
diff --git 
a/flink-libraries/flink-table/src/test/scala/org/apache/flink/table/expressions/ScalarFunctionsTest.scala
 
b/flink-libraries/flink-table/src/test/scala/org/apache/flink/table/expressions/ScalarFunctionsTest.scala
index 

[jira] [Updated] (FLINK-10805) test_confluent_schema_registry.sh and test_sql_client.sh end-to-end test broken

2018-11-06 Thread Till Rohrmann (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10805?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Till Rohrmann updated FLINK-10805:
--
Description: 
The end-to-end test {{test_confluent_schema_registry.sh}} and 
{{test_sql_client.sh}} seem to be broken after fixing FLINK-10600. 

I suspect that the kafka-common.sh import has not been adapted.

  was:
The end-to-end test {{test_confluent_schema_registry.sh}} and 
{{test_sql_client.sh}} seem to be broken after fixing FLINK-16000. 

I suspect that the kafka-common.sh import has not been adapted.


> test_confluent_schema_registry.sh and test_sql_client.sh end-to-end test 
> broken
> ---
>
> Key: FLINK-10805
> URL: https://issues.apache.org/jira/browse/FLINK-10805
> Project: Flink
>  Issue Type: Bug
>  Components: E2E Tests
>Affects Versions: 1.7.0
>Reporter: Till Rohrmann
>Assignee: Till Rohrmann
>Priority: Major
>  Labels: test-stability
>
> The end-to-end test {{test_confluent_schema_registry.sh}} and 
> {{test_sql_client.sh}} seem to be broken after fixing FLINK-10600. 
> I suspect that the kafka-common.sh import has not been adapted.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10771) Replace hard code of job graph file path with config option for FileJobGraphRetriever

2018-11-06 Thread Till Rohrmann (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676883#comment-16676883
 ] 

Till Rohrmann commented on FLINK-10771:
---

Or better to set a value in the {{Configuration}} in the 
{{AbstractYarnClusterDescriptor}}.

> Replace hard code of job graph file path with config option for 
> FileJobGraphRetriever
> -
>
> Key: FLINK-10771
> URL: https://issues.apache.org/jira/browse/FLINK-10771
> Project: Flink
>  Issue Type: Improvement
>  Components: YARN
>Reporter: vinoyang
>Assignee: vinoyang
>Priority: Major
>
> There is a config option to configure the value :
> {code:java}
> public static final ConfigOption JOB_GRAPH_FILE_PATH = ConfigOptions
>.key("internal.jobgraph-path")
>.defaultValue("job.graph");
> {code}
> However, we used the default hard code in AbstractYarnClusterDescriptor. 
> This is just a preliminary refactoring, and I finally recommend that we use 
> Zookeeper-based storage. It has been implemented in FLINK-10292.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-9699) Add api to replace registered table

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-9699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676878#comment-16676878
 ] 

ASF GitHub Bot commented on FLINK-9699:
---

yanghua commented on issue #6236: [FLINK-9699] [table] Add api to replace 
registered table
URL: https://github.com/apache/flink/pull/6236#issuecomment-436288232
 
 
   @hequn8128 I agree with you. Actually, I have given a similar suggestion 
about `CREATE VIEW` in SQL client. See 
[FLINK-10420](https://issues.apache.org/jira/browse/FLINK-10420?filter=-1). cc 
@twalthr 


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Add api to replace registered table
> ---
>
> Key: FLINK-9699
> URL: https://issues.apache.org/jira/browse/FLINK-9699
> Project: Flink
>  Issue Type: Improvement
>Reporter: Jeff Zhang
>Assignee: Jeff Zhang
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] yanghua commented on issue #6236: [FLINK-9699] [table] Add api to replace registered table

2018-11-06 Thread GitBox
yanghua commented on issue #6236: [FLINK-9699] [table] Add api to replace 
registered table
URL: https://github.com/apache/flink/pull/6236#issuecomment-436288232
 
 
   @hequn8128 I agree with you. Actually, I have given a similar suggestion 
about `CREATE VIEW` in SQL client. See 
[FLINK-10420](https://issues.apache.org/jira/browse/FLINK-10420?filter=-1). cc 
@twalthr 


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10384) Add Sinh math function supported in Table API and SQL

2018-11-06 Thread vinoyang (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10384?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676875#comment-16676875
 ] 

vinoyang commented on FLINK-10384:
--

[~pnovi...@gmail.com] I think this issue could be closed.

> Add Sinh math function supported in Table API and SQL
> -
>
> Key: FLINK-10384
> URL: https://issues.apache.org/jira/browse/FLINK-10384
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table API  SQL
>Reporter: vinoyang
>Assignee: vinoyang
>Priority: Minor
>  Labels: pull-request-available
>
> like FLINK-10340 for adding Cosh math function



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10771) Replace hard code of job graph file path with config option for FileJobGraphRetriever

2018-11-06 Thread vinoyang (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676866#comment-16676866
 ] 

vinoyang commented on FLINK-10771:
--

OK, since it may be used, why don't we refactor it as: if it's not explicitly 
set, then use the default value of the config option (rather than hard-coded 
strings) uniformly?

> Replace hard code of job graph file path with config option for 
> FileJobGraphRetriever
> -
>
> Key: FLINK-10771
> URL: https://issues.apache.org/jira/browse/FLINK-10771
> Project: Flink
>  Issue Type: Improvement
>  Components: YARN
>Reporter: vinoyang
>Assignee: vinoyang
>Priority: Major
>
> There is a config option to configure the value :
> {code:java}
> public static final ConfigOption JOB_GRAPH_FILE_PATH = ConfigOptions
>.key("internal.jobgraph-path")
>.defaultValue("job.graph");
> {code}
> However, we used the default hard code in AbstractYarnClusterDescriptor. 
> This is just a preliminary refactoring, and I finally recommend that we use 
> Zookeeper-based storage. It has been implemented in FLINK-10292.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-10805) test_confluent_schema_registry.sh and test_sql_client.sh end-to-end test broken

2018-11-06 Thread Till Rohrmann (JIRA)
Till Rohrmann created FLINK-10805:
-

 Summary: test_confluent_schema_registry.sh and test_sql_client.sh 
end-to-end test broken
 Key: FLINK-10805
 URL: https://issues.apache.org/jira/browse/FLINK-10805
 Project: Flink
  Issue Type: Bug
  Components: E2E Tests
Affects Versions: 1.7.0
Reporter: Till Rohrmann
Assignee: Till Rohrmann


The end-to-end test {{test_confluent_schema_registry.sh}} and 
{{test_sql_client.sh}} seem to be broken after fixing FLINK-16000. 

I suspect that the kafka-common.sh import has not been adapted.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10802) Add metrics on vertex in job graph picture show

2018-11-06 Thread Till Rohrmann (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676850#comment-16676850
 ] 

Till Rohrmann commented on FLINK-10802:
---

Hi [~lining], I think this could be useful feature but we might also overload 
the job vertex view. I guess it is all about how the user activates this view.

> Add metrics on vertex in job graph picture show
> ---
>
> Key: FLINK-10802
> URL: https://issues.apache.org/jira/browse/FLINK-10802
> Project: Flink
>  Issue Type: Improvement
>  Components: Webfrontend
>Reporter: lining
>Assignee: lining
>Priority: Major
> Attachments: image-2018-11-06-18-12-16-914.png
>
>
> Hi. Now, if someone want see  job metrics, have to click job overview. if we 
> can add some import metrics on graph, can impore it.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10463) Null literal cannot be properly parsed in Java Table API function call

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10463?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676846#comment-16676846
 ] 

ASF GitHub Bot commented on FLINK-10463:


hequn8128 commented on issue #6888: [FLINK-10463][table] Null literal cannot be 
properly parsed in Java Table API function call
URL: https://github.com/apache/flink/pull/6888#issuecomment-436276465
 
 
   Thank you all for the review and suggestions!


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Null literal cannot be properly parsed in Java Table API function call
> --
>
> Key: FLINK-10463
> URL: https://issues.apache.org/jira/browse/FLINK-10463
> Project: Flink
>  Issue Type: Bug
>  Components: Table API  SQL
>Reporter: Xingcan Cui
>Assignee: Hequn Cheng
>Priority: Major
>  Labels: pull-request-available
>
> For example, the expression `Null(STRING).regexpReplace('oo|ar', '')` throws 
> the following exception.
> {code:java}
> org.apache.flink.table.api.ExpressionParserException: Could not parse 
> expression at column 13: string matching regex 
> `(?i)\Qas\E(?![_$\p{javaJavaIdentifierPart}])' expected but `.' found
> Null(STRING).regexpReplace('oo|ar', '')
> ^
>   at 
> org.apache.flink.table.expressions.ExpressionParser$.throwError(ExpressionParser.scala:576)
>   at 
> org.apache.flink.table.expressions.ExpressionParser$.parseExpression(ExpressionParser.scala:569)
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] hequn8128 commented on issue #6888: [FLINK-10463][table] Null literal cannot be properly parsed in Java Table API function call

2018-11-06 Thread GitBox
hequn8128 commented on issue #6888: [FLINK-10463][table] Null literal cannot be 
properly parsed in Java Table API function call
URL: https://github.com/apache/flink/pull/6888#issuecomment-436276465
 
 
   Thank you all for the review and suggestions!


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10771) Replace hard code of job graph file path with config option for FileJobGraphRetriever

2018-11-06 Thread Till Rohrmann (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676841#comment-16676841
 ] 

Till Rohrmann commented on FLINK-10771:
---

But from the implementation of {{FileJobGraphRetriever}} you cannot tell that 
it won't be used. At the moment this might be right but some people might use 
the {{StandaloneJobClusterEntrypoint}} differently. Thus, I would not change 
this.

> Replace hard code of job graph file path with config option for 
> FileJobGraphRetriever
> -
>
> Key: FLINK-10771
> URL: https://issues.apache.org/jira/browse/FLINK-10771
> Project: Flink
>  Issue Type: Improvement
>  Components: YARN
>Reporter: vinoyang
>Assignee: vinoyang
>Priority: Major
>
> There is a config option to configure the value :
> {code:java}
> public static final ConfigOption JOB_GRAPH_FILE_PATH = ConfigOptions
>.key("internal.jobgraph-path")
>.defaultValue("job.graph");
> {code}
> However, we used the default hard code in AbstractYarnClusterDescriptor. 
> This is just a preliminary refactoring, and I finally recommend that we use 
> Zookeeper-based storage. It has been implemented in FLINK-10292.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-9699) Add api to replace registered table

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-9699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676842#comment-16676842
 ] 

ASF GitHub Bot commented on FLINK-9699:
---

hequn8128 commented on issue #6236: [FLINK-9699] [table] Add api to replace 
registered table
URL: https://github.com/apache/flink/pull/6236#issuecomment-436275448
 
 
   @zjffdu Thanks for the PR. I think this feature is very useful. I haven't 
taken a deep look at the code. Considering the api, how about change these 
`register()` functions with replace parameter to `registerOrReplaceXXX()` ?  It 
makes our api more alignment with SQL, such as CREATE OR REPLACE. What do you 
guys think here? @twalthr @yanghua 


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Add api to replace registered table
> ---
>
> Key: FLINK-9699
> URL: https://issues.apache.org/jira/browse/FLINK-9699
> Project: Flink
>  Issue Type: Improvement
>Reporter: Jeff Zhang
>Assignee: Jeff Zhang
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] hequn8128 commented on issue #6236: [FLINK-9699] [table] Add api to replace registered table

2018-11-06 Thread GitBox
hequn8128 commented on issue #6236: [FLINK-9699] [table] Add api to replace 
registered table
URL: https://github.com/apache/flink/pull/6236#issuecomment-436275448
 
 
   @zjffdu Thanks for the PR. I think this feature is very useful. I haven't 
taken a deep look at the code. Considering the api, how about change these 
`register()` functions with replace parameter to `registerOrReplaceXXX()` ?  It 
makes our api more alignment with SQL, such as CREATE OR REPLACE. What do you 
guys think here? @twalthr @yanghua 


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] shaoxuan-wang commented on issue #7025: [FLINK-10797] "IntelliJ Setup" link is broken in README.md

2018-11-06 Thread GitBox
shaoxuan-wang commented on issue #7025: [FLINK-10797] "IntelliJ Setup" link is 
broken in README.md
URL: https://github.com/apache/flink/pull/7025#issuecomment-436263199
 
 
   > @shaoxuan-wang regardless the duplication of this pr(#7025 ) and #7024 I'd 
like to point out two concerns on this thread.
   > 
   > 1. would it be better to fix the issue by replacing site link with 
relative path?(thus keep link issue isolated among versions)
   > 2. what version(s) should this fix back port to? (I guess all versions 
with `README.md` link to the site url are broken)
   > 
   > x. there is another incorrect path of "IntelliJ Setup" in `pom.xml` as 
modify within #7024 .
   
   Thanks @TisonKun .  I did not notice there is an existing PR.@xndai can 
you please close this PR as  #7024 already addressed problem reported in 
FLINK-10797


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10797) "IntelliJ Setup" link is broken in Readme.md

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10797?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676792#comment-16676792
 ] 

ASF GitHub Bot commented on FLINK-10797:


shaoxuan-wang commented on issue #7025: [FLINK-10797] "IntelliJ Setup" link is 
broken in README.md
URL: https://github.com/apache/flink/pull/7025#issuecomment-436263199
 
 
   > @shaoxuan-wang regardless the duplication of this pr(#7025 ) and #7024 I'd 
like to point out two concerns on this thread.
   > 
   > 1. would it be better to fix the issue by replacing site link with 
relative path?(thus keep link issue isolated among versions)
   > 2. what version(s) should this fix back port to? (I guess all versions 
with `README.md` link to the site url are broken)
   > 
   > x. there is another incorrect path of "IntelliJ Setup" in `pom.xml` as 
modify within #7024 .
   
   Thanks @TisonKun .  I did not notice there is an existing PR.@xndai can 
you please close this PR as  #7024 already addressed problem reported in 
FLINK-10797


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> "IntelliJ Setup" link is broken in Readme.md
> 
>
> Key: FLINK-10797
> URL: https://issues.apache.org/jira/browse/FLINK-10797
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.7.0
>Reporter: Xiening Dai
>Assignee: TisonKun
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>
> The link points to 
> https://github.com/apache/flink/blob/master/docs/internals/ide_setup.md#intellij-idea
>  which is a 404 not found.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10795) STDDEV_POP error

2018-11-06 Thread Hequn Cheng (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10795?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676791#comment-16676791
 ] 

Hequn Cheng commented on FLINK-10795:
-

Hi, I had looked into this issue. I think it is a precision issue.

As for sql, {{ROUND(STDDEV_POP(CAST(lon AS DOUBLE)),2)}},  {{STDDEV_POP}} will 
be converted to {{SQRT((SUM(x * x) - SUM( x ) * SUM( x ) / COUNT( x )) / COUNT( 
x ))}} in {{AggregateReduceFunctionsRule}} during optimization. However, due to 
precision issue, the result of {{SUM(x * x) - SUM( x ) * SUM( x ) / COUNT( x 
)}} would be negative which making sqrt return a value of NaN. {{ROUND}} throws 
an exception taking NaN as an input.

We may not be able to solve the precision problem. However, I think we can find 
ways to avoid the exception. For example, we can change  {{SQRT((SUM(x * x) - 
SUM( x ) * SUM( x ) / COUNT( x )) / COUNT( x ))}} to 
SQRT({color:red}ABS{color}((SUM(x * x) - SUM( x ) * SUM( x ) / COUNT( x ))) / 
COUNT( x )) by adding an abs(). What do you guys think?



> STDDEV_POP error
> 
>
> Key: FLINK-10795
> URL: https://issues.apache.org/jira/browse/FLINK-10795
> Project: Flink
>  Issue Type: Improvement
>  Components: Table API  SQL
>Affects Versions: 1.6.2
>Reporter: Flavio Pompermaier
>Assignee: winifredtang
>Priority: Major
> Attachments: FlinkTableApiError.java, test.tsv
>
>
> if using STDDEV_POP in the attached job the following error is thrown (with 
> Flink 1.6.1):
>  
> {code:java}
> Exception in thread "main" 
> org.apache.flink.runtime.client.JobExecutionException: 
> java.lang.NumberFormatException
>  at 
> org.apache.flink.runtime.minicluster.MiniCluster.executeJobBlocking(MiniCluster.java:623)
>  at org.apache.flink.client.LocalExecutor.executePlan(LocalExecutor.java:235)
>  at 
> org.apache.flink.api.java.LocalEnvironment.execute(LocalEnvironment.java:91)
>  at 
> org.apache.flink.api.java.ExecutionEnvironment.execute(ExecutionEnvironment.java:816)
>  at org.apache.flink.api.java.DataSet.collect(DataSet.java:413)
>  at org.apache.flink.api.java.DataSet.print(DataSet.java:1652)
>  at 
> it.okkam.datalinks.batch.flink.operations.FlinkTableApiError.main(FlinkTableApiError.java:466)
> Caused by: java.lang.NumberFormatException
>  at java.math.BigDecimal.(BigDecimal.java:494)
>  at java.math.BigDecimal.(BigDecimal.java:383)
>  at java.math.BigDecimal.(BigDecimal.java:806)
>  at java.math.BigDecimal.valueOf(BigDecimal.java:1274)
>  at org.apache.calcite.runtime.SqlFunctions.sround(SqlFunctions.java:1242)
>  at DataSetCalcRule$6909.flatMap(Unknown Source)
>  at 
> org.apache.flink.table.runtime.FlatMapRunner.flatMap(FlatMapRunner.scala:52)
>  at 
> org.apache.flink.table.runtime.FlatMapRunner.flatMap(FlatMapRunner.scala:31)
>  at 
> org.apache.flink.runtime.operators.chaining.ChainedFlatMapDriver.collect(ChainedFlatMapDriver.java:80)
>  at 
> org.apache.flink.runtime.operators.util.metrics.CountingCollector.collect(CountingCollector.java:35)
>  at DataSetSingleRowJoinRule$6450.join(Unknown Source)
>  at 
> org.apache.flink.table.runtime.MapJoinLeftRunner.flatMap(MapJoinLeftRunner.scala:35)
>  at 
> org.apache.flink.runtime.operators.FlatMapDriver.run(FlatMapDriver.java:109)
>  at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:503)
>  at org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:368)
>  at org.apache.flink.runtime.taskmanager.Task.run(Task.java:711)
>  at java.lang.Thread.run(Thread.java:748)
> {code}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-10419) ClassNotFoundException while deserializing user exceptions from checkpointing

2018-11-06 Thread ASF GitHub Bot (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10419?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-10419:
---
Labels: pull-request-available  (was: )

> ClassNotFoundException while deserializing user exceptions from checkpointing
> -
>
> Key: FLINK-10419
> URL: https://issues.apache.org/jira/browse/FLINK-10419
> Project: Flink
>  Issue Type: Bug
>  Components: Distributed Coordination, State Backends, Checkpointing
>Affects Versions: 1.5.0, 1.5.1, 1.5.2, 1.5.3, 1.5.4, 1.6.0, 1.6.1, 1.7.0
>Reporter: Nico Kruber
>Assignee: Nico Kruber
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>
> It seems that somewhere in the operator's failure handling, we hand a 
> user-code exception to the checkpoint coordinator via Java serialization but 
> it will then fail during the de-serialization because the class is not 
> available. This will result in the following error shadowing the real one:
> {code}
> java.lang.ClassNotFoundException: 
> org.apache.flink.streaming.connectors.kafka.FlinkKafka011Exception
> at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
> at sun.misc.Launcher.loadClass(Launcher.java:338)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
> at java.lang.Class.forName0(Native Method)
> at java.lang.Class.forName(Class.java:348)
> at 
> org.apache.flink.util.InstantiationUtil.resolveClass(InstantiationUtil.java:76)
> at 
> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1859)
> at 
> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1745)
> at 
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2033)
> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1567)
> at 
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2278)
> at 
> java.io.ObjectInputStream.defaultReadObject(ObjectInputStream.java:557)
> at java.lang.Throwable.readObject(Throwable.java:914)
> at sun.reflect.GeneratedMethodAccessor158.invoke(Unknown Source)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at 
> java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1158)
> at 
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2169)
> at 
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2060)
> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1567)
> at java.io.ObjectInputStream.readObject(ObjectInputStream.java:427)
> at 
> org.apache.flink.runtime.rpc.messages.RemoteRpcInvocation.readObject(RemoteRpcInvocation.java:222)
> at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at 
> java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1158)
> at 
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2169)
> at 
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2060)
> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1567)
> at java.io.ObjectInputStream.readObject(ObjectInputStream.java:427)
> at 
> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:502)
> at 
> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:489)
> at 
> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:477)
> at 
> org.apache.flink.util.SerializedValue.deserializeValue(SerializedValue.java:58)
> at 
> org.apache.flink.runtime.rpc.messages.RemoteRpcInvocation.deserializeMethodInvocation(RemoteRpcInvocation.java:118)
> at 
> org.apache.flink.runtime.rpc.messages.RemoteRpcInvocation.getMethodName(RemoteRpcInvocation.java:59)
> at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:214)
> at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:162)
> at 
> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:70)
> at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.onReceive(AkkaRpcActor.java:142)
> at 
> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.onReceive(FencedAkkaRpcActor.java:40)
>   

[jira] [Commented] (FLINK-10419) ClassNotFoundException while deserializing user exceptions from checkpointing

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10419?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676768#comment-16676768
 ] 

ASF GitHub Bot commented on FLINK-10419:


NicoK opened a new pull request #7033: [FLINK-10419][checkpoint] fix 
ClassNotFoundException while deserializing user exceptions
URL: https://github.com/apache/flink/pull/7033
 
 
   ## What is the purpose of the change
   
   `DeclineCheckpoint` tries to make sure that it does wrap any exception which 
can be and/or contain user code into a `SerializedThrowable`. It also contains 
special handling of a few exceptions that are not wrapped into 
`SerializedThrowable` (`CheckpointDeclineException` sub-classes are claimed to 
not contain user-code) but if these carry user-code exceptions as *suppressed* 
exceptions, the deserialization at the JobManager will fail and the RPC call 
will not be processed at all - we thus rely on another method to actually get 
rid of the checkpoint, e.g. a timeout.
   
   ## Brief change log
   
   - remove the special handling of `CheckpointDeclineException` sub-classes
   
   ## Verifying this change
   
   This change is already covered by existing tests, such as checkpoint tests 
and IT cases.
   
   I also manually verified that no code is relying on checks such as 
`instanceof CheckpointDeclineException` or the explicitly mentioned 
sub-classes, so that it is ok to replace these with `SerializedThrowable` in 
the RPC.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): **no**
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: **no**
 - The serializers: **no**
 - The runtime per-record code paths (performance sensitive): **no**
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: **yes**
 - The S3 file system connector: **no**
   
   ## Documentation
   
 - Does this pull request introduce a new feature? **no**
 - If yes, how is the feature documented? **JavaDocs**
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> ClassNotFoundException while deserializing user exceptions from checkpointing
> -
>
> Key: FLINK-10419
> URL: https://issues.apache.org/jira/browse/FLINK-10419
> Project: Flink
>  Issue Type: Bug
>  Components: Distributed Coordination, State Backends, Checkpointing
>Affects Versions: 1.5.0, 1.5.1, 1.5.2, 1.5.3, 1.5.4, 1.6.0, 1.6.1, 1.7.0
>Reporter: Nico Kruber
>Assignee: Nico Kruber
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>
> It seems that somewhere in the operator's failure handling, we hand a 
> user-code exception to the checkpoint coordinator via Java serialization but 
> it will then fail during the de-serialization because the class is not 
> available. This will result in the following error shadowing the real one:
> {code}
> java.lang.ClassNotFoundException: 
> org.apache.flink.streaming.connectors.kafka.FlinkKafka011Exception
> at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
> at sun.misc.Launcher.loadClass(Launcher.java:338)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
> at java.lang.Class.forName0(Native Method)
> at java.lang.Class.forName(Class.java:348)
> at 
> org.apache.flink.util.InstantiationUtil.resolveClass(InstantiationUtil.java:76)
> at 
> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1859)
> at 
> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1745)
> at 
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2033)
> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1567)
> at 
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2278)
> at 
> java.io.ObjectInputStream.defaultReadObject(ObjectInputStream.java:557)
> at java.lang.Throwable.readObject(Throwable.java:914)
> at sun.reflect.GeneratedMethodAccessor158.invoke(Unknown Source)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at 
> java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1158)
> at 
> 

[GitHub] NicoK opened a new pull request #7033: [FLINK-10419][checkpoint] fix ClassNotFoundException while deserializing user exceptions

2018-11-06 Thread GitBox
NicoK opened a new pull request #7033: [FLINK-10419][checkpoint] fix 
ClassNotFoundException while deserializing user exceptions
URL: https://github.com/apache/flink/pull/7033
 
 
   ## What is the purpose of the change
   
   `DeclineCheckpoint` tries to make sure that it does wrap any exception which 
can be and/or contain user code into a `SerializedThrowable`. It also contains 
special handling of a few exceptions that are not wrapped into 
`SerializedThrowable` (`CheckpointDeclineException` sub-classes are claimed to 
not contain user-code) but if these carry user-code exceptions as *suppressed* 
exceptions, the deserialization at the JobManager will fail and the RPC call 
will not be processed at all - we thus rely on another method to actually get 
rid of the checkpoint, e.g. a timeout.
   
   ## Brief change log
   
   - remove the special handling of `CheckpointDeclineException` sub-classes
   
   ## Verifying this change
   
   This change is already covered by existing tests, such as checkpoint tests 
and IT cases.
   
   I also manually verified that no code is relying on checks such as 
`instanceof CheckpointDeclineException` or the explicitly mentioned 
sub-classes, so that it is ok to replace these with `SerializedThrowable` in 
the RPC.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): **no**
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: **no**
 - The serializers: **no**
 - The runtime per-record code paths (performance sensitive): **no**
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: **yes**
 - The S3 file system connector: **no**
   
   ## Documentation
   
 - Does this pull request introduce a new feature? **no**
 - If yes, how is the feature documented? **JavaDocs**
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (FLINK-10804) Transfer suppressed exceptions with SerializedThrowable

2018-11-06 Thread Nico Kruber (JIRA)
Nico Kruber created FLINK-10804:
---

 Summary: Transfer suppressed exceptions with SerializedThrowable
 Key: FLINK-10804
 URL: https://issues.apache.org/jira/browse/FLINK-10804
 Project: Flink
  Issue Type: Improvement
  Components: Distributed Coordination
Affects Versions: 1.7.0
Reporter: Nico Kruber


Currently, {{SerializedThrowable}} mimics the original exception including the 
causes but not the suppressed exceptions. You may argue whether these should 
really be transferred as well or whether it would be enough to have them in the 
logging of the component which sent the message.

However, it may be surprising to silently drop suppressed exceptions.




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (FLINK-10797) "IntelliJ Setup" link is broken in Readme.md

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10797?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676750#comment-16676750
 ] 

ASF GitHub Bot commented on FLINK-10797:


TisonKun edited a comment on issue #7025: [FLINK-10797] "IntelliJ Setup" link 
is broken in README.md
URL: https://github.com/apache/flink/pull/7025#issuecomment-436249401
 
 
   @shaoxuan-wang regardless the duplication of this pr(#7025 ) and #7024 I'd 
like to point out two concerns on this thread.
   
   1. would it be better to fix the issue by replacing site link with relative 
path?(thus keep link issue isolated among versions)
   2. what version(s) should this fix back port to? (I guess all versions with 
`README.md` link to the site url are broken)
   
   x. there is another incorrect path of "IntelliJ Setup" in `pom.xml` as 
modify within #7024 .


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> "IntelliJ Setup" link is broken in Readme.md
> 
>
> Key: FLINK-10797
> URL: https://issues.apache.org/jira/browse/FLINK-10797
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.7.0
>Reporter: Xiening Dai
>Assignee: TisonKun
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>
> The link points to 
> https://github.com/apache/flink/blob/master/docs/internals/ide_setup.md#intellij-idea
>  which is a 404 not found.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] TisonKun edited a comment on issue #7025: [FLINK-10797] "IntelliJ Setup" link is broken in README.md

2018-11-06 Thread GitBox
TisonKun edited a comment on issue #7025: [FLINK-10797] "IntelliJ Setup" link 
is broken in README.md
URL: https://github.com/apache/flink/pull/7025#issuecomment-436249401
 
 
   @shaoxuan-wang regardless the duplication of this pr(#7025 ) and #7024 I'd 
like to point out two concerns on this thread.
   
   1. would it be better to fix the issue by replacing site link with relative 
path?(thus keep link issue isolated among versions)
   2. what version(s) should this fix back port to? (I guess all versions with 
`README.md` link to the site url are broken)
   
   x. there is another incorrect path of "IntelliJ Setup" in `pom.xml` as 
modify within #7024 .


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10797) "IntelliJ Setup" link is broken in Readme.md

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10797?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676747#comment-16676747
 ] 

ASF GitHub Bot commented on FLINK-10797:


TisonKun commented on issue #7025: [FLINK-10797] "IntelliJ Setup" link is 
broken in README.md
URL: https://github.com/apache/flink/pull/7025#issuecomment-436249401
 
 
   @shaoxuan-wang regardless the duplication of this pr(#7025 ) and #7024 I'd 
like to point out two concerns on this thread.
   
   1. would it be better to fix the issue by replacing site link with relative 
path?(thus keep link issue isolated among versions)
   2. what version(s) should this fix back port to?
   x. there is another incorrect path of "IntelliJ Setup" in `pom.xml` as 
modify within #7024 .


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> "IntelliJ Setup" link is broken in Readme.md
> 
>
> Key: FLINK-10797
> URL: https://issues.apache.org/jira/browse/FLINK-10797
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.7.0
>Reporter: Xiening Dai
>Assignee: TisonKun
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>
> The link points to 
> https://github.com/apache/flink/blob/master/docs/internals/ide_setup.md#intellij-idea
>  which is a 404 not found.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] TisonKun commented on issue #7025: [FLINK-10797] "IntelliJ Setup" link is broken in README.md

2018-11-06 Thread GitBox
TisonKun commented on issue #7025: [FLINK-10797] "IntelliJ Setup" link is 
broken in README.md
URL: https://github.com/apache/flink/pull/7025#issuecomment-436249401
 
 
   @shaoxuan-wang regardless the duplication of this pr(#7025 ) and #7024 I'd 
like to point out two concerns on this thread.
   
   1. would it be better to fix the issue by replacing site link with relative 
path?(thus keep link issue isolated among versions)
   2. what version(s) should this fix back port to?
   x. there is another incorrect path of "IntelliJ Setup" in `pom.xml` as 
modify within #7024 .


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10353) Restoring a KafkaProducer with Semantic.EXACTLY_ONCE from a savepoint written with Semantic.AT_LEAST_ONCE fails with NPE

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10353?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676695#comment-16676695
 ] 

ASF GitHub Bot commented on FLINK-10353:


StefanRRichter commented on issue #7010: [FLINK-10353][kafka] Support change of 
transactional semantics in Kaf…
URL: https://github.com/apache/flink/pull/7010#issuecomment-436235901
 
 
   Addressed the comments. Please have another look.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Restoring a KafkaProducer with Semantic.EXACTLY_ONCE from a savepoint written 
> with Semantic.AT_LEAST_ONCE fails with NPE
> 
>
> Key: FLINK-10353
> URL: https://issues.apache.org/jira/browse/FLINK-10353
> Project: Flink
>  Issue Type: Improvement
>  Components: Kafka Connector
>Affects Versions: 1.5.3, 1.6.0
>Reporter: Konstantin Knauf
>Assignee: Stefan Richter
>Priority: Critical
>  Labels: pull-request-available
>
> If a KafkaProducer with {{Semantic.EXACTLY_ONCE}} is restored from a 
> savepoint written with {{Semantic.AT_LEAST_ONCE}} the job fails on restore 
> with the NPE below. This makes it impossible to upgrade an AT_LEAST_ONCE 
> pipeline to an EXACTL_ONCE pipeline statefully.
> {quote}
> java.lang.NullPointerException
>   at java.util.Hashtable.put(Hashtable.java:460)
>   at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer011.initTransactionalProducer(FlinkKafkaProducer011.java:955)
>   at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer011.recoverAndCommit(FlinkKafkaProducer011.java:733)
>   at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer011.recoverAndCommit(FlinkKafkaProducer011.java:93)
>   at 
> org.apache.flink.streaming.api.functions.sink.TwoPhaseCommitSinkFunction.recoverAndCommitInternal(TwoPhaseCommitSinkFunction.java:373)
>   at 
> org.apache.flink.streaming.api.functions.sink.TwoPhaseCommitSinkFunction.initializeState(TwoPhaseCommitSinkFunction.java:333)
>   at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer011.initializeState(FlinkKafkaProducer011.java:867)
>   at 
> org.apache.flink.streaming.util.functions.StreamingFunctionUtils.tryRestoreFunction(StreamingFunctionUtils.java:178)
>   at 
> org.apache.flink.streaming.util.functions.StreamingFunctionUtils.restoreFunctionState(StreamingFunctionUtils.java:160)
>   at 
> org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.initializeState(AbstractUdfStreamOperator.java:96)
>   at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:254)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeState(StreamTask.java:738)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:289)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:711)
>   at java.lang.Thread.run(Thread.java:748){quote}
> The reason is, that for {{Semantic.AT_LEAST_ONCE}} the snapshotted state of 
> the {{TwoPhaseCommitFunction}} is of the form 
> "TransactionHolder\{handle=KafkaTransactionState [transactionalId=null, 
> producerId=-1, epoch=-1], transactionStartTime=1537175471175}".



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] StefanRRichter commented on issue #7010: [FLINK-10353][kafka] Support change of transactional semantics in Kaf…

2018-11-06 Thread GitBox
StefanRRichter commented on issue #7010: [FLINK-10353][kafka] Support change of 
transactional semantics in Kaf…
URL: https://github.com/apache/flink/pull/7010#issuecomment-436235901
 
 
   Addressed the comments. Please have another look.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10600) Provide End-to-end test cases for modern Kafka connectors

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10600?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676692#comment-16676692
 ] 

ASF GitHub Bot commented on FLINK-10600:


yanghua commented on issue #6924: [FLINK-10600] Provide End-to-end test cases 
for modern Kafka connectors
URL: https://github.com/apache/flink/pull/6924#issuecomment-436235491
 
 
   @shaoxuan-wang Can you share your error log with us? The test cases related 
to kafka connector are not very stable.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Provide End-to-end test cases for modern Kafka connectors
> -
>
> Key: FLINK-10600
> URL: https://issues.apache.org/jira/browse/FLINK-10600
> Project: Flink
>  Issue Type: Sub-task
>  Components: Kafka Connector
>Reporter: vinoyang
>Assignee: vinoyang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] yanghua commented on issue #6924: [FLINK-10600] Provide End-to-end test cases for modern Kafka connectors

2018-11-06 Thread GitBox
yanghua commented on issue #6924: [FLINK-10600] Provide End-to-end test cases 
for modern Kafka connectors
URL: https://github.com/apache/flink/pull/6924#issuecomment-436235491
 
 
   @shaoxuan-wang Can you share your error log with us? The test cases related 
to kafka connector are not very stable.


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] shaoxuan-wang commented on issue #6924: [FLINK-10600] Provide End-to-end test cases for modern Kafka connectors

2018-11-06 Thread GitBox
shaoxuan-wang commented on issue #6924: [FLINK-10600] Provide End-to-end test 
cases for modern Kafka connectors
URL: https://github.com/apache/flink/pull/6924#issuecomment-436234220
 
 
   @pnowojski @yanghuaI got this test failure when I rebase the flink 
master to my own repository which is connected with travis-ci.org.  And I did 
not find the origin travis test result in this thread. Is it possible this is a 
flaky test?


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10600) Provide End-to-end test cases for modern Kafka connectors

2018-11-06 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10600?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16676690#comment-16676690
 ] 

ASF GitHub Bot commented on FLINK-10600:


shaoxuan-wang commented on issue #6924: [FLINK-10600] Provide End-to-end test 
cases for modern Kafka connectors
URL: https://github.com/apache/flink/pull/6924#issuecomment-436234220
 
 
   @pnowojski @yanghuaI got this test failure when I rebase the flink 
master to my own repository which is connected with travis-ci.org.  And I did 
not find the origin travis test result in this thread. Is it possible this is a 
flaky test?


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Provide End-to-end test cases for modern Kafka connectors
> -
>
> Key: FLINK-10600
> URL: https://issues.apache.org/jira/browse/FLINK-10600
> Project: Flink
>  Issue Type: Sub-task
>  Components: Kafka Connector
>Reporter: vinoyang
>Assignee: vinoyang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.7.0
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-10801) Fix sql client integrate elasticsearch connector test failure

2018-11-06 Thread vinoyang (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-10801?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

vinoyang updated FLINK-10801:
-
Description: 
It usually reports : 
{code:java}
FAIL SQL Client Elasticsearch Upsert: Output hash mismatch. Got 
6187222e109ee9222e6b2f117742070c, expected 982cb32908def9801e781381c1b8a8db.
head hexdump of actual:
000 { \n " h i t s " : { \n 
010 " t o t a l " : 3 , \n
020 " m a x _ s c o r e " 
030 : 1 . 0 , \n " h i t s
040 " : [ \n { \n 
050 " _ i n d e x " :
060 " m y _ u s e r s " , \n 
070 " _ t y p e " : "
080 u s e r " , \n "
090 _ i d " : " 1 _ B o b "
0a0 , \n " _ s c o r
0b0 e " : 1 . 0 , \n 
0ba
{code}
the actual hash means : 
{code:java}
{
  "hits" : {
"total" : 3,
"max_score" : 1.0,
"hits" : [
  {
"_index" : "my_users",
"_type" : "user",
"_id" : "1_Bob  ",
"_score" : 1.0,
"_source" : {
  "user_id" : 1,
  "user_name" : "Bob  ",
  "user_count" : 1
}
  },
  {
"_index" : "my_users",
"_type" : "user",
"_id" : "22_Alice",
"_score" : 1.0,
"_source" : {
  "user_id" : 22,
  "user_name" : "Alice",
  "user_count" : 1
}
  },
  {
"_index" : "my_users",
"_type" : "user",
"_id" : "42_Greg ",
"_score" : 1.0,
"_source" : {
  "user_id" : 42,
  "user_name" : "Greg ",
  "user_count" : 3
}
  }
]
  }
}

{code}
the expected hash code means : 
{code:java}
{
  "hits" : {
"total" : 3,
"max_score" : 1.0,
"hits" : [
  {
"_index" : "my_users",
"_type" : "user",
"_id" : "1_Bob  ",
"_score" : 1.0,
"_source" : {
  "user_id" : 1,
  "user_name" : "Bob  ",
  "user_count" : 2
}
  },
  {
"_index" : "my_users",
"_type" : "user",
"_id" : "22_Alice",
"_score" : 1.0,
"_source" : {
  "user_id" : 22,
  "user_name" : "Alice",
  "user_count" : 1
}
  },
  {
"_index" : "my_users",
"_type" : "user",
"_id" : "42_Greg ",
"_score" : 1.0,
"_source" : {
  "user_id" : 42,
  "user_name" : "Greg ",
  "user_count" : 3
}
  }
]
  }
}
{code}
It seems that the user count for "Bob" is off by 1.

The speculation is due to the premature acquisition of aggregated statistics 
from Elasticsearch.

 

> Fix sql client integrate elasticsearch connector test failure
> -
>
> Key: FLINK-10801
> URL: https://issues.apache.org/jira/browse/FLINK-10801
> Project: Flink
>  Issue Type: Bug
>  Components: E2E Tests
>Reporter: vinoyang
>Assignee: vinoyang
>Priority: Major
>
> It usually reports : 
> {code:java}
> FAIL SQL Client Elasticsearch Upsert: Output hash mismatch. Got 
> 6187222e109ee9222e6b2f117742070c, expected 982cb32908def9801e781381c1b8a8db.
> head hexdump of actual:
> 000 { \n " h i t s " : { \n 
> 010 " t o t a l " : 3 , \n
> 020 " m a x _ s c o r e " 
> 030 : 1 . 0 , \n " h i t s
> 040 " : [ \n { \n 
> 050 " _ i n d e x " :
> 060 " m y _ u s e r s " , \n 
> 070 " _ t y p e " : "
> 080 u s e r " , \n "
> 090 _ i d " : " 1 _ B o b "
> 0a0 , \n " _ s c o r
> 0b0 e " : 1 . 0 , \n 
> 0ba
> {code}
> the actual hash means : 
> {code:java}
> {
>   "hits" : {
> "total" : 3,
> "max_score" : 1.0,
> "hits" : [
>   {
> "_index" : "my_users",
> "_type" : "user",
> "_id" : "1_Bob  ",
> "_score" : 1.0,
> "_source" : {
>   "user_id" : 1,
>   "user_name" : "Bob  ",
>   "user_count" : 1
> }
>   },
>   {
> "_index" : "my_users",
> "_type" : "user",
> "_id" : "22_Alice",
> "_score" : 1.0,
> "_source" : {
>   "user_id" : 22,
>   "user_name" : "Alice",
>   "user_count" : 1
> }
>   },
>   {
> "_index" : "my_users",
> "_type" : "user",
> "_id" : "42_Greg ",
> "_score" : 1.0,
> "_source" : {
>   "user_id" : 42,
>   "user_name" : "Greg ",
>   "user_count" : 3
> }
>   }
> ]
>   }
> }
> {code}
> the expected hash code means : 
> {code:java}
> {
>   "hits" : {
> "total" : 3,
> "max_score" : 1.0,
> "hits" : [
>   {
> "_index" : "my_users",
> "_type" : "user",
> "_id" : "1_Bob  ",
> "_score" : 1.0,
> "_source" : {
>   "user_id" : 1,
>   "user_name" : "Bob  ",
>   "user_count" : 2
> }
>   },
>  

  1   2   3   >