[jira] [Assigned] (FLINK-6615) tmp directory not cleaned up on shutdown
[ https://issues.apache.org/jira/browse/FLINK-6615?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bowen Li reassigned FLINK-6615: --- Assignee: (was: Bowen Li) > tmp directory not cleaned up on shutdown > > > Key: FLINK-6615 > URL: https://issues.apache.org/jira/browse/FLINK-6615 > Project: Flink > Issue Type: Bug > Components: Local Runtime >Affects Versions: 1.2.0 >Reporter: Andrey >Priority: Major > > Steps to reproduce: > 1) Stop task manager gracefully (kill -6 ) > 2) In the logs: > {code} > 2017-05-17 13:35:50,147 INFO org.apache.zookeeper.ClientCnxn > - EventThread shut down [main-EventThread] > 2017-05-17 13:35:50,200 ERROR > org.apache.flink.runtime.io.disk.iomanager.IOManager - IOManager > failed to properly clean up temp file directory: > /mnt/flink/tmp/flink-io-66f1d0ec-8976-41bf-9575-f80b181b0e47 > [flink-akka.actor.default-dispatcher-2] > java.nio.file.DirectoryNotEmptyException: > /mnt/flink/tmp/flink-io-66f1d0ec-8976-41bf-9575-f80b181b0e47 > at > sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242) > at > sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103) > at java.nio.file.Files.delete(Files.java:1126) > at org.apache.flink.util.FileUtils.deleteDirectory(FileUtils.java:154) > at > org.apache.flink.runtime.io.disk.iomanager.IOManager.shutdown(IOManager.java:109) > at > org.apache.flink.runtime.io.disk.iomanager.IOManagerAsync.shutdown(IOManagerAsync.java:185) > at > org.apache.flink.runtime.taskmanager.TaskManager.postStop(TaskManager.scala:241) > at akka.actor.Actor$class.aroundPostStop(Actor.scala:477) > {code} > Expected: > * on shutdown delete non-empty directory anyway. > Notes: > * after process terminated, I've checked > "/mnt/flink/tmp/flink-io-66f1d0ec-8976-41bf-9575-f80b181b0e47" directory and > didn't find anything there. So it looks like timing issue. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (FLINK-6615) tmp directory not cleaned up on shutdown
[ https://issues.apache.org/jira/browse/FLINK-6615?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bowen Li resolved FLINK-6615. - Resolution: Won't Fix > tmp directory not cleaned up on shutdown > > > Key: FLINK-6615 > URL: https://issues.apache.org/jira/browse/FLINK-6615 > Project: Flink > Issue Type: Bug > Components: Local Runtime >Affects Versions: 1.2.0 >Reporter: Andrey >Assignee: Bowen Li >Priority: Major > > Steps to reproduce: > 1) Stop task manager gracefully (kill -6 ) > 2) In the logs: > {code} > 2017-05-17 13:35:50,147 INFO org.apache.zookeeper.ClientCnxn > - EventThread shut down [main-EventThread] > 2017-05-17 13:35:50,200 ERROR > org.apache.flink.runtime.io.disk.iomanager.IOManager - IOManager > failed to properly clean up temp file directory: > /mnt/flink/tmp/flink-io-66f1d0ec-8976-41bf-9575-f80b181b0e47 > [flink-akka.actor.default-dispatcher-2] > java.nio.file.DirectoryNotEmptyException: > /mnt/flink/tmp/flink-io-66f1d0ec-8976-41bf-9575-f80b181b0e47 > at > sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242) > at > sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103) > at java.nio.file.Files.delete(Files.java:1126) > at org.apache.flink.util.FileUtils.deleteDirectory(FileUtils.java:154) > at > org.apache.flink.runtime.io.disk.iomanager.IOManager.shutdown(IOManager.java:109) > at > org.apache.flink.runtime.io.disk.iomanager.IOManagerAsync.shutdown(IOManagerAsync.java:185) > at > org.apache.flink.runtime.taskmanager.TaskManager.postStop(TaskManager.scala:241) > at akka.actor.Actor$class.aroundPostStop(Actor.scala:477) > {code} > Expected: > * on shutdown delete non-empty directory anyway. > Notes: > * after process terminated, I've checked > "/mnt/flink/tmp/flink-io-66f1d0ec-8976-41bf-9575-f80b181b0e47" directory and > didn't find anything there. So it looks like timing issue. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (FLINK-7947) Let ParameterTool return a dedicated GlobalJobParameters object
[ https://issues.apache.org/jira/browse/FLINK-7947?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bowen Li reassigned FLINK-7947: --- Assignee: (was: Bowen Li) > Let ParameterTool return a dedicated GlobalJobParameters object > --- > > Key: FLINK-7947 > URL: https://issues.apache.org/jira/browse/FLINK-7947 > Project: Flink > Issue Type: Improvement > Components: Client >Affects Versions: 1.4.0 >Reporter: Till Rohrmann >Priority: Major > > The {{ParameterTool}} directly implements the {{GlobalJobParameters}} > interface. Additionally it has grown over time to not only store the > configuration parameters but also to record which parameters have been > requested and what default value was set. This information is irrelevant on > the server side when setting a {{GlobalJobParameters}} object via > {{ExecutionConfig#setGlobalJobParameters}}. > Since we don't separate the {{ParameterTool}} logic and the actual data view, > users ran into problems when reusing the same {{ParameterTool}} to start > multiple jobs concurrently (see FLINK-7943). I think it would be a much > clearer separation of concerns if we would actually split the > {{GlobalJobParameters}} from the {{ParameterTool}}. > Furthermore, we should think about whether {{ParameterTool#get}} should have > side effects or not as it does right now. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #:
Github user jimmec commented on the pull request: https://github.com/apache/flink/commit/3062e29a8bbb667d098c43a0b225d5602049#commitcomment-27642949 In flink-connectors/flink-connector-kinesis/src/test/java/org/apache/flink/streaming/connectors/kinesis/internals/KinesisDataFetcherTest.java: In flink-connectors/flink-connector-kinesis/src/test/java/org/apache/flink/streaming/connectors/kinesis/internals/KinesisDataFetcherTest.java on line 113: Is DummyFlink`Kafka`Consumer a typo?? It's at the least inconsistent with the comment above this line. ---
[jira] [Closed] (FLINK-8548) Add Streaming State Machine Example
[ https://issues.apache.org/jira/browse/FLINK-8548?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stephan Ewen closed FLINK-8548. --- > Add Streaming State Machine Example > --- > > Key: FLINK-8548 > URL: https://issues.apache.org/jira/browse/FLINK-8548 > Project: Flink > Issue Type: Sub-task > Components: Examples >Reporter: Stephan Ewen >Assignee: Stephan Ewen >Priority: Major > Fix For: 1.5.0 > > > Add the example from > https://github.com/StephanEwen/flink-demos/tree/master/streaming-state-machine > to the Flink examples. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (FLINK-8548) Add Streaming State Machine Example
[ https://issues.apache.org/jira/browse/FLINK-8548?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stephan Ewen resolved FLINK-8548. - Resolution: Done Done in 80883feeeba5232ecd41fe6759c69bf2ff1a6483 > Add Streaming State Machine Example > --- > > Key: FLINK-8548 > URL: https://issues.apache.org/jira/browse/FLINK-8548 > Project: Flink > Issue Type: Sub-task > Components: Examples >Reporter: Stephan Ewen >Assignee: Stephan Ewen >Priority: Major > Fix For: 1.5.0 > > > Add the example from > https://github.com/StephanEwen/flink-demos/tree/master/streaming-state-machine > to the Flink examples. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-8680) Name printing sinks by default.
[ https://issues.apache.org/jira/browse/FLINK-8680?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stephan Ewen closed FLINK-8680. --- > Name printing sinks by default. > --- > > Key: FLINK-8680 > URL: https://issues.apache.org/jira/browse/FLINK-8680 > Project: Flink > Issue Type: Improvement >Affects Versions: 1.4.0 >Reporter: Stephan Ewen >Assignee: Stephan Ewen >Priority: Trivial > Fix For: 1.5.0 > > > The sinks that pring to std. out and std. err show up as "Sink: Unnamed" in > logs and the UI. > They should be named "Print to Std. Out" and "Print to Std. Err" by default. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (FLINK-8549) Move TimerServiceOptions to TaskManagerOptions
[ https://issues.apache.org/jira/browse/FLINK-8549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stephan Ewen resolved FLINK-8549. - Resolution: Fixed Fixed via 85bfc073680667e9daa16439d33fa86c1c0be133 > Move TimerServiceOptions to TaskManagerOptions > -- > > Key: FLINK-8549 > URL: https://issues.apache.org/jira/browse/FLINK-8549 > Project: Flink > Issue Type: Improvement > Components: Local Runtime >Reporter: Stephan Ewen >Assignee: Stephan Ewen >Priority: Major > Fix For: 1.5.0 > > > The {{TimerServiceOptions}} are in the wrong place (prohibit generation of > config docs) and cause over-fragmentation of the options in the code base. > I propose to simple move the one option from that class to the > {{TaskManagerOptions}}, as it relates to task execution. Other shutdown > related options are in there already. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (FLINK-8680) Name printing sinks by default.
[ https://issues.apache.org/jira/browse/FLINK-8680?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stephan Ewen resolved FLINK-8680. - Resolution: Fixed Fixed via 21491b2717e3a60e63dcf4fb0aa839b9885be67a > Name printing sinks by default. > --- > > Key: FLINK-8680 > URL: https://issues.apache.org/jira/browse/FLINK-8680 > Project: Flink > Issue Type: Improvement >Affects Versions: 1.4.0 >Reporter: Stephan Ewen >Assignee: Stephan Ewen >Priority: Trivial > Fix For: 1.5.0 > > > The sinks that pring to std. out and std. err show up as "Sink: Unnamed" in > logs and the UI. > They should be named "Print to Std. Out" and "Print to Std. Err" by default. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-8549) Move TimerServiceOptions to TaskManagerOptions
[ https://issues.apache.org/jira/browse/FLINK-8549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stephan Ewen closed FLINK-8549. --- > Move TimerServiceOptions to TaskManagerOptions > -- > > Key: FLINK-8549 > URL: https://issues.apache.org/jira/browse/FLINK-8549 > Project: Flink > Issue Type: Improvement > Components: Local Runtime >Reporter: Stephan Ewen >Assignee: Stephan Ewen >Priority: Major > Fix For: 1.5.0 > > > The {{TimerServiceOptions}} are in the wrong place (prohibit generation of > config docs) and cause over-fragmentation of the options in the code base. > I propose to simple move the one option from that class to the > {{TaskManagerOptions}}, as it relates to task execution. Other shutdown > related options are in there already. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-8548) Add Streaming State Machine Example
[ https://issues.apache.org/jira/browse/FLINK-8548?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368705#comment-16368705 ] ASF GitHub Bot commented on FLINK-8548: --- Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5401 > Add Streaming State Machine Example > --- > > Key: FLINK-8548 > URL: https://issues.apache.org/jira/browse/FLINK-8548 > Project: Flink > Issue Type: Sub-task > Components: Examples >Reporter: Stephan Ewen >Assignee: Stephan Ewen >Priority: Major > Fix For: 1.5.0 > > > Add the example from > https://github.com/StephanEwen/flink-demos/tree/master/streaming-state-machine > to the Flink examples. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-8682) Make start/stop cluster scripts work without SSH for local HA setups
[ https://issues.apache.org/jira/browse/FLINK-8682?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stephan Ewen closed FLINK-8682. --- > Make start/stop cluster scripts work without SSH for local HA setups > > > Key: FLINK-8682 > URL: https://issues.apache.org/jira/browse/FLINK-8682 > Project: Flink > Issue Type: Improvement > Components: Startup Shell Scripts >Affects Versions: 1.4.0 >Reporter: Stephan Ewen >Assignee: Stephan Ewen >Priority: Minor > Fix For: 1.5.0 > > > The startup should work for purely local (testing) cluster without setups > without SSH. > While the shell scripts handle this correctly for TaskManagers, they don't > handle it correctly for JobManagers. As a consequence, {{start-cluster.sh}} > does not work without SSH when high availability is enabled. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (FLINK-8682) Make start/stop cluster scripts work without SSH for local HA setups
[ https://issues.apache.org/jira/browse/FLINK-8682?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stephan Ewen resolved FLINK-8682. - Resolution: Fixed Fixed via 39711fb6f1d801e067e121803ea82b1c2742c956 > Make start/stop cluster scripts work without SSH for local HA setups > > > Key: FLINK-8682 > URL: https://issues.apache.org/jira/browse/FLINK-8682 > Project: Flink > Issue Type: Improvement > Components: Startup Shell Scripts >Affects Versions: 1.4.0 >Reporter: Stephan Ewen >Assignee: Stephan Ewen >Priority: Minor > Fix For: 1.5.0 > > > The startup should work for purely local (testing) cluster without setups > without SSH. > While the shell scripts handle this correctly for TaskManagers, they don't > handle it correctly for JobManagers. As a consequence, {{start-cluster.sh}} > does not work without SSH when high availability is enabled. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5402: [FLINK-8549] [config] Move TimerServiceOptions int...
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5402 ---
[jira] [Commented] (FLINK-8549) Move TimerServiceOptions to TaskManagerOptions
[ https://issues.apache.org/jira/browse/FLINK-8549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368704#comment-16368704 ] ASF GitHub Bot commented on FLINK-8549: --- Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5402 > Move TimerServiceOptions to TaskManagerOptions > -- > > Key: FLINK-8549 > URL: https://issues.apache.org/jira/browse/FLINK-8549 > Project: Flink > Issue Type: Improvement > Components: Local Runtime >Reporter: Stephan Ewen >Assignee: Stephan Ewen >Priority: Major > Fix For: 1.5.0 > > > The {{TimerServiceOptions}} are in the wrong place (prohibit generation of > config docs) and cause over-fragmentation of the options in the code base. > I propose to simple move the one option from that class to the > {{TaskManagerOptions}}, as it relates to task execution. Other shutdown > related options are in there already. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-8681) Remove planVisualizer.html move notice
[ https://issues.apache.org/jira/browse/FLINK-8681?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stephan Ewen closed FLINK-8681. --- > Remove planVisualizer.html move notice > -- > > Key: FLINK-8681 > URL: https://issues.apache.org/jira/browse/FLINK-8681 > Project: Flink > Issue Type: Improvement > Components: Build System >Affects Versions: 1.4.0 >Reporter: Stephan Ewen >Assignee: Stephan Ewen >Priority: Trivial > Fix For: 1.5.0 > > > The {{planVisualizer.html}} for optimizer plans is no longer in the Flink > distribution, but we hold a notice there that the visualizer has moved to the > website. > That notice has been there for many versions (since Flink 1.0) and can be > removed now. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (FLINK-8681) Remove planVisualizer.html move notice
[ https://issues.apache.org/jira/browse/FLINK-8681?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stephan Ewen resolved FLINK-8681. - Resolution: Fixed Fixed via 096bf80d93d4f27b2e57aaff8754ebbe54151ecb > Remove planVisualizer.html move notice > -- > > Key: FLINK-8681 > URL: https://issues.apache.org/jira/browse/FLINK-8681 > Project: Flink > Issue Type: Improvement > Components: Build System >Affects Versions: 1.4.0 >Reporter: Stephan Ewen >Assignee: Stephan Ewen >Priority: Trivial > Fix For: 1.5.0 > > > The {{planVisualizer.html}} for optimizer plans is no longer in the Flink > distribution, but we hold a notice there that the visualizer has moved to the > website. > That notice has been there for many versions (since Flink 1.0) and can be > removed now. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5401: [FLINK-8548] [examples] Add state machine example
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5401 ---
[jira] [Commented] (FLINK-8599) Improve the failure behavior of the FileInputFormat for bad files
[ https://issues.apache.org/jira/browse/FLINK-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368678#comment-16368678 ] ASF GitHub Bot commented on FLINK-8599: --- GitHub user ChengzhiZhao opened a pull request: https://github.com/apache/flink/pull/5521 [FLINK-8599] Improve the failure behavior of the FileInputFormat for … ## What is the purpose of the change This pull request is intent to improve the failure behavior of the ContinuousFileReader, currently if a bad file (for example, a different schema been dropped in this folder) came to the path and flink will do several retries. However, since the file path persist in the checkpoint, when people tried to resume from external checkpoint, it threw the following error on no file been found and the process cannot move forward. `java.io.IOException: Error opening the Input Split s3a://myfile [0,904]: No such file or directory: s3a://myfile` The change is to check if the path exist before open the file, if error occurs and bad file removed, flink should resume the process and continue. ## Brief change log - *Add a file exist check before open the file * ## Verifying this change - *Manually verified the change by introduce a bad file while continuously monitoring the folder, after remove the bad file, the process continued.* ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: (no) - The S3 file system connector: (no) ## Documentation - Does this pull request introduce a new feature? (no) - If yes, how is the feature documented? (not applicable) You can merge this pull request into a Git repository by running: $ git pull https://github.com/ChengzhiZhao/flink Improve_failure_behavior_FileInputFormat Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5521.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5521 commit 6fa8ef212c536acee56b2e9831ec92d1059449ff Author: Chengzhi ZhaoDate: 2018-02-18T18:23:32Z [FLINK-8599] Improve the failure behavior of the FileInputFormat for bad files > Improve the failure behavior of the FileInputFormat for bad files > - > > Key: FLINK-8599 > URL: https://issues.apache.org/jira/browse/FLINK-8599 > Project: Flink > Issue Type: New Feature > Components: DataStream API >Affects Versions: 1.4.0, 1.3.2 >Reporter: Chengzhi Zhao >Priority: Major > > So we have a s3 path that flink is monitoring that path to see new files > available. > {code:java} > val avroInputStream_activity = env.readFile(format, path, > FileProcessingMode.PROCESS_CONTINUOUSLY, 1) {code} > > I am doing both internal and external check pointing and let's say there is a > bad file (for example, a different schema been dropped in this folder) came > to the path and flink will do several retries. I want to take those bad files > and let the process continue. However, since the file path persist in the > checkpoint, when I try to resume from external checkpoint, it threw the > following error on no file been found. > > {code:java} > java.io.IOException: Error opening the Input Split s3a://myfile [0,904]: No > such file or directory: s3a://myfile{code} > > As [~fhue...@gmail.com] suggested, we could check if a path exists and before > trying to read a file and ignore the input split instead of throwing an > exception and causing a failure. > > Also, I am thinking about add an error output for bad files as an option to > users. So if there is any bad files exist we could move them in a separated > path and do further analysis. > > Not sure how people feel about it, but I'd like to contribute on it if people > think this can be an improvement. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-7477) Use "hadoop classpath" to augment classpath when available
[ https://issues.apache.org/jira/browse/FLINK-7477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368677#comment-16368677 ] Ken Krugler commented on FLINK-7477: It works (at least when running with YARN via EMR). I believe that's because the version of Hadoop on the EMR master matches what we're running against; on my machine, I have to switch between multiple versions of Hadoop for various (consulting) clients who are on different versions of Hadoop, and my {{hadoop}} symlink wound up pointing to a different version of Hadoop than what Flink was using. Related note - the 1.4 release fixed some shading issues we were running into with AWS client classes (mostly around {{HttpCore}} stuff), but to get everything working properly I felt like I did some voodoo with class exclusions in the {{maven-shade-plugin}} section of my {{pom.xml}}, which still feels fragile. > Use "hadoop classpath" to augment classpath when available > -- > > Key: FLINK-7477 > URL: https://issues.apache.org/jira/browse/FLINK-7477 > Project: Flink > Issue Type: Bug > Components: Startup Shell Scripts >Reporter: Aljoscha Krettek >Assignee: Aljoscha Krettek >Priority: Major > Fix For: 1.4.0 > > > Currently, some cloud environments don't properly put the Hadoop jars into > {{HADOOP_CLASSPATH}} (or don't set {{HADOOP_CLASSPATH}}) at all. We should > check in {{config.sh}} if the {{hadoop}} binary is on the path and augment > our {{INTERNAL_HADOOP_CLASSPATHS}} with the result of {{hadoop classpath}} in > our scripts. > This will improve the out-of-box experience of users that otherwise have to > manually set {{HADOOP_CLASSPATH}}. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5521: [FLINK-8599] Improve the failure behavior of the F...
GitHub user ChengzhiZhao opened a pull request: https://github.com/apache/flink/pull/5521 [FLINK-8599] Improve the failure behavior of the FileInputFormat for ⦠## What is the purpose of the change This pull request is intent to improve the failure behavior of the ContinuousFileReader, currently if a bad file (for example, a different schema been dropped in this folder) came to the path and flink will do several retries. However, since the file path persist in the checkpoint, when people tried to resume from external checkpoint, it threw the following error on no file been found and the process cannot move forward. `java.io.IOException: Error opening the Input Split s3a://myfile [0,904]: No such file or directory: s3a://myfile` The change is to check if the path exist before open the file, if error occurs and bad file removed, flink should resume the process and continue. ## Brief change log - *Add a file exist check before open the file * ## Verifying this change - *Manually verified the change by introduce a bad file while continuously monitoring the folder, after remove the bad file, the process continued.* ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: (no) - The S3 file system connector: (no) ## Documentation - Does this pull request introduce a new feature? (no) - If yes, how is the feature documented? (not applicable) You can merge this pull request into a Git repository by running: $ git pull https://github.com/ChengzhiZhao/flink Improve_failure_behavior_FileInputFormat Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5521.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5521 commit 6fa8ef212c536acee56b2e9831ec92d1059449ff Author: Chengzhi ZhaoDate: 2018-02-18T18:23:32Z [FLINK-8599] Improve the failure behavior of the FileInputFormat for bad files ---
[jira] [Closed] (FLINK-8692) Mistake in MyMapFunction code snippet
[ https://issues.apache.org/jira/browse/FLINK-8692?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler closed FLINK-8692. --- Resolution: Fixed master: 0644f4f77dde4532714fb60bf8c646365fca5fc0 1.4: 528317c8f23a7f20493510e57c21c72a1fc42b3d > Mistake in MyMapFunction code snippet > - > > Key: FLINK-8692 > URL: https://issues.apache.org/jira/browse/FLINK-8692 > Project: Flink > Issue Type: Bug > Components: Documentation >Affects Versions: 1.5.0, 1.4.1 >Reporter: Matt Hagen >Assignee: Chesnay Schepler >Priority: Trivial > Fix For: 1.5.0, 1.4.2 > > Attachments: my-map-function.png > > Original Estimate: 1h > Remaining Estimate: 1h > > The MyMapFunction code snippet on the [Basic API > Concepts|https://ci.apache.org/projects/flink/flink-docs-release-1.4/dev/api_concepts.html] > page has an extra parenthesis. Just remove the last parenthesis. See the > attached screenshot. Thanks. > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-8692) Mistake in MyMapFunction code snippet
[ https://issues.apache.org/jira/browse/FLINK-8692?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler updated FLINK-8692: Labels: (was: documentation) > Mistake in MyMapFunction code snippet > - > > Key: FLINK-8692 > URL: https://issues.apache.org/jira/browse/FLINK-8692 > Project: Flink > Issue Type: Bug > Components: Documentation >Affects Versions: 1.5.0, 1.4.1 >Reporter: Matt Hagen >Assignee: Chesnay Schepler >Priority: Trivial > Fix For: 1.5.0, 1.4.2 > > Attachments: my-map-function.png > > Original Estimate: 1h > Remaining Estimate: 1h > > The MyMapFunction code snippet on the [Basic API > Concepts|https://ci.apache.org/projects/flink/flink-docs-release-1.4/dev/api_concepts.html] > page has an extra parenthesis. Just remove the last parenthesis. See the > attached screenshot. Thanks. > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-8692) Mistake in MyMapFunction code snippet
[ https://issues.apache.org/jira/browse/FLINK-8692?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler updated FLINK-8692: Affects Version/s: 1.5.0 > Mistake in MyMapFunction code snippet > - > > Key: FLINK-8692 > URL: https://issues.apache.org/jira/browse/FLINK-8692 > Project: Flink > Issue Type: Bug > Components: Documentation >Affects Versions: 1.5.0, 1.4.1 >Reporter: Matt Hagen >Assignee: Chesnay Schepler >Priority: Trivial > Fix For: 1.5.0, 1.4.2 > > Attachments: my-map-function.png > > Original Estimate: 1h > Remaining Estimate: 1h > > The MyMapFunction code snippet on the [Basic API > Concepts|https://ci.apache.org/projects/flink/flink-docs-release-1.4/dev/api_concepts.html] > page has an extra parenthesis. Just remove the last parenthesis. See the > attached screenshot. Thanks. > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (FLINK-8692) Mistake in MyMapFunction code snippet
[ https://issues.apache.org/jira/browse/FLINK-8692?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler reassigned FLINK-8692: --- Assignee: Chesnay Schepler > Mistake in MyMapFunction code snippet > - > > Key: FLINK-8692 > URL: https://issues.apache.org/jira/browse/FLINK-8692 > Project: Flink > Issue Type: Bug > Components: Documentation >Affects Versions: 1.5.0, 1.4.1 >Reporter: Matt Hagen >Assignee: Chesnay Schepler >Priority: Trivial > Fix For: 1.5.0, 1.4.2 > > Attachments: my-map-function.png > > Original Estimate: 1h > Remaining Estimate: 1h > > The MyMapFunction code snippet on the [Basic API > Concepts|https://ci.apache.org/projects/flink/flink-docs-release-1.4/dev/api_concepts.html] > page has an extra parenthesis. Just remove the last parenthesis. See the > attached screenshot. Thanks. > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-8692) Mistake in MyMapFunction code snippet
[ https://issues.apache.org/jira/browse/FLINK-8692?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler updated FLINK-8692: Fix Version/s: 1.4.2 1.5.0 > Mistake in MyMapFunction code snippet > - > > Key: FLINK-8692 > URL: https://issues.apache.org/jira/browse/FLINK-8692 > Project: Flink > Issue Type: Bug > Components: Documentation >Affects Versions: 1.5.0, 1.4.1 >Reporter: Matt Hagen >Assignee: Chesnay Schepler >Priority: Trivial > Fix For: 1.5.0, 1.4.2 > > Attachments: my-map-function.png > > Original Estimate: 1h > Remaining Estimate: 1h > > The MyMapFunction code snippet on the [Basic API > Concepts|https://ci.apache.org/projects/flink/flink-docs-release-1.4/dev/api_concepts.html] > page has an extra parenthesis. Just remove the last parenthesis. See the > attached screenshot. Thanks. > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-8693) initPythonInterpreter is called twice with same arguments in InterpreterUtils#initAndExecPythonScript
[ https://issues.apache.org/jira/browse/FLINK-8693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler closed FLINK-8693. --- Resolution: Fixed master: 8cea2f6576da713e5625e00ef2dbe34a410a1df1 > initPythonInterpreter is called twice with same arguments in > InterpreterUtils#initAndExecPythonScript > - > > Key: FLINK-8693 > URL: https://issues.apache.org/jira/browse/FLINK-8693 > Project: Flink > Issue Type: Bug > Components: Python API >Affects Versions: 1.5.0 >Reporter: Ted Yu >Assignee: Chesnay Schepler >Priority: Trivial > Fix For: 1.5.0 > > > {code} > initPythonInterpreter(fullArgs, scriptDirectory.toUri().getPath(), > scriptName); > PythonInterpreter pythonInterpreter = initPythonInterpreter(fullArgs, > scriptDirectory.toUri().getPath(), scriptName); > {code} > The two calls to initPythonInterpreter() have the same parameters. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-8693) initPythonInterpreter is called twice with same arguments in InterpreterUtils#initAndExecPythonScript
[ https://issues.apache.org/jira/browse/FLINK-8693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler updated FLINK-8693: Priority: Trivial (was: Minor) > initPythonInterpreter is called twice with same arguments in > InterpreterUtils#initAndExecPythonScript > - > > Key: FLINK-8693 > URL: https://issues.apache.org/jira/browse/FLINK-8693 > Project: Flink > Issue Type: Bug > Components: Python API >Affects Versions: 1.5.0 >Reporter: Ted Yu >Assignee: Chesnay Schepler >Priority: Trivial > Fix For: 1.5.0 > > > {code} > initPythonInterpreter(fullArgs, scriptDirectory.toUri().getPath(), > scriptName); > PythonInterpreter pythonInterpreter = initPythonInterpreter(fullArgs, > scriptDirectory.toUri().getPath(), scriptName); > {code} > The two calls to initPythonInterpreter() have the same parameters. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-8693) initPythonInterpreter is called twice with same arguments in InterpreterUtils#initAndExecPythonScript
[ https://issues.apache.org/jira/browse/FLINK-8693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler updated FLINK-8693: Fix Version/s: 1.5.0 > initPythonInterpreter is called twice with same arguments in > InterpreterUtils#initAndExecPythonScript > - > > Key: FLINK-8693 > URL: https://issues.apache.org/jira/browse/FLINK-8693 > Project: Flink > Issue Type: Bug > Components: Python API >Affects Versions: 1.5.0 >Reporter: Ted Yu >Assignee: Chesnay Schepler >Priority: Minor > Fix For: 1.5.0 > > > {code} > initPythonInterpreter(fullArgs, scriptDirectory.toUri().getPath(), > scriptName); > PythonInterpreter pythonInterpreter = initPythonInterpreter(fullArgs, > scriptDirectory.toUri().getPath(), scriptName); > {code} > The two calls to initPythonInterpreter() have the same parameters. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-8693) initPythonInterpreter is called twice with same arguments in InterpreterUtils#initAndExecPythonScript
[ https://issues.apache.org/jira/browse/FLINK-8693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler updated FLINK-8693: Component/s: Python API > initPythonInterpreter is called twice with same arguments in > InterpreterUtils#initAndExecPythonScript > - > > Key: FLINK-8693 > URL: https://issues.apache.org/jira/browse/FLINK-8693 > Project: Flink > Issue Type: Bug > Components: Python API >Affects Versions: 1.5.0 >Reporter: Ted Yu >Assignee: Chesnay Schepler >Priority: Minor > Fix For: 1.5.0 > > > {code} > initPythonInterpreter(fullArgs, scriptDirectory.toUri().getPath(), > scriptName); > PythonInterpreter pythonInterpreter = initPythonInterpreter(fullArgs, > scriptDirectory.toUri().getPath(), scriptName); > {code} > The two calls to initPythonInterpreter() have the same parameters. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-8693) initPythonInterpreter is called twice with same arguments in InterpreterUtils#initAndExecPythonScript
[ https://issues.apache.org/jira/browse/FLINK-8693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler updated FLINK-8693: Affects Version/s: 1.5.0 > initPythonInterpreter is called twice with same arguments in > InterpreterUtils#initAndExecPythonScript > - > > Key: FLINK-8693 > URL: https://issues.apache.org/jira/browse/FLINK-8693 > Project: Flink > Issue Type: Bug >Affects Versions: 1.5.0 >Reporter: Ted Yu >Assignee: Chesnay Schepler >Priority: Minor > > {code} > initPythonInterpreter(fullArgs, scriptDirectory.toUri().getPath(), > scriptName); > PythonInterpreter pythonInterpreter = initPythonInterpreter(fullArgs, > scriptDirectory.toUri().getPath(), scriptName); > {code} > The two calls to initPythonInterpreter() have the same parameters. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (FLINK-8693) initPythonInterpreter is called twice with same arguments in InterpreterUtils#initAndExecPythonScript
[ https://issues.apache.org/jira/browse/FLINK-8693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler reassigned FLINK-8693: --- Assignee: Chesnay Schepler > initPythonInterpreter is called twice with same arguments in > InterpreterUtils#initAndExecPythonScript > - > > Key: FLINK-8693 > URL: https://issues.apache.org/jira/browse/FLINK-8693 > Project: Flink > Issue Type: Bug >Reporter: Ted Yu >Assignee: Chesnay Schepler >Priority: Minor > > {code} > initPythonInterpreter(fullArgs, scriptDirectory.toUri().getPath(), > scriptName); > PythonInterpreter pythonInterpreter = initPythonInterpreter(fullArgs, > scriptDirectory.toUri().getPath(), scriptName); > {code} > The two calls to initPythonInterpreter() have the same parameters. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (FLINK-8693) initPythonInterpreter is called twice with same arguments in InterpreterUtils#initAndExecPythonScript
Ted Yu created FLINK-8693: - Summary: initPythonInterpreter is called twice with same arguments in InterpreterUtils#initAndExecPythonScript Key: FLINK-8693 URL: https://issues.apache.org/jira/browse/FLINK-8693 Project: Flink Issue Type: Bug Reporter: Ted Yu {code} initPythonInterpreter(fullArgs, scriptDirectory.toUri().getPath(), scriptName); PythonInterpreter pythonInterpreter = initPythonInterpreter(fullArgs, scriptDirectory.toUri().getPath(), scriptName); {code} The two calls to initPythonInterpreter() have the same parameters. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (FLINK-8692) Mistake in MyMapFunction code snippet
Matt Hagen created FLINK-8692: - Summary: Mistake in MyMapFunction code snippet Key: FLINK-8692 URL: https://issues.apache.org/jira/browse/FLINK-8692 Project: Flink Issue Type: Bug Components: Documentation Affects Versions: 1.4.1 Reporter: Matt Hagen Attachments: my-map-function.png The MyMapFunction code snippet on the [Basic API Concepts|https://ci.apache.org/projects/flink/flink-docs-release-1.4/dev/api_concepts.html] page has an extra parenthesis. Just remove the last parenthesis. See the attached screenshot. Thanks. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5520: [FLINK-8599] Improve the failure behavior of the F...
Github user ChengzhiZhao closed the pull request at: https://github.com/apache/flink/pull/5520 ---
[jira] [Commented] (FLINK-8599) Improve the failure behavior of the FileInputFormat for bad files
[ https://issues.apache.org/jira/browse/FLINK-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368614#comment-16368614 ] ASF GitHub Bot commented on FLINK-8599: --- Github user ChengzhiZhao closed the pull request at: https://github.com/apache/flink/pull/5520 > Improve the failure behavior of the FileInputFormat for bad files > - > > Key: FLINK-8599 > URL: https://issues.apache.org/jira/browse/FLINK-8599 > Project: Flink > Issue Type: New Feature > Components: DataStream API >Affects Versions: 1.4.0, 1.3.2 >Reporter: Chengzhi Zhao >Priority: Major > > So we have a s3 path that flink is monitoring that path to see new files > available. > {code:java} > val avroInputStream_activity = env.readFile(format, path, > FileProcessingMode.PROCESS_CONTINUOUSLY, 1) {code} > > I am doing both internal and external check pointing and let's say there is a > bad file (for example, a different schema been dropped in this folder) came > to the path and flink will do several retries. I want to take those bad files > and let the process continue. However, since the file path persist in the > checkpoint, when I try to resume from external checkpoint, it threw the > following error on no file been found. > > {code:java} > java.io.IOException: Error opening the Input Split s3a://myfile [0,904]: No > such file or directory: s3a://myfile{code} > > As [~fhue...@gmail.com] suggested, we could check if a path exists and before > trying to read a file and ignore the input split instead of throwing an > exception and causing a failure. > > Also, I am thinking about add an error output for bad files as an option to > users. So if there is any bad files exist we could move them in a separated > path and do further analysis. > > Not sure how people feel about it, but I'd like to contribute on it if people > think this can be an improvement. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (FLINK-8684) Rework MesosTaskManagerParameters#MESOS_RM_TASKS_SLOTS
[ https://issues.apache.org/jira/browse/FLINK-8684?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler reassigned FLINK-8684: --- Assignee: Chesnay Schepler > Rework MesosTaskManagerParameters#MESOS_RM_TASKS_SLOTS > -- > > Key: FLINK-8684 > URL: https://issues.apache.org/jira/browse/FLINK-8684 > Project: Flink > Issue Type: Improvement > Components: Documentation, Mesos >Affects Versions: 1.5.0 >Reporter: Chesnay Schepler >Assignee: Chesnay Schepler >Priority: Major > Fix For: 1.5.0 > > > Currently, {{MesosTaskManagerParameters#MESOS_RM_TASKS_SLOTS}} mimics > {{TaskManagerOptions#NUM_TASK_SLOTS}}: > {code:java} > public static final ConfigOption MESOS_RM_TASKS_SLOTS = > key(ConfigConstants.TASK_MANAGER_NUM_TASK_SLOTS) > .defaultValue(1); > public static final ConfigOption NUM_TASK_SLOTS = > key("taskmanager.numberOfTaskSlots") > .defaultValue(1) > .withDescription("..."); > {code} > This pattern is problematic as this creates 2 documentation entries for > {{taskmanager.numberOfTaskSlots}} with different descriptions, and opens the > potential for different defaults. Ultimately this causes the documentation to > become ambiguous. > I thus propose to either outright remove this option or turn it into an > actual alias: > {code:java} > public static final ConfigOption MESOS_RM_TASKS_SLOTS = > TaskManagerOptions.NUM_TASK_SLOTS; > {code} > As a side-effect of FLINK-8683 we can ensure that no differing config options > exist for a given key. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (FLINK-8689) Add runtime support of distinct filter using MapView for GenerateAggregation
[ https://issues.apache.org/jira/browse/FLINK-8689?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rong Rong reassigned FLINK-8689: Assignee: Rong Rong > Add runtime support of distinct filter using MapView for GenerateAggregation > > > Key: FLINK-8689 > URL: https://issues.apache.org/jira/browse/FLINK-8689 > Project: Flink > Issue Type: Sub-task >Reporter: Rong Rong >Assignee: Rong Rong >Priority: Major > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (FLINK-8690) Update logical rule set to generate FlinkLogicalAggregate explicitly allow distinct operator
[ https://issues.apache.org/jira/browse/FLINK-8690?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rong Rong reassigned FLINK-8690: Assignee: Rong Rong > Update logical rule set to generate FlinkLogicalAggregate explicitly allow > distinct operator > - > > Key: FLINK-8690 > URL: https://issues.apache.org/jira/browse/FLINK-8690 > Project: Flink > Issue Type: Sub-task >Reporter: Rong Rong >Assignee: Rong Rong >Priority: Major > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (FLINK-8691) Update table API to support distinct operator on data stream
Rong Rong created FLINK-8691: Summary: Update table API to support distinct operator on data stream Key: FLINK-8691 URL: https://issues.apache.org/jira/browse/FLINK-8691 Project: Flink Issue Type: Sub-task Reporter: Rong Rong -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (FLINK-8690) Update logical rule set to generate FlinkLogicalAggregate explicitly allow distinct operator
Rong Rong created FLINK-8690: Summary: Update logical rule set to generate FlinkLogicalAggregate explicitly allow distinct operator Key: FLINK-8690 URL: https://issues.apache.org/jira/browse/FLINK-8690 Project: Flink Issue Type: Sub-task Reporter: Rong Rong -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (FLINK-8689) Add runtime support of distinct filter using MapView for GenerateAggregation
Rong Rong created FLINK-8689: Summary: Add runtime support of distinct filter using MapView for GenerateAggregation Key: FLINK-8689 URL: https://issues.apache.org/jira/browse/FLINK-8689 Project: Flink Issue Type: Sub-task Reporter: Rong Rong -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-8688) Enable distinct aggregation for data stream on Table/SQL API
[ https://issues.apache.org/jira/browse/FLINK-8688?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368585#comment-16368585 ] Rong Rong commented on FLINK-8688: -- Related distinct aggregate feature request JIRA > Enable distinct aggregation for data stream on Table/SQL API > > > Key: FLINK-8688 > URL: https://issues.apache.org/jira/browse/FLINK-8688 > Project: Flink > Issue Type: Improvement > Components: Table API SQL >Reporter: Rong Rong >Assignee: Rong Rong >Priority: Major > > Distinct aggregation is not currently supported on data stream with Table/SQL > API. This is an umbrella task for enabling distinct aggregation in various > use cases. > Discussion doc can be found here: > https://docs.google.com/document/d/1zj6OA-K2hi7ah8Fo-xTQB-mVmYfm6LsN2_NHgTCVmJI/edit?usp=sharing > > Distinct aggregation is a very important feature in SQL processing and there > are many JIRAs currently open with various use cases. The goal is to create > one solution to both unbounded and bounded distinct aggregation on data > stream so that it can easily be extended to support these use cases. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-8688) Enable distinct aggregation for data stream on Table/SQL API
[ https://issues.apache.org/jira/browse/FLINK-8688?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rong Rong updated FLINK-8688: - Description: Distinct aggregation is not currently supported on data stream with Table/SQL API. This is an umbrella task for enabling distinct aggregation in various use cases. Discussion doc can be found here: https://docs.google.com/document/d/1zj6OA-K2hi7ah8Fo-xTQB-mVmYfm6LsN2_NHgTCVmJI/edit?usp=sharing Distinct aggregation is a very important feature in SQL processing and there are many JIRAs currently open with various use cases. The goal is to create one solution to both unbounded and bounded distinct aggregation on data stream so that it can easily be extended to support these use cases. was: Distinct aggregation is not currently supported on data stream with Table/SQL API. This is an umbrella task for enabling distinct aggregation in various use cases. Discussion doc can be found [ here | [https://docs.google.com/document/d/1zj6OA-K2hi7ah8Fo-xTQB-mVmYfm6LsN2_NHgTCVmJI/edit?usp=sharing] ] Distinct aggregation is a very important feature in SQL processing and there are many JIRAs currently open with various use cases. The goal is to create one solution to both unbounded and bounded distinct aggregation on data stream so that it can easily be extended to support these use cases. > Enable distinct aggregation for data stream on Table/SQL API > > > Key: FLINK-8688 > URL: https://issues.apache.org/jira/browse/FLINK-8688 > Project: Flink > Issue Type: Improvement > Components: Table API SQL >Reporter: Rong Rong >Assignee: Rong Rong >Priority: Major > > Distinct aggregation is not currently supported on data stream with Table/SQL > API. This is an umbrella task for enabling distinct aggregation in various > use cases. > Discussion doc can be found here: > https://docs.google.com/document/d/1zj6OA-K2hi7ah8Fo-xTQB-mVmYfm6LsN2_NHgTCVmJI/edit?usp=sharing > > Distinct aggregation is a very important feature in SQL processing and there > are many JIRAs currently open with various use cases. The goal is to create > one solution to both unbounded and bounded distinct aggregation on data > stream so that it can easily be extended to support these use cases. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-8688) Enable distinct aggregation for data stream on Table/SQL API
[ https://issues.apache.org/jira/browse/FLINK-8688?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rong Rong updated FLINK-8688: - Description: Distinct aggregation is not currently supported on data stream with Table/SQL API. This is an umbrella task for enabling distinct aggregation in various use cases. Discussion doc can be found [ here | [https://docs.google.com/document/d/1zj6OA-K2hi7ah8Fo-xTQB-mVmYfm6LsN2_NHgTCVmJI/edit?usp=sharing] ] Distinct aggregation is a very important feature in SQL processing and there are many JIRAs currently open with various use cases. The goal is to create one solution to both unbounded and bounded distinct aggregation on data stream so that it can easily be extended to support these use cases. was: Distinct aggregation is not currently supported on data stream with Table/SQL API. This is an umbrella task for enabling distinct aggregation in various use cases. Discussion doc can be found [here|[https://docs.google.com/document/d/1zj6OA-K2hi7ah8Fo-xTQB-mVmYfm6LsN2_NHgTCVmJI/edit?usp=sharing]] Distinct aggregation is a very important feature in SQL processing and there are many JIRAs currently open with various use cases. The goal is to create one solution to both unbounded and bounded distinct aggregation on data stream so that it can easily be extended to support these use cases. > Enable distinct aggregation for data stream on Table/SQL API > > > Key: FLINK-8688 > URL: https://issues.apache.org/jira/browse/FLINK-8688 > Project: Flink > Issue Type: Improvement > Components: Table API SQL >Reporter: Rong Rong >Assignee: Rong Rong >Priority: Major > > Distinct aggregation is not currently supported on data stream with Table/SQL > API. This is an umbrella task for enabling distinct aggregation in various > use cases. > Discussion doc can be found [ here | > [https://docs.google.com/document/d/1zj6OA-K2hi7ah8Fo-xTQB-mVmYfm6LsN2_NHgTCVmJI/edit?usp=sharing] > ] > > Distinct aggregation is a very important feature in SQL processing and there > are many JIRAs currently open with various use cases. The goal is to create > one solution to both unbounded and bounded distinct aggregation on data > stream so that it can easily be extended to support these use cases. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (FLINK-8688) Enable distinct aggregation for data stream on Table/SQL API
Rong Rong created FLINK-8688: Summary: Enable distinct aggregation for data stream on Table/SQL API Key: FLINK-8688 URL: https://issues.apache.org/jira/browse/FLINK-8688 Project: Flink Issue Type: Improvement Components: Table API SQL Reporter: Rong Rong Assignee: Rong Rong Distinct aggregation is not currently supported on data stream with Table/SQL API. This is an umbrella task for enabling distinct aggregation in various use cases. Discussion doc can be found [here|[https://docs.google.com/document/d/1zj6OA-K2hi7ah8Fo-xTQB-mVmYfm6LsN2_NHgTCVmJI/edit?usp=sharing]] Distinct aggregation is a very important feature in SQL processing and there are many JIRAs currently open with various use cases. The goal is to create one solution to both unbounded and bounded distinct aggregation on data stream so that it can easily be extended to support these use cases. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-8684) Rework MesosTaskManagerParameters#MESOS_RM_TASKS_SLOTS
[ https://issues.apache.org/jira/browse/FLINK-8684?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368557#comment-16368557 ] ASF GitHub Bot commented on FLINK-8684: --- Github user zhangminglei closed the pull request at: https://github.com/apache/flink/pull/5519 > Rework MesosTaskManagerParameters#MESOS_RM_TASKS_SLOTS > -- > > Key: FLINK-8684 > URL: https://issues.apache.org/jira/browse/FLINK-8684 > Project: Flink > Issue Type: Improvement > Components: Documentation, Mesos >Affects Versions: 1.5.0 >Reporter: Chesnay Schepler >Priority: Major > Fix For: 1.5.0 > > > Currently, {{MesosTaskManagerParameters#MESOS_RM_TASKS_SLOTS}} mimics > {{TaskManagerOptions#NUM_TASK_SLOTS}}: > {code:java} > public static final ConfigOption MESOS_RM_TASKS_SLOTS = > key(ConfigConstants.TASK_MANAGER_NUM_TASK_SLOTS) > .defaultValue(1); > public static final ConfigOption NUM_TASK_SLOTS = > key("taskmanager.numberOfTaskSlots") > .defaultValue(1) > .withDescription("..."); > {code} > This pattern is problematic as this creates 2 documentation entries for > {{taskmanager.numberOfTaskSlots}} with different descriptions, and opens the > potential for different defaults. Ultimately this causes the documentation to > become ambiguous. > I thus propose to either outright remove this option or turn it into an > actual alias: > {code:java} > public static final ConfigOption MESOS_RM_TASKS_SLOTS = > TaskManagerOptions.NUM_TASK_SLOTS; > {code} > As a side-effect of FLINK-8683 we can ensure that no differing config options > exist for a given key. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5519: [FLINK-8684] [config] Rework MesosTaskManagerParam...
Github user zhangminglei closed the pull request at: https://github.com/apache/flink/pull/5519 ---
[jira] [Commented] (FLINK-8599) Improve the failure behavior of the FileInputFormat for bad files
[ https://issues.apache.org/jira/browse/FLINK-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368547#comment-16368547 ] ASF GitHub Bot commented on FLINK-8599: --- GitHub user ChengzhiZhao opened a pull request: https://github.com/apache/flink/pull/5520 [FLINK-8599] Improve the failure behavior of the FileInputFormat for bad files ## What is the purpose of the change This pull request is intent to improve the failure behavior of the FileInputFormat, currently if a bad file (for example, a different schema been dropped in this folder) came to the path and flink will do several retries. However, since the file path persist in the checkpoint, when people tried to resume from external checkpoint, it threw the following error on no file been found and the process cannot move forward. `java.io.IOException: Error opening the Input Split s3a://myfile [0,904]: No such file or directory: s3a://myfile` The change is to check if the path exist before open the file, if error occurs and bad file removed, flink should resume the process and continue. ## Brief change log - *Add a file exist check before open the file * ## Verifying this change - *Manually verified the change by introduce a bad file while continuously monitoring the folder, after remove the bad file, the process continued.* ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: (no) - The S3 file system connector: (no) ## Documentation - Does this pull request introduce a new feature? (no) - If yes, how is the feature documented? (not applicable) You can merge this pull request into a Git repository by running: $ git pull https://github.com/ChengzhiZhao/flink Improve_failure_behavior_ContinuousFileReaderOperator Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5520.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5520 commit e1605306d5f4a7c7c52eb1e1f3d213ea9872c71b Author: Chengzhi ZhaoDate: 2018-02-16T16:03:01Z [FLINK-8599] Improve the failure behavior of the ContinuousFileReaderOperator commit d8074e5141e0e6dc9f66e1fab8275cec3803904f Author: Chengzhi Zhao Date: 2018-02-17T21:04:59Z [FLINK-8599] break while loop commit 27c031466e1ea2d31cace5cd6a7f33f0d4c896c6 Author: Chengzhi Zhao Date: 2018-02-18T03:14:05Z [FLINK-8599] Check file exists in FileInputFormat commit a087587d12c5118f7ba8e55692dae6f3e95ef50b Author: Chengzhi Zhao Date: 2018-02-18T04:13:50Z [FLINK-8599] Add debug > Improve the failure behavior of the FileInputFormat for bad files > - > > Key: FLINK-8599 > URL: https://issues.apache.org/jira/browse/FLINK-8599 > Project: Flink > Issue Type: New Feature > Components: DataStream API >Affects Versions: 1.4.0, 1.3.2 >Reporter: Chengzhi Zhao >Priority: Major > > So we have a s3 path that flink is monitoring that path to see new files > available. > {code:java} > val avroInputStream_activity = env.readFile(format, path, > FileProcessingMode.PROCESS_CONTINUOUSLY, 1) {code} > > I am doing both internal and external check pointing and let's say there is a > bad file (for example, a different schema been dropped in this folder) came > to the path and flink will do several retries. I want to take those bad files > and let the process continue. However, since the file path persist in the > checkpoint, when I try to resume from external checkpoint, it threw the > following error on no file been found. > > {code:java} > java.io.IOException: Error opening the Input Split s3a://myfile [0,904]: No > such file or directory: s3a://myfile{code} > > As [~fhue...@gmail.com] suggested, we could check if a path exists and before > trying to read a file and ignore the input split instead of throwing an > exception and causing a failure. > > Also, I am thinking about add an error output for bad files as an option to > users. So if there is any bad files exist we could move them in a separated > path and do further analysis. > > Not sure how people feel about it, but I'd like to contribute on it if people >
[GitHub] flink pull request #5520: [FLINK-8599] Improve the failure behavior of the F...
GitHub user ChengzhiZhao opened a pull request: https://github.com/apache/flink/pull/5520 [FLINK-8599] Improve the failure behavior of the FileInputFormat for bad files ## What is the purpose of the change This pull request is intent to improve the failure behavior of the FileInputFormat, currently if a bad file (for example, a different schema been dropped in this folder) came to the path and flink will do several retries. However, since the file path persist in the checkpoint, when people tried to resume from external checkpoint, it threw the following error on no file been found and the process cannot move forward. `java.io.IOException: Error opening the Input Split s3a://myfile [0,904]: No such file or directory: s3a://myfile` The change is to check if the path exist before open the file, if error occurs and bad file removed, flink should resume the process and continue. ## Brief change log - *Add a file exist check before open the file * ## Verifying this change - *Manually verified the change by introduce a bad file while continuously monitoring the folder, after remove the bad file, the process continued.* ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: (no) - The S3 file system connector: (no) ## Documentation - Does this pull request introduce a new feature? (no) - If yes, how is the feature documented? (not applicable) You can merge this pull request into a Git repository by running: $ git pull https://github.com/ChengzhiZhao/flink Improve_failure_behavior_ContinuousFileReaderOperator Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5520.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5520 commit e1605306d5f4a7c7c52eb1e1f3d213ea9872c71b Author: Chengzhi ZhaoDate: 2018-02-16T16:03:01Z [FLINK-8599] Improve the failure behavior of the ContinuousFileReaderOperator commit d8074e5141e0e6dc9f66e1fab8275cec3803904f Author: Chengzhi Zhao Date: 2018-02-17T21:04:59Z [FLINK-8599] break while loop commit 27c031466e1ea2d31cace5cd6a7f33f0d4c896c6 Author: Chengzhi Zhao Date: 2018-02-18T03:14:05Z [FLINK-8599] Check file exists in FileInputFormat commit a087587d12c5118f7ba8e55692dae6f3e95ef50b Author: Chengzhi Zhao Date: 2018-02-18T04:13:50Z [FLINK-8599] Add debug ---
[jira] [Updated] (FLINK-8599) Improve the failure behavior of the FileInputFormat for bad files
[ https://issues.apache.org/jira/browse/FLINK-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chengzhi Zhao updated FLINK-8599: - Summary: Improve the failure behavior of the FileInputFormat for bad files (was: Improve the failure behavior of the ContinuousFileReaderOperator for bad files) > Improve the failure behavior of the FileInputFormat for bad files > - > > Key: FLINK-8599 > URL: https://issues.apache.org/jira/browse/FLINK-8599 > Project: Flink > Issue Type: New Feature > Components: DataStream API >Affects Versions: 1.4.0, 1.3.2 >Reporter: Chengzhi Zhao >Priority: Major > > So we have a s3 path that flink is monitoring that path to see new files > available. > {code:java} > val avroInputStream_activity = env.readFile(format, path, > FileProcessingMode.PROCESS_CONTINUOUSLY, 1) {code} > > I am doing both internal and external check pointing and let's say there is a > bad file (for example, a different schema been dropped in this folder) came > to the path and flink will do several retries. I want to take those bad files > and let the process continue. However, since the file path persist in the > checkpoint, when I try to resume from external checkpoint, it threw the > following error on no file been found. > > {code:java} > java.io.IOException: Error opening the Input Split s3a://myfile [0,904]: No > such file or directory: s3a://myfile{code} > > As [~fhue...@gmail.com] suggested, we could check if a path exists and before > trying to read a file and ignore the input split instead of throwing an > exception and causing a failure. > > Also, I am thinking about add an error output for bad files as an option to > users. So if there is any bad files exist we could move them in a separated > path and do further analysis. > > Not sure how people feel about it, but I'd like to contribute on it if people > think this can be an improvement. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-8546) Respect savepoint settings and recover from latest checkpoint in Flip-6
[ https://issues.apache.org/jira/browse/FLINK-8546?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368494#comment-16368494 ] ASF GitHub Bot commented on FLINK-8546: --- Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5444 > Respect savepoint settings and recover from latest checkpoint in Flip-6 > --- > > Key: FLINK-8546 > URL: https://issues.apache.org/jira/browse/FLINK-8546 > Project: Flink > Issue Type: Improvement > Components: JobManager >Affects Versions: 1.5.0 >Reporter: Till Rohrmann >Assignee: Till Rohrmann >Priority: Major > Labels: flip-6 > Fix For: 1.5.0 > > > The {{JobMaster}} should respect savepoints and recover from the latest > checkpoint if possible. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-8614) Enable Flip-6 per default
[ https://issues.apache.org/jira/browse/FLINK-8614?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368493#comment-16368493 ] ASF GitHub Bot commented on FLINK-8614: --- Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5437 > Enable Flip-6 per default > - > > Key: FLINK-8614 > URL: https://issues.apache.org/jira/browse/FLINK-8614 > Project: Flink > Issue Type: Sub-task > Components: Distributed Coordination >Affects Versions: 1.5.0 >Reporter: Till Rohrmann >Assignee: Till Rohrmann >Priority: Blocker > Labels: flip-6 > Fix For: 1.5.0 > > > After adding the FLINK-8471, the next step is to enable Flip-6 per default by > setting the configuration switch to {{flip6}}. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-8673) Don't let JobManagerRunner shut down itself
[ https://issues.apache.org/jira/browse/FLINK-8673?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368492#comment-16368492 ] ASF GitHub Bot commented on FLINK-8673: --- Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5510 > Don't let JobManagerRunner shut down itself > --- > > Key: FLINK-8673 > URL: https://issues.apache.org/jira/browse/FLINK-8673 > Project: Flink > Issue Type: Improvement > Components: Distributed Coordination >Affects Versions: 1.5.0 >Reporter: Till Rohrmann >Assignee: Till Rohrmann >Priority: Major > Labels: flip-6 > Fix For: 1.5.0 > > > Currently, the {{JobManagerRunner}} is allowed to shut down itself in case of > a job completion. This, however, can cause problems when the {{Dispatcher}} > receives a request for a {{JobMaster}}. If the {{Dispatcher}} is not told > about the shut down of the {{JobMaster}} then it might still try to send > requests to it. This will lead to time outs. > It would be better to simply let the {{JobManagerRunner}} not shut down > itself and defer it to the owner (the {{Dispatcher}}). We can do this by > listening on the {{JobManagerRunner#resultFuture}} which is completed by the > {{JobManagerRunner}} in case of a successful job completion or a failure. > That way we could also get rid of the {{OnCompletionActions}} and the > {{FatalErrorHandler}}. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-8653) Remove slot request timeout from SlotPool
[ https://issues.apache.org/jira/browse/FLINK-8653?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368495#comment-16368495 ] ASF GitHub Bot commented on FLINK-8653: --- Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5483 > Remove slot request timeout from SlotPool > - > > Key: FLINK-8653 > URL: https://issues.apache.org/jira/browse/FLINK-8653 > Project: Flink > Issue Type: Improvement > Components: Distributed Coordination >Affects Versions: 1.5.0 >Reporter: Till Rohrmann >Assignee: Till Rohrmann >Priority: Major > Labels: flip-6 > Fix For: 1.5.0 > > > After addressing FLINK-8643, we can further simplify the {{SlotPool}} by > replacing the internal slot request timeout by the timeout given to > {{SlotPool#allocateSlot}}. Since this request will timeout on the > {{ProviderAndOwner}} side anyway, we should do the same on the {{SlotPool}} > side. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5437: [FLINK-8614] [flip6] Activate Flip-6 mode per defa...
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5437 ---
[GitHub] flink pull request #5483: [FLINK-8653] [flip6] Remove internal slot request ...
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5483 ---
[jira] [Closed] (FLINK-8673) Don't let JobManagerRunner shut down itself
[ https://issues.apache.org/jira/browse/FLINK-8673?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Till Rohrmann closed FLINK-8673. Resolution: Fixed Fixed via 075f5b6930b86e110a32e290229b9800be72a3a7 > Don't let JobManagerRunner shut down itself > --- > > Key: FLINK-8673 > URL: https://issues.apache.org/jira/browse/FLINK-8673 > Project: Flink > Issue Type: Improvement > Components: Distributed Coordination >Affects Versions: 1.5.0 >Reporter: Till Rohrmann >Assignee: Till Rohrmann >Priority: Major > Labels: flip-6 > Fix For: 1.5.0 > > > Currently, the {{JobManagerRunner}} is allowed to shut down itself in case of > a job completion. This, however, can cause problems when the {{Dispatcher}} > receives a request for a {{JobMaster}}. If the {{Dispatcher}} is not told > about the shut down of the {{JobMaster}} then it might still try to send > requests to it. This will lead to time outs. > It would be better to simply let the {{JobManagerRunner}} not shut down > itself and defer it to the owner (the {{Dispatcher}}). We can do this by > listening on the {{JobManagerRunner#resultFuture}} which is completed by the > {{JobManagerRunner}} in case of a successful job completion or a failure. > That way we could also get rid of the {{OnCompletionActions}} and the > {{FatalErrorHandler}}. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-8614) Enable Flip-6 per default
[ https://issues.apache.org/jira/browse/FLINK-8614?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Till Rohrmann closed FLINK-8614. Resolution: Fixed Fixed via ab8316f31a4de0648ab3ffa9f19fcd419aaa3bb9 > Enable Flip-6 per default > - > > Key: FLINK-8614 > URL: https://issues.apache.org/jira/browse/FLINK-8614 > Project: Flink > Issue Type: Sub-task > Components: Distributed Coordination >Affects Versions: 1.5.0 >Reporter: Till Rohrmann >Assignee: Till Rohrmann >Priority: Blocker > Labels: flip-6 > Fix For: 1.5.0 > > > After adding the FLINK-8471, the next step is to enable Flip-6 per default by > setting the configuration switch to {{flip6}}. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5510: [FLINK-8673] [flip6] Use JobManagerRunner#resultFu...
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5510 ---
[GitHub] flink pull request #5444: [FLINK-8546] [flip6] Respect savepoints and restor...
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5444 ---
[jira] [Closed] (FLINK-8546) Respect savepoint settings and recover from latest checkpoint in Flip-6
[ https://issues.apache.org/jira/browse/FLINK-8546?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Till Rohrmann closed FLINK-8546. Resolution: Fixed Fixed via 3a4e89643d7d7642dde9b5644491f261d4d545bd > Respect savepoint settings and recover from latest checkpoint in Flip-6 > --- > > Key: FLINK-8546 > URL: https://issues.apache.org/jira/browse/FLINK-8546 > Project: Flink > Issue Type: Improvement > Components: JobManager >Affects Versions: 1.5.0 >Reporter: Till Rohrmann >Assignee: Till Rohrmann >Priority: Major > Labels: flip-6 > Fix For: 1.5.0 > > > The {{JobMaster}} should respect savepoints and recover from the latest > checkpoint if possible. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-8653) Remove slot request timeout from SlotPool
[ https://issues.apache.org/jira/browse/FLINK-8653?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Till Rohrmann closed FLINK-8653. Resolution: Fixed Fixed via d8a8866973f7e0463047963e6b242cdc2cb82fec > Remove slot request timeout from SlotPool > - > > Key: FLINK-8653 > URL: https://issues.apache.org/jira/browse/FLINK-8653 > Project: Flink > Issue Type: Improvement > Components: Distributed Coordination >Affects Versions: 1.5.0 >Reporter: Till Rohrmann >Assignee: Till Rohrmann >Priority: Major > Labels: flip-6 > Fix For: 1.5.0 > > > After addressing FLINK-8643, we can further simplify the {{SlotPool}} by > replacing the internal slot request timeout by the timeout given to > {{SlotPool#allocateSlot}}. Since this request will timeout on the > {{ProviderAndOwner}} side anyway, we should do the same on the {{SlotPool}} > side. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-8684) Rework MesosTaskManagerParameters#MESOS_RM_TASKS_SLOTS
[ https://issues.apache.org/jira/browse/FLINK-8684?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368482#comment-16368482 ] ASF GitHub Bot commented on FLINK-8684: --- Github user zentol commented on the issue: https://github.com/apache/flink/pull/5519 Correct, I've included a fix for FLINK-8684 in the [PR](https://github.com/apache/flink/pull/5515) for FLINK-8683. > Rework MesosTaskManagerParameters#MESOS_RM_TASKS_SLOTS > -- > > Key: FLINK-8684 > URL: https://issues.apache.org/jira/browse/FLINK-8684 > Project: Flink > Issue Type: Improvement > Components: Documentation, Mesos >Affects Versions: 1.5.0 >Reporter: Chesnay Schepler >Priority: Major > Fix For: 1.5.0 > > > Currently, {{MesosTaskManagerParameters#MESOS_RM_TASKS_SLOTS}} mimics > {{TaskManagerOptions#NUM_TASK_SLOTS}}: > {code:java} > public static final ConfigOption MESOS_RM_TASKS_SLOTS = > key(ConfigConstants.TASK_MANAGER_NUM_TASK_SLOTS) > .defaultValue(1); > public static final ConfigOption NUM_TASK_SLOTS = > key("taskmanager.numberOfTaskSlots") > .defaultValue(1) > .withDescription("..."); > {code} > This pattern is problematic as this creates 2 documentation entries for > {{taskmanager.numberOfTaskSlots}} with different descriptions, and opens the > potential for different defaults. Ultimately this causes the documentation to > become ambiguous. > I thus propose to either outright remove this option or turn it into an > actual alias: > {code:java} > public static final ConfigOption MESOS_RM_TASKS_SLOTS = > TaskManagerOptions.NUM_TASK_SLOTS; > {code} > As a side-effect of FLINK-8683 we can ensure that no differing config options > exist for a given key. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink issue #5519: [FLINK-8684] [config] Rework MesosTaskManagerParameters#M...
Github user zentol commented on the issue: https://github.com/apache/flink/pull/5519 Correct, I've included a fix for FLINK-8684 in the [PR](https://github.com/apache/flink/pull/5515) for FLINK-8683. ---
[jira] [Commented] (FLINK-8274) Fix Java 64K method compiling limitation for CommonCalc
[ https://issues.apache.org/jira/browse/FLINK-8274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16368465#comment-16368465 ] ASF GitHub Bot commented on FLINK-8274: --- Github user jahandarm commented on a diff in the pull request: https://github.com/apache/flink/pull/5174#discussion_r168942867 --- Diff: flink-libraries/flink-table/src/test/scala/org/apache/flink/table/runtime/stream/table/CalcWithSplitCodeGenITCase.scala --- @@ -0,0 +1,385 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.table.runtime.stream.table + +import org.apache.flink.api.scala._ +import org.apache.flink.streaming.api.scala.StreamExecutionEnvironment +import org.apache.flink.streaming.util.StreamingMultipleProgramsTestBase --- End diff -- I think this class doesn't exist. > Fix Java 64K method compiling limitation for CommonCalc > --- > > Key: FLINK-8274 > URL: https://issues.apache.org/jira/browse/FLINK-8274 > Project: Flink > Issue Type: Improvement > Components: Table API SQL >Affects Versions: 1.5.0 >Reporter: Ruidong Li >Assignee: Ruidong Li >Priority: Major > > For complex SQL Queries, the generated code for {code}DataStreamCalc{code}, > {code}DataSetCalc{code} may exceed Java's method length limitation 64kb. > > This issue will split long method to several sub method calls. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5174: [FLINK-8274][TableAPI & SQL] Fix Java 64K method c...
Github user jahandarm commented on a diff in the pull request: https://github.com/apache/flink/pull/5174#discussion_r168942867 --- Diff: flink-libraries/flink-table/src/test/scala/org/apache/flink/table/runtime/stream/table/CalcWithSplitCodeGenITCase.scala --- @@ -0,0 +1,385 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.table.runtime.stream.table + +import org.apache.flink.api.scala._ +import org.apache.flink.streaming.api.scala.StreamExecutionEnvironment +import org.apache.flink.streaming.util.StreamingMultipleProgramsTestBase --- End diff -- I think this class doesn't exist. ---