[jira] [Closed] (FLINK-7566) if there's only one checkpointing metadata file in , `flink run -s ` should successfully resume from that metadata file
[ https://issues.apache.org/jira/browse/FLINK-7566?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bowen Li closed FLINK-7566. --- Resolution: Won't Fix > if there's only one checkpointing metadata file in , `flink run -s > ` should successfully resume from that metadata file > -- > > Key: FLINK-7566 > URL: https://issues.apache.org/jira/browse/FLINK-7566 > Project: Flink > Issue Type: Improvement > Components: State Backends, Checkpointing >Affects Versions: 1.3.2 >Reporter: Bowen Li >Assignee: Bowen Li >Priority: Major > > Currently, if we want to start a Flink job from a checkpointing file, we have > to run `flink run -s /checkpoint_metadata-x` by explicitly > specifying the checkpoint metadata file name 'checkpoint_metadata-x'. > Since metadata file name always changes, it's not easy to programmatically > restart a failed Flink job. The error from jobmanager.log looks like: > {code:java} > 2017-08-30 07:25:04,907 INFO > org.apache.flink.runtime.executiongraph.ExecutionGraph- Job > (22defcf962ff2ac2e7fe99354f5ab168) switched from state FAILING to FAILED. > org.apache.flink.runtime.execution.SuppressRestartsException: Unrecoverable > failure. This suppresses job restarts. Please check the stack trace for the > root cause. > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$org$apache$flink$runtime$jobmanager$JobManager$$submitJob$1.apply$mcV$sp(JobManager.scala:1396) > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$org$apache$flink$runtime$jobmanager$JobManager$$submitJob$1.apply(JobManager.scala:1372) > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$org$apache$flink$runtime$jobmanager$JobManager$$submitJob$1.apply(JobManager.scala:1372) > at > scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24) > at > scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24) > at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40) > at > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397) > at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > at > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > Caused by: java.io.IOException: Cannot find meta data file in directory > s3:///checkpoints. Please try to load the savepoint directly from the > meta data file instead of the directory. > at > org.apache.flink.runtime.checkpoint.savepoint.SavepointStore.loadSavepointWithHandle(SavepointStore.java:262) > at > org.apache.flink.runtime.checkpoint.savepoint.SavepointLoader.loadAndValidateSavepoint(SavepointLoader.java:69) > at > org.apache.flink.runtime.checkpoint.CheckpointCoordinator.restoreSavepoint(CheckpointCoordinator.java:1140) > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$org$apache$flink$runtime$jobmanager$JobManager$$submitJob$1.apply$mcV$sp(JobManager.scala:1386) > ... 10 more > {code} > What I want is like this: users should be able to start a Flink job by > running `flink run -s ` if there's only one checkpointing metadata file > in . If there's none or more than 1 metadata file, the command can fail > like it is right now. This way, we can programmatically restart a failed > Flink job by hardcoding . > To achieve that, I think there're two appraches we can do: > 1) modify {{CheckpointCoordinator.restoreSavepoint}} to check how many > metadata files are in > 2) add another commandline option like '-sd' / '--savepointdirectory' to > explicitly load a dir -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-8837) add @Experimental annotation and properly annotate some classes
[ https://issues.apache.org/jira/browse/FLINK-8837?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bowen Li updated FLINK-8837: Affects Version/s: 1.5.0 > add @Experimental annotation and properly annotate some classes > > > Key: FLINK-8837 > URL: https://issues.apache.org/jira/browse/FLINK-8837 > Project: Flink > Issue Type: Bug > Components: Streaming >Affects Versions: 1.5.0 >Reporter: Stephan Ewen >Assignee: Bowen Li >Priority: Blocker > Fix For: 1.5.0 > > > The class {{DataStreamUtils}} came from 'flink-contrib' and now accidentally > moved to the fully supported API packages. It should be in package > 'experimental' to properly communicate that it is not guaranteed to be API > stable. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9127) Filesystem State Backend logged incorrectly
[ https://issues.apache.org/jira/browse/FLINK-9127?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16426530#comment-16426530 ] ASF GitHub Bot commented on FLINK-9127: --- Github user bowenli86 commented on the issue: https://github.com/apache/flink/pull/5810 The original logging is correct - filesystem state backend is actually memory state backend + filesystem checkpointing. No need to change the logging. > Filesystem State Backend logged incorrectly > --- > > Key: FLINK-9127 > URL: https://issues.apache.org/jira/browse/FLINK-9127 > Project: Flink > Issue Type: Bug > Components: Core >Affects Versions: 1.3.2, 1.4.2 >Reporter: Scott Kidder >Priority: Trivial > > When using a filesystem backend, the > '[StateBackendLoader|https://github.com/apache/flink/blob/1f9c2d9740ffea2b59b8f5f3da287a0dc890ddbf/flink-runtime/src/main/java/org/apache/flink/runtime/state/StateBackendLoader.java#L123]' > class produces a log message stating: "State backend is set to heap memory". > Example: > {{2018-04-04 00:45:49,591 INFO > org.apache.flink.streaming.runtime.tasks.StreamTask - State backend > is set to heap memory (checkpoints to filesystem > "hdfs://hdfs:8020/flink/checkpoints")}} > It looks like this resulted from some copy-pasta of the previous > case-statement that matches on the memory backend. This bug is also present > in earlier releases (1.3.2, 1.4.0) of Flink in the 'AbstractStateBackend' > class. > This log statement should be corrected to indicate that a filesystem backend > is in use. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink issue #5810: [FLINK-9127] [Core] Filesystem State Backend logged incor...
Github user bowenli86 commented on the issue: https://github.com/apache/flink/pull/5810 The original logging is correct - filesystem state backend is actually memory state backend + filesystem checkpointing. No need to change the logging. ---
[jira] [Commented] (FLINK-9015) Upgrade Calcite dependency to 1.17
[ https://issues.apache.org/jira/browse/FLINK-9015?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16426458#comment-16426458 ] Shuyi Chen commented on FLINK-9015: --- Duplicate of [FLINK-9134|https://issues.apache.org/jira/browse/FLINK-9134]. I'll close this one. > Upgrade Calcite dependency to 1.17 > -- > > Key: FLINK-9015 > URL: https://issues.apache.org/jira/browse/FLINK-9015 > Project: Flink > Issue Type: Task >Reporter: Shuyi Chen >Assignee: Shuyi Chen >Priority: Major > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-9015) Upgrade Calcite dependency to 1.17
[ https://issues.apache.org/jira/browse/FLINK-9015?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shuyi Chen closed FLINK-9015. - Resolution: Duplicate > Upgrade Calcite dependency to 1.17 > -- > > Key: FLINK-9015 > URL: https://issues.apache.org/jira/browse/FLINK-9015 > Project: Flink > Issue Type: Task >Reporter: Shuyi Chen >Assignee: Shuyi Chen >Priority: Major > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9134) Update Calcite dependency to 1.17
[ https://issues.apache.org/jira/browse/FLINK-9134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16426455#comment-16426455 ] Shuyi Chen commented on FLINK-9134: --- Actually, let's close [FLINK-9015|https://issues.apache.org/jira/browse/FLINK-9015]. > Update Calcite dependency to 1.17 > - > > Key: FLINK-9134 > URL: https://issues.apache.org/jira/browse/FLINK-9134 > Project: Flink > Issue Type: Improvement > Components: Table API & SQL >Reporter: Timo Walther >Assignee: Shuyi Chen >Priority: Major > > This is an umbrella issue for tasks that need to be performed when upgrading > to Calcite 1.17 once it is released. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Reopened] (FLINK-9134) Update Calcite dependency to 1.17
[ https://issues.apache.org/jira/browse/FLINK-9134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shuyi Chen reopened FLINK-9134: --- Assignee: Shuyi Chen > Update Calcite dependency to 1.17 > - > > Key: FLINK-9134 > URL: https://issues.apache.org/jira/browse/FLINK-9134 > Project: Flink > Issue Type: Improvement > Components: Table API & SQL >Reporter: Timo Walther >Assignee: Shuyi Chen >Priority: Major > > This is an umbrella issue for tasks that need to be performed when upgrading > to Calcite 1.17 once it is released. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-9134) Update Calcite dependency to 1.17
[ https://issues.apache.org/jira/browse/FLINK-9134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shuyi Chen closed FLINK-9134. - Resolution: Duplicate > Update Calcite dependency to 1.17 > - > > Key: FLINK-9134 > URL: https://issues.apache.org/jira/browse/FLINK-9134 > Project: Flink > Issue Type: Improvement > Components: Table API & SQL >Reporter: Timo Walther >Priority: Major > > This is an umbrella issue for tasks that need to be performed when upgrading > to Calcite 1.17 once it is released. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-9089) Upgrade Orc dependency to 1.4.3
[ https://issues.apache.org/jira/browse/FLINK-9089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ted Yu updated FLINK-9089: -- Description: Currently flink-orc uses Orc 1.4.1 release. This issue upgrades to Orc 1.4.3 was: Currently flink-orc uses Orc 1.4.1 release. This issue upgrades to Orc 1.4.3 > Upgrade Orc dependency to 1.4.3 > --- > > Key: FLINK-9089 > URL: https://issues.apache.org/jira/browse/FLINK-9089 > Project: Flink > Issue Type: Improvement >Reporter: Ted Yu >Assignee: vinoyang >Priority: Minor > > Currently flink-orc uses Orc 1.4.1 release. > This issue upgrades to Orc 1.4.3 -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5818: sql() is deprecated.
GitHub user mayyamus opened a pull request: https://github.com/apache/flink/pull/5818 sql() is deprecated. ## Brief change log - Replace the deprecatd method(sql()method) with sqlQuery() ## Verifying this change This change is a trivial rework / code cleanup without any test coverage. You can merge this pull request into a Git repository by running: $ git pull https://github.com/mayyamus/flink fix_table_docs Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5818.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5818 commit a80733769591f4f87fdf38794771604583c4d49b Author: mayyamus Date: 2018-04-05T01:54:11Z sql() is deprecated. ---
[jira] [Commented] (FLINK-8703) Migrate tests from LocalFlinkMiniCluster to MiniClusterResource
[ https://issues.apache.org/jira/browse/FLINK-8703?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16426402#comment-16426402 ] ASF GitHub Bot commented on FLINK-8703: --- Github user pnowojski commented on a diff in the pull request: https://github.com/apache/flink/pull/5669#discussion_r179332025 --- Diff: flink-connectors/flink-connector-kafka-base/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaConsumerTestBase.java --- @@ -1063,23 +1078,27 @@ public void runCancelingOnEmptyInputTest() throws Exception { final AtomicReference error = new AtomicReference<>(); + final StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); --- End diff -- As you wish, I can open a follow up since it's a trivial fixup. > Migrate tests from LocalFlinkMiniCluster to MiniClusterResource > --- > > Key: FLINK-8703 > URL: https://issues.apache.org/jira/browse/FLINK-8703 > Project: Flink > Issue Type: Sub-task > Components: Tests >Reporter: Aljoscha Krettek >Assignee: Chesnay Schepler >Priority: Blocker > Fix For: 1.5.0 > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5669: [FLINK-8703][tests] Port KafkaTestBase to MiniClus...
Github user pnowojski commented on a diff in the pull request: https://github.com/apache/flink/pull/5669#discussion_r179332025 --- Diff: flink-connectors/flink-connector-kafka-base/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaConsumerTestBase.java --- @@ -1063,23 +1078,27 @@ public void runCancelingOnEmptyInputTest() throws Exception { final AtomicReference error = new AtomicReference<>(); + final StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); --- End diff -- As you wish, I can open a follow up since it's a trivial fixup. ---
[jira] [Created] (FLINK-9139) Allow yarn.tags as a job parameter
Nikhil Simha created FLINK-9139: --- Summary: Allow yarn.tags as a job parameter Key: FLINK-9139 URL: https://issues.apache.org/jira/browse/FLINK-9139 Project: Flink Issue Type: Improvement Components: Cluster Management, Configuration Reporter: Nikhil Simha Currently `yarn.tags` is accepted as a parameter from flink-conf.yml. I want to be able to pass that as a job commandline argument like `yarnslots` `yarnqueue`. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-5153) Allow setting custom application tags for Flink on YARN
[ https://issues.apache.org/jira/browse/FLINK-5153?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16426359#comment-16426359 ] Nikhil Simha commented on FLINK-5153: - I want to be able to supply an application tag at job submission time - as a parameter like yarn.taskmanagerslots or yarn.name - instead of modifying the flink-conf. > Allow setting custom application tags for Flink on YARN > --- > > Key: FLINK-5153 > URL: https://issues.apache.org/jira/browse/FLINK-5153 > Project: Flink > Issue Type: Improvement > Components: YARN >Reporter: Robert Metzger >Assignee: Patrick Lucas >Priority: Major > Fix For: 1.3.0 > > > https://issues.apache.org/jira/browse/YARN-1399 added support in YARN to tag > applications. > We should introduce a configuration variable in Flink allowing users to > specify a comma-separated list of tags they want to assign to their Flink on > YARN applications. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Comment Edited] (FLINK-5153) Allow setting custom application tags for Flink on YARN
[ https://issues.apache.org/jira/browse/FLINK-5153?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16426359#comment-16426359 ] Nikhil Simha edited comment on FLINK-5153 at 4/5/18 12:17 AM: -- I want to be able to supply an application tag at job submission time - as a parameter like yarn.taskmanagerslots or yarn.name - instead of modifying the flink-conf. How do I do that? was (Author: nikhilsimha): I want to be able to supply an application tag at job submission time - as a parameter like yarn.taskmanagerslots or yarn.name - instead of modifying the flink-conf. > Allow setting custom application tags for Flink on YARN > --- > > Key: FLINK-5153 > URL: https://issues.apache.org/jira/browse/FLINK-5153 > Project: Flink > Issue Type: Improvement > Components: YARN >Reporter: Robert Metzger >Assignee: Patrick Lucas >Priority: Major > Fix For: 1.3.0 > > > https://issues.apache.org/jira/browse/YARN-1399 added support in YARN to tag > applications. > We should introduce a configuration variable in Flink allowing users to > specify a comma-separated list of tags they want to assign to their Flink on > YARN applications. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Comment Edited] (FLINK-6105) Properly handle InterruptedException in HadoopInputFormatBase
[ https://issues.apache.org/jira/browse/FLINK-6105?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16307281#comment-16307281 ] Ted Yu edited comment on FLINK-6105 at 4/4/18 11:18 PM: In flink-connectors/flink-connector-filesystem/src/main/java/org/apache/flink/streaming/connectors/fs/RollingSink.java : {code} try { Thread.sleep(500); } catch (InterruptedException e1) { // ignore it } {code} Interrupt status should be restored, or throw InterruptedIOException . was (Author: yuzhih...@gmail.com): In flink-connectors/flink-connector-filesystem/src/main/java/org/apache/flink/streaming/connectors/fs/RollingSink.java : {code} try { Thread.sleep(500); } catch (InterruptedException e1) { // ignore it } {code} Interrupt status should be restored, or throw InterruptedIOException . > Properly handle InterruptedException in HadoopInputFormatBase > - > > Key: FLINK-6105 > URL: https://issues.apache.org/jira/browse/FLINK-6105 > Project: Flink > Issue Type: Bug > Components: DataStream API >Reporter: Ted Yu >Assignee: mingleizhang >Priority: Major > > When catching InterruptedException, we should throw InterruptedIOException > instead of IOException. > The following example is from HadoopInputFormatBase : > {code} > try { > splits = this.mapreduceInputFormat.getSplits(jobContext); > } catch (InterruptedException e) { > throw new IOException("Could not get Splits.", e); > } > {code} > There may be other places where IOE is thrown. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-7151) FLINK SQL support create temporary function and table
[ https://issues.apache.org/jira/browse/FLINK-7151?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16426316#comment-16426316 ] Shuyi Chen commented on FLINK-7151: --- I don't have a concrete timeline, but will try to implement the table DDL before Flink 1.6 release. > FLINK SQL support create temporary function and table > - > > Key: FLINK-7151 > URL: https://issues.apache.org/jira/browse/FLINK-7151 > Project: Flink > Issue Type: New Feature > Components: Table API & SQL >Reporter: yuemeng >Assignee: yuemeng >Priority: Major > > Based on create temporary function and table.we can register a udf,udaf,udtf > use sql: > {code} > CREATE TEMPORARY function 'TOPK' AS > 'com..aggregate.udaf.distinctUdaf.topk.ITopKUDAF'; > INSERT INTO db_sink SELECT id, TOPK(price, 5, 'DESC') FROM kafka_source GROUP > BY id; > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-8554) Upgrade AWS SDK
[ https://issues.apache.org/jira/browse/FLINK-8554?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ted Yu updated FLINK-8554: -- Description: AWS SDK 1.11.271 fixes a lot of bugs. One of which would exhibit the following: {code} Caused by: java.lang.NullPointerException at com.amazonaws.metrics.AwsSdkMetrics.getRegion(AwsSdkMetrics.java:729) at com.amazonaws.metrics.MetricAdmin.getRegion(MetricAdmin.java:67) at sun.reflect.GeneratedMethodAccessor132.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) {code} was: AWS SDK 1.11.271 fixes a lot of bugs. One of which would exhibit the following: {code} Caused by: java.lang.NullPointerException at com.amazonaws.metrics.AwsSdkMetrics.getRegion(AwsSdkMetrics.java:729) at com.amazonaws.metrics.MetricAdmin.getRegion(MetricAdmin.java:67) at sun.reflect.GeneratedMethodAccessor132.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) {code} > Upgrade AWS SDK > --- > > Key: FLINK-8554 > URL: https://issues.apache.org/jira/browse/FLINK-8554 > Project: Flink > Issue Type: Improvement >Reporter: Ted Yu >Priority: Minor > > AWS SDK 1.11.271 fixes a lot of bugs. > One of which would exhibit the following: > {code} > Caused by: java.lang.NullPointerException > at com.amazonaws.metrics.AwsSdkMetrics.getRegion(AwsSdkMetrics.java:729) > at com.amazonaws.metrics.MetricAdmin.getRegion(MetricAdmin.java:67) > at sun.reflect.GeneratedMethodAccessor132.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9061) S3 checkpoint data not partitioned well -- causes errors and poor performance
[ https://issues.apache.org/jira/browse/FLINK-9061?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16426132#comment-16426132 ] Jamie Grier commented on FLINK-9061: Yup, sounds good to me :) > S3 checkpoint data not partitioned well -- causes errors and poor performance > - > > Key: FLINK-9061 > URL: https://issues.apache.org/jira/browse/FLINK-9061 > Project: Flink > Issue Type: Bug > Components: FileSystem, State Backends, Checkpointing >Affects Versions: 1.4.2 >Reporter: Jamie Grier >Priority: Critical > > I think we need to modify the way we write checkpoints to S3 for high-scale > jobs (those with many total tasks). The issue is that we are writing all the > checkpoint data under a common key prefix. This is the worst case scenario > for S3 performance since the key is used as a partition key. > > In the worst case checkpoints fail with a 500 status code coming back from S3 > and an internal error type of TooBusyException. > > One possible solution would be to add a hook in the Flink filesystem code > that allows me to "rewrite" paths. For example say I have the checkpoint > directory set to: > > s3://bucket/flink/checkpoints > > I would hook that and rewrite that path to: > > s3://bucket/[HASH]/flink/checkpoints, where HASH is the hash of the original > path > > This would distribute the checkpoint write load around the S3 cluster evenly. > > For reference: > https://aws.amazon.com/premiumsupport/knowledge-center/s3-bucket-performance-improve/ > > Any other people hit this issue? Any other ideas for solutions? This is a > pretty serious problem for people trying to checkpoint to S3. > > -Jamie > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Comment Edited] (FLINK-9061) S3 checkpoint data not partitioned well -- causes errors and poor performance
[ https://issues.apache.org/jira/browse/FLINK-9061?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16426124#comment-16426124 ] Steven Zhen Wu edited comment on FLINK-9061 at 4/4/18 8:24 PM: --- [~jgrier] Amazon doesn't want to reveal internal details, hence sometimes are pretty vague. My understanding is that prefix (before the random/entropy part) has to be fixed. Either way, latest proposal doesn't prevent user from setting the first a few chars as random/entropy part. I just think it is important to give user the full control on key names. was (Author: stevenz3wu): [~jgrier] Amazon doesn't want to reveal internal details, hence sometimes are pretty vague. Mu understanding is that prefix (before the random/entropy part) has to be fixed. Either way, latest proposal doesn't prevent user from setting the first a few chars as random/entropy part. I just think it is important to give user the full control on key names. > S3 checkpoint data not partitioned well -- causes errors and poor performance > - > > Key: FLINK-9061 > URL: https://issues.apache.org/jira/browse/FLINK-9061 > Project: Flink > Issue Type: Bug > Components: FileSystem, State Backends, Checkpointing >Affects Versions: 1.4.2 >Reporter: Jamie Grier >Priority: Critical > > I think we need to modify the way we write checkpoints to S3 for high-scale > jobs (those with many total tasks). The issue is that we are writing all the > checkpoint data under a common key prefix. This is the worst case scenario > for S3 performance since the key is used as a partition key. > > In the worst case checkpoints fail with a 500 status code coming back from S3 > and an internal error type of TooBusyException. > > One possible solution would be to add a hook in the Flink filesystem code > that allows me to "rewrite" paths. For example say I have the checkpoint > directory set to: > > s3://bucket/flink/checkpoints > > I would hook that and rewrite that path to: > > s3://bucket/[HASH]/flink/checkpoints, where HASH is the hash of the original > path > > This would distribute the checkpoint write load around the S3 cluster evenly. > > For reference: > https://aws.amazon.com/premiumsupport/knowledge-center/s3-bucket-performance-improve/ > > Any other people hit this issue? Any other ideas for solutions? This is a > pretty serious problem for people trying to checkpoint to S3. > > -Jamie > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Comment Edited] (FLINK-9061) S3 checkpoint data not partitioned well -- causes errors and poor performance
[ https://issues.apache.org/jira/browse/FLINK-9061?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16426124#comment-16426124 ] Steven Zhen Wu edited comment on FLINK-9061 at 4/4/18 8:20 PM: --- [~jgrier] Amazon doesn't want to reveal internal details, hence sometimes are pretty vague. Mu understanding is that prefix (before the random/entropy part) has to be fixed. Either way, latest proposal doesn't prevent user from setting the first a few chars as random/entropy part. I just think it is important to give user the full control on key names. was (Author: stevenz3wu): [~jgrier] Amazon doesn't want to reveal internal details, hence sometimes are pretty vague. Mu understanding is that prefix (before the random/entropy part) has to be fixed. Either way, latest proposal doesn't prevent user from setting the first a few chars as random/entropy part. > S3 checkpoint data not partitioned well -- causes errors and poor performance > - > > Key: FLINK-9061 > URL: https://issues.apache.org/jira/browse/FLINK-9061 > Project: Flink > Issue Type: Bug > Components: FileSystem, State Backends, Checkpointing >Affects Versions: 1.4.2 >Reporter: Jamie Grier >Priority: Critical > > I think we need to modify the way we write checkpoints to S3 for high-scale > jobs (those with many total tasks). The issue is that we are writing all the > checkpoint data under a common key prefix. This is the worst case scenario > for S3 performance since the key is used as a partition key. > > In the worst case checkpoints fail with a 500 status code coming back from S3 > and an internal error type of TooBusyException. > > One possible solution would be to add a hook in the Flink filesystem code > that allows me to "rewrite" paths. For example say I have the checkpoint > directory set to: > > s3://bucket/flink/checkpoints > > I would hook that and rewrite that path to: > > s3://bucket/[HASH]/flink/checkpoints, where HASH is the hash of the original > path > > This would distribute the checkpoint write load around the S3 cluster evenly. > > For reference: > https://aws.amazon.com/premiumsupport/knowledge-center/s3-bucket-performance-improve/ > > Any other people hit this issue? Any other ideas for solutions? This is a > pretty serious problem for people trying to checkpoint to S3. > > -Jamie > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9061) S3 checkpoint data not partitioned well -- causes errors and poor performance
[ https://issues.apache.org/jira/browse/FLINK-9061?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16426124#comment-16426124 ] Steven Zhen Wu commented on FLINK-9061: --- [~jgrier] Amazon doesn't want to reveal internal details, hence sometimes are pretty vague. Mu understanding is that prefix (before the random/entropy part) has to be fixed. Either way, latest proposal doesn't prevent user from setting the first a few chars as random/entropy part. > S3 checkpoint data not partitioned well -- causes errors and poor performance > - > > Key: FLINK-9061 > URL: https://issues.apache.org/jira/browse/FLINK-9061 > Project: Flink > Issue Type: Bug > Components: FileSystem, State Backends, Checkpointing >Affects Versions: 1.4.2 >Reporter: Jamie Grier >Priority: Critical > > I think we need to modify the way we write checkpoints to S3 for high-scale > jobs (those with many total tasks). The issue is that we are writing all the > checkpoint data under a common key prefix. This is the worst case scenario > for S3 performance since the key is used as a partition key. > > In the worst case checkpoints fail with a 500 status code coming back from S3 > and an internal error type of TooBusyException. > > One possible solution would be to add a hook in the Flink filesystem code > that allows me to "rewrite" paths. For example say I have the checkpoint > directory set to: > > s3://bucket/flink/checkpoints > > I would hook that and rewrite that path to: > > s3://bucket/[HASH]/flink/checkpoints, where HASH is the hash of the original > path > > This would distribute the checkpoint write load around the S3 cluster evenly. > > For reference: > https://aws.amazon.com/premiumsupport/knowledge-center/s3-bucket-performance-improve/ > > Any other people hit this issue? Any other ideas for solutions? This is a > pretty serious problem for people trying to checkpoint to S3. > > -Jamie > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9061) S3 checkpoint data not partitioned well -- causes errors and poor performance
[ https://issues.apache.org/jira/browse/FLINK-9061?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16426103#comment-16426103 ] Jamie Grier commented on FLINK-9061: Okay, this is the best documentation I've found on this: [https://docs.aws.amazon.com/AmazonS3/latest/dev/request-rate-perf-considerations.html] and even it is very vague. It does appear that it doesn't have to be the very first characters but it brings up an interesting question. What are the exact constraints here? Which part of the key name is and isn't used for partitioning exactly? I mean technically all of our checkpoint objects do in fact have several characters of uniqueness since the last part of the full object key name is the GUID. Anyway, not having full info sucks. [~stevenz3wu] I think your proposal sounds good. Thanks for offering to do the PR :) That should work well and logical listing of sub-directories should still be possible in this scheme by issuing parallel s3 list requests for each possible prefix and merging the results. Shall we proceed with this approach then? > S3 checkpoint data not partitioned well -- causes errors and poor performance > - > > Key: FLINK-9061 > URL: https://issues.apache.org/jira/browse/FLINK-9061 > Project: Flink > Issue Type: Bug > Components: FileSystem, State Backends, Checkpointing >Affects Versions: 1.4.2 >Reporter: Jamie Grier >Priority: Critical > > I think we need to modify the way we write checkpoints to S3 for high-scale > jobs (those with many total tasks). The issue is that we are writing all the > checkpoint data under a common key prefix. This is the worst case scenario > for S3 performance since the key is used as a partition key. > > In the worst case checkpoints fail with a 500 status code coming back from S3 > and an internal error type of TooBusyException. > > One possible solution would be to add a hook in the Flink filesystem code > that allows me to "rewrite" paths. For example say I have the checkpoint > directory set to: > > s3://bucket/flink/checkpoints > > I would hook that and rewrite that path to: > > s3://bucket/[HASH]/flink/checkpoints, where HASH is the hash of the original > path > > This would distribute the checkpoint write load around the S3 cluster evenly. > > For reference: > https://aws.amazon.com/premiumsupport/knowledge-center/s3-bucket-performance-improve/ > > Any other people hit this issue? Any other ideas for solutions? This is a > pretty serious problem for people trying to checkpoint to S3. > > -Jamie > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9136) Remove StreamingProgramTestBase
[ https://issues.apache.org/jira/browse/FLINK-9136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425945#comment-16425945 ] ASF GitHub Bot commented on FLINK-9136: --- GitHub user zentol opened a pull request: https://github.com/apache/flink/pull/5817 [FLINK-9136][tests] Remove StreamingProgramTestBase Builds on #5816. ## What is the purpose of the change This PR removes the `StreamingProgramTestBase` class. The class discourages reusing cluster resources as every test must be setup in a new class. Additionally, it appears to mimic to jUnits `@Before/@After` life-cycle, but actually doesn't as the `postSubmit` method is not called not called if `testPogram` fails with an exception. This can lead to resource leaks, like for example in the `ContinuousFileProcessingITCase`. Existing usages were ported to the `AbstractTestBase`. If a tests was using preSubmit to setup data and postSubmit to verify the result, then the methods were merged into a single method. Otherwise, `preSubmit` implementations were annotated with `@Before`, `postSubmit` implementations with `@After`, and `testProgram` implementations with `@Test`. Additionally the visibility of these methods was set to `public`. You can merge this pull request into a Git repository by running: $ git pull https://github.com/zentol/flink 9136 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5817.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5817 commit b3a31933432f4ad4eaeffd28ced16a570910986d Author: zentol Date: 2018-04-04T17:49:35Z [FLINK-9137][tests] Merge TopSpeedWindowingExampleITCase into StreamingExamplesITCase commit a56c6f969a5fb03260fa0d2705356f0d01badb18 Author: zentol Date: 2018-04-04T17:22:53Z [FLINK-9136][tests] Remove StreamingProgramTestBase > Remove StreamingProgramTestBase > --- > > Key: FLINK-9136 > URL: https://issues.apache.org/jira/browse/FLINK-9136 > Project: Flink > Issue Type: Improvement > Components: Tests >Affects Versions: 1.5.0 >Reporter: Chesnay Schepler >Assignee: Chesnay Schepler >Priority: Minor > > The {{StreamingProgramTestBase}} should be removed. We can move all existing > tests to the {{AbstractTestBase}} with junit annotations. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5817: [FLINK-9136][tests] Remove StreamingProgramTestBas...
GitHub user zentol opened a pull request: https://github.com/apache/flink/pull/5817 [FLINK-9136][tests] Remove StreamingProgramTestBase Builds on #5816. ## What is the purpose of the change This PR removes the `StreamingProgramTestBase` class. The class discourages reusing cluster resources as every test must be setup in a new class. Additionally, it appears to mimic to jUnits `@Before/@After` life-cycle, but actually doesn't as the `postSubmit` method is not called not called if `testPogram` fails with an exception. This can lead to resource leaks, like for example in the `ContinuousFileProcessingITCase`. Existing usages were ported to the `AbstractTestBase`. If a tests was using preSubmit to setup data and postSubmit to verify the result, then the methods were merged into a single method. Otherwise, `preSubmit` implementations were annotated with `@Before`, `postSubmit` implementations with `@After`, and `testProgram` implementations with `@Test`. Additionally the visibility of these methods was set to `public`. You can merge this pull request into a Git repository by running: $ git pull https://github.com/zentol/flink 9136 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5817.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5817 commit b3a31933432f4ad4eaeffd28ced16a570910986d Author: zentol Date: 2018-04-04T17:49:35Z [FLINK-9137][tests] Merge TopSpeedWindowingExampleITCase into StreamingExamplesITCase commit a56c6f969a5fb03260fa0d2705356f0d01badb18 Author: zentol Date: 2018-04-04T17:22:53Z [FLINK-9136][tests] Remove StreamingProgramTestBase ---
[jira] [Created] (FLINK-9138) Enhance BucketingSink to also flush data by time interval
Narayanan Arunachalam created FLINK-9138: Summary: Enhance BucketingSink to also flush data by time interval Key: FLINK-9138 URL: https://issues.apache.org/jira/browse/FLINK-9138 Project: Flink Issue Type: Improvement Components: filesystem-connector Affects Versions: 1.4.2 Reporter: Narayanan Arunachalam BucketingSink now supports flushing data to the file system by size limit and by period of inactivity. It will be useful to also flush data by a specified time period. This way, the data will be written out when write throughput is low but there is no significant time period gaps between the writes. This reduces ETA for the data in the file system and should help move the checkpoints faster as well. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9134) Update Calcite dependency to 1.17
[ https://issues.apache.org/jira/browse/FLINK-9134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425926#comment-16425926 ] Shuyi Chen commented on FLINK-9134: --- Hi [~twalthr], this is duplicate of [FLINK-9015|https://issues.apache.org/jira/browse/FLINK-9015]. I'll merge and close this one. > Update Calcite dependency to 1.17 > - > > Key: FLINK-9134 > URL: https://issues.apache.org/jira/browse/FLINK-9134 > Project: Flink > Issue Type: Improvement > Components: Table API & SQL >Reporter: Timo Walther >Priority: Major > > This is an umbrella issue for tasks that need to be performed when upgrading > to Calcite 1.17 once it is released. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9137) Merge TopSpeedWindowingExampleITCase into StreamingExamplesITCase
[ https://issues.apache.org/jira/browse/FLINK-9137?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425924#comment-16425924 ] ASF GitHub Bot commented on FLINK-9137: --- GitHub user zentol opened a pull request: https://github.com/apache/flink/pull/5816 [FLINK-9137][tests] Merge TopSpeedWindowingExampleITCase into StreamingExamplesITCase ## What is the purpose of the change This PR merges the `TopSpeedWindowingExampleITCase` into the `StreamingExamplesITCase`. The latter tests all streaming examples, and I don't see a reason to not include the `TopSpeedWindowingExample` as well. You can merge this pull request into a Git repository by running: $ git pull https://github.com/zentol/flink 9137 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5816.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5816 commit b3a31933432f4ad4eaeffd28ced16a570910986d Author: zentol Date: 2018-04-04T17:49:35Z [FLINK-9137][tests] Merge TopSpeedWindowingExampleITCase into StreamingExamplesITCase > Merge TopSpeedWindowingExampleITCase into StreamingExamplesITCase > - > > Key: FLINK-9137 > URL: https://issues.apache.org/jira/browse/FLINK-9137 > Project: Flink > Issue Type: Improvement > Components: Tests >Affects Versions: 1.5.0 >Reporter: Chesnay Schepler >Assignee: Chesnay Schepler >Priority: Minor > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5816: [FLINK-9137][tests] Merge TopSpeedWindowingExample...
GitHub user zentol opened a pull request: https://github.com/apache/flink/pull/5816 [FLINK-9137][tests] Merge TopSpeedWindowingExampleITCase into StreamingExamplesITCase ## What is the purpose of the change This PR merges the `TopSpeedWindowingExampleITCase` into the `StreamingExamplesITCase`. The latter tests all streaming examples, and I don't see a reason to not include the `TopSpeedWindowingExample` as well. You can merge this pull request into a Git repository by running: $ git pull https://github.com/zentol/flink 9137 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5816.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5816 commit b3a31933432f4ad4eaeffd28ced16a570910986d Author: zentol Date: 2018-04-04T17:49:35Z [FLINK-9137][tests] Merge TopSpeedWindowingExampleITCase into StreamingExamplesITCase ---
[jira] [Created] (FLINK-9137) Merge TopSpeedWindowingExampleITCase into StreamingExamplesITCase
Chesnay Schepler created FLINK-9137: --- Summary: Merge TopSpeedWindowingExampleITCase into StreamingExamplesITCase Key: FLINK-9137 URL: https://issues.apache.org/jira/browse/FLINK-9137 Project: Flink Issue Type: Improvement Components: Tests Affects Versions: 1.5.0 Reporter: Chesnay Schepler Assignee: Chesnay Schepler -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (FLINK-9136) Remove StreamingProgramTestBase
Chesnay Schepler created FLINK-9136: --- Summary: Remove StreamingProgramTestBase Key: FLINK-9136 URL: https://issues.apache.org/jira/browse/FLINK-9136 Project: Flink Issue Type: Improvement Components: Tests Affects Versions: 1.5.0 Reporter: Chesnay Schepler Assignee: Chesnay Schepler The {{StreamingProgramTestBase}} should be removed. We can move all existing tests to the {{AbstractTestBase}} with junit annotations. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Comment Edited] (FLINK-9110) Building docs with Ruby 2.5 fails if bundler is not globally installed
[ https://issues.apache.org/jira/browse/FLINK-9110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425798#comment-16425798 ] Timo Walther edited comment on FLINK-9110 at 4/4/18 5:21 PM: - Fixed in 1.6.0: 29fbc95cad2ad05fd08fb82eeac89e0ade011ea6 Fixed in 1.5.0: 4e7b15dad6eb36397f3a07f9f58a1216657493bb Fixed in 1.4.3: 7610b597b49676299b2dc609d6fd60d4bdccfa2e was (Author: twalthr): Fixed in 1.6.0: 29fbc95cad2ad05fd08fb82eeac89e0ade011ea6 Fixed in 1.5.0: 4e7b15dad6eb36397f3a07f9f58a1216657493bb Fixed in 1.4.0: 7610b597b49676299b2dc609d6fd60d4bdccfa2e > Building docs with Ruby 2.5 fails if bundler is not globally installed > -- > > Key: FLINK-9110 > URL: https://issues.apache.org/jira/browse/FLINK-9110 > Project: Flink > Issue Type: Bug > Components: Documentation >Affects Versions: 1.5.0, 1.6.0 >Reporter: Nico Kruber >Assignee: Nico Kruber >Priority: Blocker > Fix For: 1.5.0, 1.4.3 > > > If {{bundler}} is not installed, {{build_docs.sh}} attempts to install it > locally but updating the {{$PATH}} environment variable is broken at least in > my setup with ruby 2.5 because of this command failing: > {code} > > ruby -rubygems -e 'puts Gem.user_dir' > Traceback (most recent call last): > 1: from > /usr/lib64/ruby/2.5.0/rubygems/core_ext/kernel_require.rb:59:in `require' > /usr/lib64/ruby/2.5.0/rubygems/core_ext/kernel_require.rb:59:in `require': > cannot load such file -- ubygems (LoadError) > > ruby -e 'puts Gem.user_dir' > /home/nico/.gem/ruby/2.5.0 > {code} > Additionally, the {{bundle}} binary is not even in that path:: > {code} > > find ~/.gem/ruby/2.*/bin > /home/nico/.gem/ruby/2.4.0/bin > /home/nico/.gem/ruby/2.4.0/bin/bundle.ruby2.4 > /home/nico/.gem/ruby/2.4.0/bin/bundler.ruby2.4 > /home/nico/.gem/ruby/2.5.0/bin > /home/nico/.gem/ruby/2.5.0/bin/bundle.ruby2.5 > /home/nico/.gem/ruby/2.5.0/bin/bundler.ruby2.5 > {code} > but indeed here: > {code} > > ls ~/.gem/ruby/2.*/gems/bundler-*/exe/bundle > /home/nico/.gem/ruby/2.4.0/gems/bundler-1.15.3/exe/bundle > /home/nico/.gem/ruby/2.5.0/gems/bundler-1.16.1/exe/bundle > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (FLINK-9107) Document timer coalescing for ProcessFunctions
[ https://issues.apache.org/jira/browse/FLINK-9107?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Timo Walther resolved FLINK-9107. - Resolution: Fixed Fix Version/s: (was: 1.3.4) Fixed in 1.6.0: 7b0fc58f75494c9a2c71d551632445ded85c0a45 Fixed in 1.5.0: f083622a200c79395ecf16e2be6f8b540fe85178 Fixed in 1.4.3: ca8f4ca4fdf4fbd98a86b32a2a77dbb00742e164 > Document timer coalescing for ProcessFunctions > -- > > Key: FLINK-9107 > URL: https://issues.apache.org/jira/browse/FLINK-9107 > Project: Flink > Issue Type: Improvement > Components: Documentation, Streaming >Affects Versions: 1.3.0, 1.4.0, 1.5.0, 1.6.0 >Reporter: Nico Kruber >Assignee: Nico Kruber >Priority: Major > Fix For: 1.5.0, 1.4.3 > > > In a {{ProcessFunction}}, registering timers for each event via > {{ctx.timerService().registerEventTimeTimer()}} using times like > {{ctx.timestamp() + timeout}} will get a millisecond accuracy and may thus > create one timer per millisecond which may lead to some overhead in the > {{TimerService}}. > This problem can be mitigated by using timer coalescing if the desired > accuracy of the timer can be larger than 1ms. A timer firing at full seconds > only, for example, can be realised like this: > {code} > coalescedTime = ((ctx.timestamp() + timeout) / 1000) * 1000; > ctx.timerService().registerEventTimeTimer(coalescedTime); > {code} > As a result, only a single timer may exist for every second since we do not > add timers for timestamps that are already there. > This should be documented in the {{ProcessFunction}} docs. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9107) Document timer coalescing for ProcessFunctions
[ https://issues.apache.org/jira/browse/FLINK-9107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425881#comment-16425881 ] ASF GitHub Bot commented on FLINK-9107: --- Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5790 > Document timer coalescing for ProcessFunctions > -- > > Key: FLINK-9107 > URL: https://issues.apache.org/jira/browse/FLINK-9107 > Project: Flink > Issue Type: Improvement > Components: Documentation, Streaming >Affects Versions: 1.3.0, 1.4.0, 1.5.0, 1.6.0 >Reporter: Nico Kruber >Assignee: Nico Kruber >Priority: Major > Fix For: 1.5.0, 1.4.3, 1.3.4 > > > In a {{ProcessFunction}}, registering timers for each event via > {{ctx.timerService().registerEventTimeTimer()}} using times like > {{ctx.timestamp() + timeout}} will get a millisecond accuracy and may thus > create one timer per millisecond which may lead to some overhead in the > {{TimerService}}. > This problem can be mitigated by using timer coalescing if the desired > accuracy of the timer can be larger than 1ms. A timer firing at full seconds > only, for example, can be realised like this: > {code} > coalescedTime = ((ctx.timestamp() + timeout) / 1000) * 1000; > ctx.timerService().registerEventTimeTimer(coalescedTime); > {code} > As a result, only a single timer may exist for every second since we do not > add timers for timestamps that are already there. > This should be documented in the {{ProcessFunction}} docs. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5790: [FLINK-9107][docs] document timer coalescing for P...
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5790 ---
[jira] [Commented] (FLINK-9131) Disable spotbugs on travis
[ https://issues.apache.org/jira/browse/FLINK-9131?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425874#comment-16425874 ] ASF GitHub Bot commented on FLINK-9131: --- Github user zentol closed the pull request at: https://github.com/apache/flink/pull/5815 > Disable spotbugs on travis > -- > > Key: FLINK-9131 > URL: https://issues.apache.org/jira/browse/FLINK-9131 > Project: Flink > Issue Type: Improvement > Components: Travis >Affects Versions: 1.5.0 >Reporter: Chesnay Schepler >Assignee: Chesnay Schepler >Priority: Critical > Fix For: 1.5.0 > > > The misc profile that also runs spotbugs is consistently timing out on travis > at the moment. > The spotbugs plugin is a major contributor to the compilation time, for > example it doubles the compile time for flink-runtime. > I suggest to temporarily disable spotbugs, and re-enable it at a lter point > when we figure out the daily cron jobs. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-9131) Disable spotbugs on travis
[ https://issues.apache.org/jira/browse/FLINK-9131?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler closed FLINK-9131. --- Resolution: Fixed Fix Version/s: 1.5.0 master: bd8b47956bde0e7ff7ed5cc6d4bd79a875957835 1.5: 2ae48534ed6df42da9530f95d2b9179a6dbc9015 > Disable spotbugs on travis > -- > > Key: FLINK-9131 > URL: https://issues.apache.org/jira/browse/FLINK-9131 > Project: Flink > Issue Type: Improvement > Components: Travis >Affects Versions: 1.5.0 >Reporter: Chesnay Schepler >Assignee: Chesnay Schepler >Priority: Critical > Fix For: 1.5.0 > > > The misc profile that also runs spotbugs is consistently timing out on travis > at the moment. > The spotbugs plugin is a major contributor to the compilation time, for > example it doubles the compile time for flink-runtime. > I suggest to temporarily disable spotbugs, and re-enable it at a lter point > when we figure out the daily cron jobs. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5815: [FLINK-9131][travis] Disable spotbugs plugin
Github user zentol closed the pull request at: https://github.com/apache/flink/pull/5815 ---
[jira] [Commented] (FLINK-9107) Document timer coalescing for ProcessFunctions
[ https://issues.apache.org/jira/browse/FLINK-9107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425865#comment-16425865 ] ASF GitHub Bot commented on FLINK-9107: --- Github user twalthr commented on the issue: https://github.com/apache/flink/pull/5790 Thank you @NicoK. I will merge this... > Document timer coalescing for ProcessFunctions > -- > > Key: FLINK-9107 > URL: https://issues.apache.org/jira/browse/FLINK-9107 > Project: Flink > Issue Type: Improvement > Components: Documentation, Streaming >Affects Versions: 1.3.0, 1.4.0, 1.5.0, 1.6.0 >Reporter: Nico Kruber >Assignee: Nico Kruber >Priority: Major > Fix For: 1.5.0, 1.4.3, 1.3.4 > > > In a {{ProcessFunction}}, registering timers for each event via > {{ctx.timerService().registerEventTimeTimer()}} using times like > {{ctx.timestamp() + timeout}} will get a millisecond accuracy and may thus > create one timer per millisecond which may lead to some overhead in the > {{TimerService}}. > This problem can be mitigated by using timer coalescing if the desired > accuracy of the timer can be larger than 1ms. A timer firing at full seconds > only, for example, can be realised like this: > {code} > coalescedTime = ((ctx.timestamp() + timeout) / 1000) * 1000; > ctx.timerService().registerEventTimeTimer(coalescedTime); > {code} > As a result, only a single timer may exist for every second since we do not > add timers for timestamps that are already there. > This should be documented in the {{ProcessFunction}} docs. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink issue #5790: [FLINK-9107][docs] document timer coalescing for ProcessF...
Github user twalthr commented on the issue: https://github.com/apache/flink/pull/5790 Thank you @NicoK. I will merge this... ---
[jira] [Created] (FLINK-9135) Remove AggregateReduceFunctionsRule once CALCITE-2216 is fixed
Fabian Hueske created FLINK-9135: Summary: Remove AggregateReduceFunctionsRule once CALCITE-2216 is fixed Key: FLINK-9135 URL: https://issues.apache.org/jira/browse/FLINK-9135 Project: Flink Issue Type: Sub-task Components: Table API & SQL Affects Versions: 1.5.0, 1.6.0 Reporter: Fabian Hueske We had to copy and slightly modify {{AggregateReduceFunctionsRule}} from Calcite to fix FLINK-8903. We proposed the changes to Calcite as CALCITE-2216. Once this issue is fixed and we updated to Calcite dependency to a version that includes the fix, we can remove our custom rule. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9110) Building docs with Ruby 2.5 fails if bundler is not globally installed
[ https://issues.apache.org/jira/browse/FLINK-9110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425834#comment-16425834 ] Timo Walther commented on FLINK-9110: - Also merged to flink-web: 8b7c2dcb2e46ee1dbc6f4dab49c8b4d89b46027d > Building docs with Ruby 2.5 fails if bundler is not globally installed > -- > > Key: FLINK-9110 > URL: https://issues.apache.org/jira/browse/FLINK-9110 > Project: Flink > Issue Type: Bug > Components: Documentation >Affects Versions: 1.5.0, 1.6.0 >Reporter: Nico Kruber >Assignee: Nico Kruber >Priority: Blocker > Fix For: 1.5.0, 1.4.3 > > > If {{bundler}} is not installed, {{build_docs.sh}} attempts to install it > locally but updating the {{$PATH}} environment variable is broken at least in > my setup with ruby 2.5 because of this command failing: > {code} > > ruby -rubygems -e 'puts Gem.user_dir' > Traceback (most recent call last): > 1: from > /usr/lib64/ruby/2.5.0/rubygems/core_ext/kernel_require.rb:59:in `require' > /usr/lib64/ruby/2.5.0/rubygems/core_ext/kernel_require.rb:59:in `require': > cannot load such file -- ubygems (LoadError) > > ruby -e 'puts Gem.user_dir' > /home/nico/.gem/ruby/2.5.0 > {code} > Additionally, the {{bundle}} binary is not even in that path:: > {code} > > find ~/.gem/ruby/2.*/bin > /home/nico/.gem/ruby/2.4.0/bin > /home/nico/.gem/ruby/2.4.0/bin/bundle.ruby2.4 > /home/nico/.gem/ruby/2.4.0/bin/bundler.ruby2.4 > /home/nico/.gem/ruby/2.5.0/bin > /home/nico/.gem/ruby/2.5.0/bin/bundle.ruby2.5 > /home/nico/.gem/ruby/2.5.0/bin/bundler.ruby2.5 > {code} > but indeed here: > {code} > > ls ~/.gem/ruby/2.*/gems/bundler-*/exe/bundle > /home/nico/.gem/ruby/2.4.0/gems/bundler-1.15.3/exe/bundle > /home/nico/.gem/ruby/2.5.0/gems/bundler-1.16.1/exe/bundle > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (FLINK-9110) Building docs with Ruby 2.5 fails if bundler is not globally installed
[ https://issues.apache.org/jira/browse/FLINK-9110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Timo Walther resolved FLINK-9110. - Resolution: Fixed Fix Version/s: 1.4.3 Fixed in 1.6.0: 29fbc95cad2ad05fd08fb82eeac89e0ade011ea6 Fixed in 1.5.0: 4e7b15dad6eb36397f3a07f9f58a1216657493bb Fixed in 1.4.0: 7610b597b49676299b2dc609d6fd60d4bdccfa2e > Building docs with Ruby 2.5 fails if bundler is not globally installed > -- > > Key: FLINK-9110 > URL: https://issues.apache.org/jira/browse/FLINK-9110 > Project: Flink > Issue Type: Bug > Components: Documentation >Affects Versions: 1.5.0, 1.6.0 >Reporter: Nico Kruber >Assignee: Nico Kruber >Priority: Blocker > Fix For: 1.5.0, 1.4.3 > > > If {{bundler}} is not installed, {{build_docs.sh}} attempts to install it > locally but updating the {{$PATH}} environment variable is broken at least in > my setup with ruby 2.5 because of this command failing: > {code} > > ruby -rubygems -e 'puts Gem.user_dir' > Traceback (most recent call last): > 1: from > /usr/lib64/ruby/2.5.0/rubygems/core_ext/kernel_require.rb:59:in `require' > /usr/lib64/ruby/2.5.0/rubygems/core_ext/kernel_require.rb:59:in `require': > cannot load such file -- ubygems (LoadError) > > ruby -e 'puts Gem.user_dir' > /home/nico/.gem/ruby/2.5.0 > {code} > Additionally, the {{bundle}} binary is not even in that path:: > {code} > > find ~/.gem/ruby/2.*/bin > /home/nico/.gem/ruby/2.4.0/bin > /home/nico/.gem/ruby/2.4.0/bin/bundle.ruby2.4 > /home/nico/.gem/ruby/2.4.0/bin/bundler.ruby2.4 > /home/nico/.gem/ruby/2.5.0/bin > /home/nico/.gem/ruby/2.5.0/bin/bundle.ruby2.5 > /home/nico/.gem/ruby/2.5.0/bin/bundler.ruby2.5 > {code} > but indeed here: > {code} > > ls ~/.gem/ruby/2.*/gems/bundler-*/exe/bundle > /home/nico/.gem/ruby/2.4.0/gems/bundler-1.15.3/exe/bundle > /home/nico/.gem/ruby/2.5.0/gems/bundler-1.16.1/exe/bundle > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9110) Building docs with Ruby 2.5 fails if bundler is not globally installed
[ https://issues.apache.org/jira/browse/FLINK-9110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425787#comment-16425787 ] ASF GitHub Bot commented on FLINK-9110: --- Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5788 > Building docs with Ruby 2.5 fails if bundler is not globally installed > -- > > Key: FLINK-9110 > URL: https://issues.apache.org/jira/browse/FLINK-9110 > Project: Flink > Issue Type: Bug > Components: Documentation >Affects Versions: 1.5.0, 1.6.0 >Reporter: Nico Kruber >Assignee: Nico Kruber >Priority: Blocker > Fix For: 1.5.0 > > > If {{bundler}} is not installed, {{build_docs.sh}} attempts to install it > locally but updating the {{$PATH}} environment variable is broken at least in > my setup with ruby 2.5 because of this command failing: > {code} > > ruby -rubygems -e 'puts Gem.user_dir' > Traceback (most recent call last): > 1: from > /usr/lib64/ruby/2.5.0/rubygems/core_ext/kernel_require.rb:59:in `require' > /usr/lib64/ruby/2.5.0/rubygems/core_ext/kernel_require.rb:59:in `require': > cannot load such file -- ubygems (LoadError) > > ruby -e 'puts Gem.user_dir' > /home/nico/.gem/ruby/2.5.0 > {code} > Additionally, the {{bundle}} binary is not even in that path:: > {code} > > find ~/.gem/ruby/2.*/bin > /home/nico/.gem/ruby/2.4.0/bin > /home/nico/.gem/ruby/2.4.0/bin/bundle.ruby2.4 > /home/nico/.gem/ruby/2.4.0/bin/bundler.ruby2.4 > /home/nico/.gem/ruby/2.5.0/bin > /home/nico/.gem/ruby/2.5.0/bin/bundle.ruby2.5 > /home/nico/.gem/ruby/2.5.0/bin/bundler.ruby2.5 > {code} > but indeed here: > {code} > > ls ~/.gem/ruby/2.*/gems/bundler-*/exe/bundle > /home/nico/.gem/ruby/2.4.0/gems/bundler-1.15.3/exe/bundle > /home/nico/.gem/ruby/2.5.0/gems/bundler-1.16.1/exe/bundle > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5788: [FLINK-9110][docs] fix local bundler installation
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5788 ---
[jira] [Updated] (FLINK-9031) DataSet Job result changes when adding rebalance after union
[ https://issues.apache.org/jira/browse/FLINK-9031?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fabian Hueske updated FLINK-9031: - Fix Version/s: 1.3.4 > DataSet Job result changes when adding rebalance after union > > > Key: FLINK-9031 > URL: https://issues.apache.org/jira/browse/FLINK-9031 > Project: Flink > Issue Type: Bug > Components: DataSet API, Local Runtime, Optimizer >Affects Versions: 1.3.1 >Reporter: Fabian Hueske >Assignee: Fabian Hueske >Priority: Critical > Fix For: 1.5.0, 1.4.3, 1.3.4 > > Attachments: Person.java, RunAll.java, newplan.txt, oldplan.txt > > > A user [reported this issue on the user mailing > list|https://lists.apache.org/thread.html/075f1a487b044079b5d61f199439cb77dd4174bd425bcb3327ed7dfc@%3Cuser.flink.apache.org%3E]. > {quote}I am using Flink 1.3.1 and I have found a strange behavior on running > the following logic: > # Read data from file and store into DataSet > # Split dataset in two, by checking if "field1" of POJOs is empty or not, so > that the first dataset contains only elements with non empty "field1", and > the second dataset will contain the other elements. > # Each dataset is then grouped by, one by "field1" and other by another > field, and subsequently reduced. > # The 2 datasets are merged together by union. > # The final dataset is written as json. > What I was expected, from output, was to find only one element with a > specific value of "field1" because: > # Reducing the first dataset grouped by "field1" should generate only one > element with a specific value of "field1". > # The second dataset should contain only elements with empty "field1". > # Making an union of them should not duplicate any record. > This does not happen. When i read the generated jsons i see some duplicate > (non empty) values of "field1". > Strangely this does not happen when the union between the two datasets is > not computed. In this case the first dataset produces elements only with > distinct values of "field1", while second dataset produces only records with > empty field "value1". > {quote} > The user has not enable object reuse. > Later he reports that the problem disappears when he injects a rebalance() > after a union resolves the problem. I had a look at the execution plans for > both cases (attached to this issue) but could not identify a problem. > Hence I assume, this might be an issue with the runtime code but we need to > look deeper into this. The user also provided an example program consisting > of two classes which are attached to the issue as well. > > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Reopened] (FLINK-9031) DataSet Job result changes when adding rebalance after union
[ https://issues.apache.org/jira/browse/FLINK-9031?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fabian Hueske reopened FLINK-9031: -- Added fix for Flink 1.3.4 > DataSet Job result changes when adding rebalance after union > > > Key: FLINK-9031 > URL: https://issues.apache.org/jira/browse/FLINK-9031 > Project: Flink > Issue Type: Bug > Components: DataSet API, Local Runtime, Optimizer >Affects Versions: 1.3.1 >Reporter: Fabian Hueske >Assignee: Fabian Hueske >Priority: Critical > Fix For: 1.5.0, 1.4.3, 1.3.4 > > Attachments: Person.java, RunAll.java, newplan.txt, oldplan.txt > > > A user [reported this issue on the user mailing > list|https://lists.apache.org/thread.html/075f1a487b044079b5d61f199439cb77dd4174bd425bcb3327ed7dfc@%3Cuser.flink.apache.org%3E]. > {quote}I am using Flink 1.3.1 and I have found a strange behavior on running > the following logic: > # Read data from file and store into DataSet > # Split dataset in two, by checking if "field1" of POJOs is empty or not, so > that the first dataset contains only elements with non empty "field1", and > the second dataset will contain the other elements. > # Each dataset is then grouped by, one by "field1" and other by another > field, and subsequently reduced. > # The 2 datasets are merged together by union. > # The final dataset is written as json. > What I was expected, from output, was to find only one element with a > specific value of "field1" because: > # Reducing the first dataset grouped by "field1" should generate only one > element with a specific value of "field1". > # The second dataset should contain only elements with empty "field1". > # Making an union of them should not duplicate any record. > This does not happen. When i read the generated jsons i see some duplicate > (non empty) values of "field1". > Strangely this does not happen when the union between the two datasets is > not computed. In this case the first dataset produces elements only with > distinct values of "field1", while second dataset produces only records with > empty field "value1". > {quote} > The user has not enable object reuse. > Later he reports that the problem disappears when he injects a rebalance() > after a union resolves the problem. I had a look at the execution plans for > both cases (attached to this issue) but could not identify a problem. > Hence I assume, this might be an issue with the runtime code but we need to > look deeper into this. The user also provided an example program consisting > of two classes which are attached to the issue as well. > > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-9031) DataSet Job result changes when adding rebalance after union
[ https://issues.apache.org/jira/browse/FLINK-9031?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fabian Hueske closed FLINK-9031. Resolution: Fixed > DataSet Job result changes when adding rebalance after union > > > Key: FLINK-9031 > URL: https://issues.apache.org/jira/browse/FLINK-9031 > Project: Flink > Issue Type: Bug > Components: DataSet API, Local Runtime, Optimizer >Affects Versions: 1.3.1 >Reporter: Fabian Hueske >Assignee: Fabian Hueske >Priority: Critical > Fix For: 1.5.0, 1.4.3, 1.3.4 > > Attachments: Person.java, RunAll.java, newplan.txt, oldplan.txt > > > A user [reported this issue on the user mailing > list|https://lists.apache.org/thread.html/075f1a487b044079b5d61f199439cb77dd4174bd425bcb3327ed7dfc@%3Cuser.flink.apache.org%3E]. > {quote}I am using Flink 1.3.1 and I have found a strange behavior on running > the following logic: > # Read data from file and store into DataSet > # Split dataset in two, by checking if "field1" of POJOs is empty or not, so > that the first dataset contains only elements with non empty "field1", and > the second dataset will contain the other elements. > # Each dataset is then grouped by, one by "field1" and other by another > field, and subsequently reduced. > # The 2 datasets are merged together by union. > # The final dataset is written as json. > What I was expected, from output, was to find only one element with a > specific value of "field1" because: > # Reducing the first dataset grouped by "field1" should generate only one > element with a specific value of "field1". > # The second dataset should contain only elements with empty "field1". > # Making an union of them should not duplicate any record. > This does not happen. When i read the generated jsons i see some duplicate > (non empty) values of "field1". > Strangely this does not happen when the union between the two datasets is > not computed. In this case the first dataset produces elements only with > distinct values of "field1", while second dataset produces only records with > empty field "value1". > {quote} > The user has not enable object reuse. > Later he reports that the problem disappears when he injects a rebalance() > after a union resolves the problem. I had a look at the execution plans for > both cases (attached to this issue) but could not identify a problem. > Hence I assume, this might be an issue with the runtime code but we need to > look deeper into this. The user also provided an example program consisting > of two classes which are attached to the issue as well. > > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9031) DataSet Job result changes when adding rebalance after union
[ https://issues.apache.org/jira/browse/FLINK-9031?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425774#comment-16425774 ] Fabian Hueske commented on FLINK-9031: -- Fixed for 1.3.4 with cb38b6defbea5f92b6f3a5874acacb56523534f0 > DataSet Job result changes when adding rebalance after union > > > Key: FLINK-9031 > URL: https://issues.apache.org/jira/browse/FLINK-9031 > Project: Flink > Issue Type: Bug > Components: DataSet API, Local Runtime, Optimizer >Affects Versions: 1.3.1 >Reporter: Fabian Hueske >Assignee: Fabian Hueske >Priority: Critical > Fix For: 1.5.0, 1.4.3 > > Attachments: Person.java, RunAll.java, newplan.txt, oldplan.txt > > > A user [reported this issue on the user mailing > list|https://lists.apache.org/thread.html/075f1a487b044079b5d61f199439cb77dd4174bd425bcb3327ed7dfc@%3Cuser.flink.apache.org%3E]. > {quote}I am using Flink 1.3.1 and I have found a strange behavior on running > the following logic: > # Read data from file and store into DataSet > # Split dataset in two, by checking if "field1" of POJOs is empty or not, so > that the first dataset contains only elements with non empty "field1", and > the second dataset will contain the other elements. > # Each dataset is then grouped by, one by "field1" and other by another > field, and subsequently reduced. > # The 2 datasets are merged together by union. > # The final dataset is written as json. > What I was expected, from output, was to find only one element with a > specific value of "field1" because: > # Reducing the first dataset grouped by "field1" should generate only one > element with a specific value of "field1". > # The second dataset should contain only elements with empty "field1". > # Making an union of them should not duplicate any record. > This does not happen. When i read the generated jsons i see some duplicate > (non empty) values of "field1". > Strangely this does not happen when the union between the two datasets is > not computed. In this case the first dataset produces elements only with > distinct values of "field1", while second dataset produces only records with > empty field "value1". > {quote} > The user has not enable object reuse. > Later he reports that the problem disappears when he injects a rebalance() > after a union resolves the problem. I had a look at the execution plans for > both cases (attached to this issue) but could not identify a problem. > Hence I assume, this might be an issue with the runtime code but we need to > look deeper into this. The user also provided an example program consisting > of two classes which are attached to the issue as well. > > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink issue #5788: [FLINK-9110][docs] fix local bundler installation
Github user twalthr commented on the issue: https://github.com/apache/flink/pull/5788 Merging... ---
[jira] [Commented] (FLINK-9110) Building docs with Ruby 2.5 fails if bundler is not globally installed
[ https://issues.apache.org/jira/browse/FLINK-9110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425754#comment-16425754 ] ASF GitHub Bot commented on FLINK-9110: --- Github user twalthr commented on the issue: https://github.com/apache/flink/pull/5788 Merging... > Building docs with Ruby 2.5 fails if bundler is not globally installed > -- > > Key: FLINK-9110 > URL: https://issues.apache.org/jira/browse/FLINK-9110 > Project: Flink > Issue Type: Bug > Components: Documentation >Affects Versions: 1.5.0, 1.6.0 >Reporter: Nico Kruber >Assignee: Nico Kruber >Priority: Blocker > Fix For: 1.5.0 > > > If {{bundler}} is not installed, {{build_docs.sh}} attempts to install it > locally but updating the {{$PATH}} environment variable is broken at least in > my setup with ruby 2.5 because of this command failing: > {code} > > ruby -rubygems -e 'puts Gem.user_dir' > Traceback (most recent call last): > 1: from > /usr/lib64/ruby/2.5.0/rubygems/core_ext/kernel_require.rb:59:in `require' > /usr/lib64/ruby/2.5.0/rubygems/core_ext/kernel_require.rb:59:in `require': > cannot load such file -- ubygems (LoadError) > > ruby -e 'puts Gem.user_dir' > /home/nico/.gem/ruby/2.5.0 > {code} > Additionally, the {{bundle}} binary is not even in that path:: > {code} > > find ~/.gem/ruby/2.*/bin > /home/nico/.gem/ruby/2.4.0/bin > /home/nico/.gem/ruby/2.4.0/bin/bundle.ruby2.4 > /home/nico/.gem/ruby/2.4.0/bin/bundler.ruby2.4 > /home/nico/.gem/ruby/2.5.0/bin > /home/nico/.gem/ruby/2.5.0/bin/bundle.ruby2.5 > /home/nico/.gem/ruby/2.5.0/bin/bundler.ruby2.5 > {code} > but indeed here: > {code} > > ls ~/.gem/ruby/2.*/gems/bundler-*/exe/bundle > /home/nico/.gem/ruby/2.4.0/gems/bundler-1.15.3/exe/bundle > /home/nico/.gem/ruby/2.5.0/gems/bundler-1.16.1/exe/bundle > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9110) Building docs with Ruby 2.5 fails if bundler is not globally installed
[ https://issues.apache.org/jira/browse/FLINK-9110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425753#comment-16425753 ] ASF GitHub Bot commented on FLINK-9110: --- Github user twalthr commented on the issue: https://github.com/apache/flink/pull/5788 Thank you @NicoK. I tested it locally on MacOS. The changes look good. > Building docs with Ruby 2.5 fails if bundler is not globally installed > -- > > Key: FLINK-9110 > URL: https://issues.apache.org/jira/browse/FLINK-9110 > Project: Flink > Issue Type: Bug > Components: Documentation >Affects Versions: 1.5.0, 1.6.0 >Reporter: Nico Kruber >Assignee: Nico Kruber >Priority: Blocker > Fix For: 1.5.0 > > > If {{bundler}} is not installed, {{build_docs.sh}} attempts to install it > locally but updating the {{$PATH}} environment variable is broken at least in > my setup with ruby 2.5 because of this command failing: > {code} > > ruby -rubygems -e 'puts Gem.user_dir' > Traceback (most recent call last): > 1: from > /usr/lib64/ruby/2.5.0/rubygems/core_ext/kernel_require.rb:59:in `require' > /usr/lib64/ruby/2.5.0/rubygems/core_ext/kernel_require.rb:59:in `require': > cannot load such file -- ubygems (LoadError) > > ruby -e 'puts Gem.user_dir' > /home/nico/.gem/ruby/2.5.0 > {code} > Additionally, the {{bundle}} binary is not even in that path:: > {code} > > find ~/.gem/ruby/2.*/bin > /home/nico/.gem/ruby/2.4.0/bin > /home/nico/.gem/ruby/2.4.0/bin/bundle.ruby2.4 > /home/nico/.gem/ruby/2.4.0/bin/bundler.ruby2.4 > /home/nico/.gem/ruby/2.5.0/bin > /home/nico/.gem/ruby/2.5.0/bin/bundle.ruby2.5 > /home/nico/.gem/ruby/2.5.0/bin/bundler.ruby2.5 > {code} > but indeed here: > {code} > > ls ~/.gem/ruby/2.*/gems/bundler-*/exe/bundle > /home/nico/.gem/ruby/2.4.0/gems/bundler-1.15.3/exe/bundle > /home/nico/.gem/ruby/2.5.0/gems/bundler-1.16.1/exe/bundle > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink issue #5788: [FLINK-9110][docs] fix local bundler installation
Github user twalthr commented on the issue: https://github.com/apache/flink/pull/5788 Thank you @NicoK. I tested it locally on MacOS. The changes look good. ---
[jira] [Updated] (FLINK-7235) Backport CALCITE-1884 to the Flink repository before Calcite 1.14
[ https://issues.apache.org/jira/browse/FLINK-7235?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Timo Walther updated FLINK-7235: Parent Issue: FLINK-9134 (was: FLINK-8507) > Backport CALCITE-1884 to the Flink repository before Calcite 1.14 > - > > Key: FLINK-7235 > URL: https://issues.apache.org/jira/browse/FLINK-7235 > Project: Flink > Issue Type: Sub-task > Components: Table API & SQL >Reporter: Haohui Mai >Assignee: Haohui Mai >Priority: Major > > We need to backport CALCITE-1884 in order to unblock upgrading Calcite to > 1.13. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-7237) Remove DateTimeUtils from Flink once Calcite is upgraded to 1.14
[ https://issues.apache.org/jira/browse/FLINK-7237?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Timo Walther updated FLINK-7237: Parent Issue: FLINK-9134 (was: FLINK-8507) > Remove DateTimeUtils from Flink once Calcite is upgraded to 1.14 > > > Key: FLINK-7237 > URL: https://issues.apache.org/jira/browse/FLINK-7237 > Project: Flink > Issue Type: Sub-task > Components: Table API & SQL >Reporter: Haohui Mai >Priority: Major > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9059) Add support for unified table source and sink declaration in environment file
[ https://issues.apache.org/jira/browse/FLINK-9059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425717#comment-16425717 ] ASF GitHub Bot commented on FLINK-9059: --- Github user twalthr commented on a diff in the pull request: https://github.com/apache/flink/pull/5758#discussion_r179182994 --- Diff: flink-libraries/flink-table/src/main/scala/org/apache/flink/table/descriptors/TableDescriptor.scala --- @@ -0,0 +1,75 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.table.descriptors + +import org.apache.flink.table.descriptors.DescriptorProperties.toScala +import org.apache.flink.table.descriptors.StatisticsValidator.{STATISTICS_COLUMNS, STATISTICS_ROW_COUNT, readColumnStats} +import org.apache.flink.table.plan.stats.TableStats + +import scala.collection.JavaConverters._ + +/** + * Common class for all descriptors describing table sources and sinks. + */ +abstract class TableDescriptor extends Descriptor { + + protected var connectorDescriptor: Option[ConnectorDescriptor] = None + protected var formatDescriptor: Option[FormatDescriptor] = None + protected var schemaDescriptor: Option[Schema] = None + protected var statisticsDescriptor: Option[Statistics] = None --- End diff -- Thanks for pointing this out. You are right. Sinks can have a schema but no statistics. I was just wondering if we really need most of the refactorings in this PR. We need to rework the `TableSourceDescriptor` class in the near future because a Java user can access all `protected` field which is not very nice API design. > Add support for unified table source and sink declaration in environment file > - > > Key: FLINK-9059 > URL: https://issues.apache.org/jira/browse/FLINK-9059 > Project: Flink > Issue Type: Task > Components: Table API & SQL >Reporter: Shuyi Chen >Assignee: Shuyi Chen >Priority: Major > Fix For: 1.5.0 > > > 1) Add a common property called "type" with single value 'source'. > 2) in yaml file, replace "sources" with "tables". -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5758: [FLINK-9059][Table API & SQL] add table type attri...
Github user twalthr commented on a diff in the pull request: https://github.com/apache/flink/pull/5758#discussion_r179182994 --- Diff: flink-libraries/flink-table/src/main/scala/org/apache/flink/table/descriptors/TableDescriptor.scala --- @@ -0,0 +1,75 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.table.descriptors + +import org.apache.flink.table.descriptors.DescriptorProperties.toScala +import org.apache.flink.table.descriptors.StatisticsValidator.{STATISTICS_COLUMNS, STATISTICS_ROW_COUNT, readColumnStats} +import org.apache.flink.table.plan.stats.TableStats + +import scala.collection.JavaConverters._ + +/** + * Common class for all descriptors describing table sources and sinks. + */ +abstract class TableDescriptor extends Descriptor { + + protected var connectorDescriptor: Option[ConnectorDescriptor] = None + protected var formatDescriptor: Option[FormatDescriptor] = None + protected var schemaDescriptor: Option[Schema] = None + protected var statisticsDescriptor: Option[Statistics] = None --- End diff -- Thanks for pointing this out. You are right. Sinks can have a schema but no statistics. I was just wondering if we really need most of the refactorings in this PR. We need to rework the `TableSourceDescriptor` class in the near future because a Java user can access all `protected` field which is not very nice API design. ---
[jira] [Created] (FLINK-9134) Update Calcite dependency to 1.17
Timo Walther created FLINK-9134: --- Summary: Update Calcite dependency to 1.17 Key: FLINK-9134 URL: https://issues.apache.org/jira/browse/FLINK-9134 Project: Flink Issue Type: Improvement Components: Table API & SQL Reporter: Timo Walther This is an umbrella issue for tasks that need to be performed when upgrading to Calcite 1.17 once it is released. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (FLINK-8508) Remove RexSimplify from Flink repo
[ https://issues.apache.org/jira/browse/FLINK-8508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Timo Walther resolved FLINK-8508. - Resolution: Fixed Fix Version/s: 1.5.0 Fixed in 1.6.0: 176a893d2084cc48f9b2b7849ada9a1bd56e Fixed in 1.5.0: 2c626d1404439f8fdc81f64d9db4531e5530771a > Remove RexSimplify from Flink repo > -- > > Key: FLINK-8508 > URL: https://issues.apache.org/jira/browse/FLINK-8508 > Project: Flink > Issue Type: Task > Components: Table API & SQL >Reporter: Shuyi Chen >Assignee: Shuyi Chen >Priority: Major > Fix For: 1.5.0 > > > RexSimplify is copied to the Flink repo due to > [CALCITE-2110|https://issues.apache.org/jira/browse/CALCITE-2110], we should > remove it once flink upgrade Calcite dependency to 1.16. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-8508) Remove RexSimplify from Flink repo
[ https://issues.apache.org/jira/browse/FLINK-8508?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425703#comment-16425703 ] ASF GitHub Bot commented on FLINK-8508: --- Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5793 > Remove RexSimplify from Flink repo > -- > > Key: FLINK-8508 > URL: https://issues.apache.org/jira/browse/FLINK-8508 > Project: Flink > Issue Type: Task > Components: Table API & SQL >Reporter: Shuyi Chen >Assignee: Shuyi Chen >Priority: Major > Fix For: 1.5.0 > > > RexSimplify is copied to the Flink repo due to > [CALCITE-2110|https://issues.apache.org/jira/browse/CALCITE-2110], we should > remove it once flink upgrade Calcite dependency to 1.16. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5793: [FLINK-8508][Table API & SQL] Remove RexSimplify c...
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5793 ---
[jira] [Commented] (FLINK-9059) Add support for unified table source and sink declaration in environment file
[ https://issues.apache.org/jira/browse/FLINK-9059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425676#comment-16425676 ] ASF GitHub Bot commented on FLINK-9059: --- Github user walterddr commented on a diff in the pull request: https://github.com/apache/flink/pull/5758#discussion_r179178156 --- Diff: flink-libraries/flink-table/src/main/scala/org/apache/flink/table/descriptors/TableDescriptor.scala --- @@ -0,0 +1,75 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.table.descriptors + +import org.apache.flink.table.descriptors.DescriptorProperties.toScala +import org.apache.flink.table.descriptors.StatisticsValidator.{STATISTICS_COLUMNS, STATISTICS_ROW_COUNT, readColumnStats} +import org.apache.flink.table.plan.stats.TableStats + +import scala.collection.JavaConverters._ + +/** + * Common class for all descriptors describing table sources and sinks. + */ +abstract class TableDescriptor extends Descriptor { + + protected var connectorDescriptor: Option[ConnectorDescriptor] = None + protected var formatDescriptor: Option[FormatDescriptor] = None + protected var schemaDescriptor: Option[Schema] = None + protected var statisticsDescriptor: Option[Statistics] = None --- End diff -- I was wondering if a sink could have its own "schema configurations" for alignment with the output table schema? For example a CassandraTableSink / JDBCTableSink would definitely throw exceptions when trying to execute an insert with mismatched schemas. > Add support for unified table source and sink declaration in environment file > - > > Key: FLINK-9059 > URL: https://issues.apache.org/jira/browse/FLINK-9059 > Project: Flink > Issue Type: Task > Components: Table API & SQL >Reporter: Shuyi Chen >Assignee: Shuyi Chen >Priority: Major > Fix For: 1.5.0 > > > 1) Add a common property called "type" with single value 'source'. > 2) in yaml file, replace "sources" with "tables". -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5758: [FLINK-9059][Table API & SQL] add table type attri...
Github user walterddr commented on a diff in the pull request: https://github.com/apache/flink/pull/5758#discussion_r179178156 --- Diff: flink-libraries/flink-table/src/main/scala/org/apache/flink/table/descriptors/TableDescriptor.scala --- @@ -0,0 +1,75 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.table.descriptors + +import org.apache.flink.table.descriptors.DescriptorProperties.toScala +import org.apache.flink.table.descriptors.StatisticsValidator.{STATISTICS_COLUMNS, STATISTICS_ROW_COUNT, readColumnStats} +import org.apache.flink.table.plan.stats.TableStats + +import scala.collection.JavaConverters._ + +/** + * Common class for all descriptors describing table sources and sinks. + */ +abstract class TableDescriptor extends Descriptor { + + protected var connectorDescriptor: Option[ConnectorDescriptor] = None + protected var formatDescriptor: Option[FormatDescriptor] = None + protected var schemaDescriptor: Option[Schema] = None + protected var statisticsDescriptor: Option[Statistics] = None --- End diff -- I was wondering if a sink could have its own "schema configurations" for alignment with the output table schema? For example a CassandraTableSink / JDBCTableSink would definitely throw exceptions when trying to execute an insert with mismatched schemas. ---
[jira] [Resolved] (FLINK-8509) Remove SqlGroupedWindowFunction from Flink repo
[ https://issues.apache.org/jira/browse/FLINK-8509?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Timo Walther resolved FLINK-8509. - Resolution: Fixed Fix Version/s: 1.5.0 Fixed in 1.6.0: 98a8b642f24c813a4929dfd780163778dc5bd010 Fixed in 1.5.0: a298e6e4f81ec2cab4c86dd07f441eaebada4915 > Remove SqlGroupedWindowFunction from Flink repo > --- > > Key: FLINK-8509 > URL: https://issues.apache.org/jira/browse/FLINK-8509 > Project: Flink > Issue Type: Task > Components: Table API & SQL >Reporter: Shuyi Chen >Assignee: Shuyi Chen >Priority: Major > Fix For: 1.5.0 > > > SqlGroupedWindowFunction is copied to the Flink repo due to > [CALCITE-2133|https://issues.apache.org/jira/browse/CALCITE-2133], we should > remove it once flink upgrade Calcite dependency to 1.16. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (FLINK-9133) Improve documentability of REST API
Chesnay Schepler created FLINK-9133: --- Summary: Improve documentability of REST API Key: FLINK-9133 URL: https://issues.apache.org/jira/browse/FLINK-9133 Project: Flink Issue Type: Improvement Components: Documentation, REST Affects Versions: 1.5.0 Reporter: Chesnay Schepler Assignee: Chesnay Schepler In several places of the REST API we use custom JSON (de)serializers for writing data. This is very problematic in regards to the documentation, as there is no way to actually generated it when these serializers are used. I doubt we can fix this issue entirely at the moment, but I already found areas that we can improve. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5794: [Flink-8509][Table API & SQL] Remove SqlGroupedWin...
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5794 ---
[jira] [Created] (FLINK-9132) Cluster runs out of task slots when a job falls into restart loop
Alex Smirnov created FLINK-9132: --- Summary: Cluster runs out of task slots when a job falls into restart loop Key: FLINK-9132 URL: https://issues.apache.org/jira/browse/FLINK-9132 Project: Flink Issue Type: Bug Affects Versions: 1.4.2 Environment: env.java.opts in flink-conf.yaml file: env.java.opts: -Xloggc:/home/user/flink/log/flinkServer-gc.log -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseG1GC -XX:MaxGCPauseMillis=150 -XX:InitiatingHeapOccupancyPercent=55 -XX:+ParallelRefProcEnabled -XX:ParallelGCThreads=2 -XX:-ResizePLAB -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=5 -XX:GCLogFileSize=100M Reporter: Alex Smirnov Attachments: FailedJob.java, jconsole-classes.png If there's a job which is restarting in a loop, then Task Manager hosting it goes down after some time. Job manager automatically assigns the job to another Task Manager and the new Task Manager goes down as well. After some time, all Task Managers are gone. Cluster becomes paralyzed. I've attached to TaskManager's java process using jconsole and noticed that number of loaded classes increases dramatically if a job is in restarting loop and restores from checkpoint. See attachment for the graph with G1GC enabled for the node. Standard GC performs even worse - task manager shuts down within 20 minutes since the restart loop start. I've also attached minimal program to reproduce the problem please let me know if additional information is required from me. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Comment Edited] (FLINK-9008) End-to-end test: Quickstarts
[ https://issues.apache.org/jira/browse/FLINK-9008?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425648#comment-16425648 ] mingleizhang edited comment on FLINK-9008 at 4/4/18 2:54 PM: - Hi, [~till.rohrmann] I would like to confirm one stuff with you here for the example of a flink application program that will package into a jar file. I know there is already an example of {{PopularPlacesToES}}, should I package this into a jar or instead I can write a simple job which support only one operator like {{filter}} into that jar. Any suggestions ? Thank you very much. was (Author: mingleizhang): Hi, [~till.rohrmann] I would like to confirm one stuff with you here for the example of a flink application program that will package into a jar file. I know there is already an example of {{PopularPlacesToES}}, should I package this into a jar or instead I can write a simple job which support only one operator like {{filter}}. Any suggestions ? Thank you very much. > End-to-end test: Quickstarts > > > Key: FLINK-9008 > URL: https://issues.apache.org/jira/browse/FLINK-9008 > Project: Flink > Issue Type: Sub-task > Components: Quickstarts, Tests >Affects Versions: 1.5.0 >Reporter: Till Rohrmann >Assignee: mingleizhang >Priority: Critical > Fix For: 1.5.0 > > > We could add an end-to-end test which verifies Flink's quickstarts. It should > do the following: > # create a new Flink project using the quickstarts archetype > # add a new Flink dependency to the {{pom.xml}} (e.g. Flink connector or > library) > # run {{mvn clean package -Pbuild-jar}} > # verify that no core dependencies are contained in the jar file > # Run the program -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9008) End-to-end test: Quickstarts
[ https://issues.apache.org/jira/browse/FLINK-9008?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425648#comment-16425648 ] mingleizhang commented on FLINK-9008: - Hi, [~till.rohrmann] I would like to confirm one stuff with you here for the example of a flink application program that will package into a jar file. I know there is already an example of {{PopularPlacesToES}}, should I package this into a jar or instead I can write a simple job which support only one operator like {{filter}}. Any suggestions ? Thank you very much. > End-to-end test: Quickstarts > > > Key: FLINK-9008 > URL: https://issues.apache.org/jira/browse/FLINK-9008 > Project: Flink > Issue Type: Sub-task > Components: Quickstarts, Tests >Affects Versions: 1.5.0 >Reporter: Till Rohrmann >Assignee: mingleizhang >Priority: Critical > Fix For: 1.5.0 > > > We could add an end-to-end test which verifies Flink's quickstarts. It should > do the following: > # create a new Flink project using the quickstarts archetype > # add a new Flink dependency to the {{pom.xml}} (e.g. Flink connector or > library) > # run {{mvn clean package -Pbuild-jar}} > # verify that no core dependencies are contained in the jar file > # Run the program -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (FLINK-8563) Support consecutive DOT operators
[ https://issues.apache.org/jira/browse/FLINK-8563?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Timo Walther resolved FLINK-8563. - Resolution: Fixed Fix Version/s: 1.5.0 Fixed in 1.6.0: 284995172bad3cbc5844d8198654e1de7f513591 Fixed in 1.5.0: 4b10dd684b2ee8a1b74f1297b79fe0852f0172e5 > Support consecutive DOT operators > -- > > Key: FLINK-8563 > URL: https://issues.apache.org/jira/browse/FLINK-8563 > Project: Flink > Issue Type: Sub-task > Components: Table API & SQL >Reporter: Timo Walther >Assignee: Shuyi Chen >Priority: Major > Fix For: 1.5.0 > > > We added support for accessing fields of arrays of composite types in > FLINK-7923. However, accessing another nested subfield is not supported by > Calcite. See CALCITE-2162. We should fix this once we upgrade to Calcite 1.16. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9131) Disable spotbugs on travis
[ https://issues.apache.org/jira/browse/FLINK-9131?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425641#comment-16425641 ] ASF GitHub Bot commented on FLINK-9131: --- GitHub user zentol opened a pull request: https://github.com/apache/flink/pull/5815 [FLINK-9131][travis] Disable spotbugs plugin ## What is the purpose of the change This PR disables spotbugs on travis. This will buy us 7-8 minutes in the misc profile that currently times out, You can merge this pull request into a Git repository by running: $ git pull https://github.com/zentol/flink 9131 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5815.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5815 commit 521462ad59dcfea7f12d2d3827846028154c8a80 Author: zentol Date: 2018-04-04T14:43:14Z [FLINK-9131][travis] Disable spotbugs plugin > Disable spotbugs on travis > -- > > Key: FLINK-9131 > URL: https://issues.apache.org/jira/browse/FLINK-9131 > Project: Flink > Issue Type: Improvement > Components: Travis >Affects Versions: 1.5.0 >Reporter: Chesnay Schepler >Assignee: Chesnay Schepler >Priority: Critical > > The misc profile that also runs spotbugs is consistently timing out on travis > at the moment. > The spotbugs plugin is a major contributor to the compilation time, for > example it doubles the compile time for flink-runtime. > I suggest to temporarily disable spotbugs, and re-enable it at a lter point > when we figure out the daily cron jobs. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-9108) invalid ProcessWindowFunction link in Document
[ https://issues.apache.org/jira/browse/FLINK-9108?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler closed FLINK-9108. --- Resolution: Fixed Fix Version/s: 1.5.0 master: a71b9031821b80b74df855df4a565467bb32550a 1.5: 198446962c67a297cf98d45b3efa94ee56d1dd7a > invalid ProcessWindowFunction link in Document > --- > > Key: FLINK-9108 > URL: https://issues.apache.org/jira/browse/FLINK-9108 > Project: Flink > Issue Type: Bug > Components: Documentation >Reporter: Matrix42 >Assignee: Matrix42 >Priority: Trivial > Fix For: 1.5.0 > > Attachments: QQ截图20180329184203.png > > > !QQ截图20180329184203.png! -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5815: [FLINK-9131][travis] Disable spotbugs plugin
GitHub user zentol opened a pull request: https://github.com/apache/flink/pull/5815 [FLINK-9131][travis] Disable spotbugs plugin ## What is the purpose of the change This PR disables spotbugs on travis. This will buy us 7-8 minutes in the misc profile that currently times out, You can merge this pull request into a Git repository by running: $ git pull https://github.com/zentol/flink 9131 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5815.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5815 commit 521462ad59dcfea7f12d2d3827846028154c8a80 Author: zentol Date: 2018-04-04T14:43:14Z [FLINK-9131][travis] Disable spotbugs plugin ---
[GitHub] flink pull request #5792: [Flink-8563][Table API & SQL] add unittest for con...
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5792 ---
[jira] [Commented] (FLINK-9108) invalid ProcessWindowFunction link in Document
[ https://issues.apache.org/jira/browse/FLINK-9108?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425636#comment-16425636 ] ASF GitHub Bot commented on FLINK-9108: --- Github user zentol commented on the issue: https://github.com/apache/flink/pull/5785 ehh...whoops. Yes i closed it by accident, thanks for catching it. I'll merge the commit in a second... > invalid ProcessWindowFunction link in Document > --- > > Key: FLINK-9108 > URL: https://issues.apache.org/jira/browse/FLINK-9108 > Project: Flink > Issue Type: Bug > Components: Documentation >Reporter: Matrix42 >Assignee: Matrix42 >Priority: Trivial > Attachments: QQ截图20180329184203.png > > > !QQ截图20180329184203.png! -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink issue #5785: [FLINK-9108][docs] Fix invalid link
Github user zentol commented on the issue: https://github.com/apache/flink/pull/5785 ehh...whoops. Yes i closed it by accident, thanks for catching it. I'll merge the commit in a second... ---
[jira] [Created] (FLINK-9131) Disable spotbugs on travis
Chesnay Schepler created FLINK-9131: --- Summary: Disable spotbugs on travis Key: FLINK-9131 URL: https://issues.apache.org/jira/browse/FLINK-9131 Project: Flink Issue Type: Improvement Components: Travis Affects Versions: 1.5.0 Reporter: Chesnay Schepler Assignee: Chesnay Schepler The misc profile that also runs spotbugs is consistently timing out on travis at the moment. The spotbugs plugin is a major contributor to the compilation time, for example it doubles the compile time for flink-runtime. I suggest to temporarily disable spotbugs, and re-enable it at a lter point when we figure out the daily cron jobs. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink issue #5792: [Flink-8563][Table API & SQL] add unittest for consecutiv...
Github user twalthr commented on the issue: https://github.com/apache/flink/pull/5792 Thank you @suez1224. Merging... ---
[jira] [Commented] (FLINK-9130) Add cancel-job option to SavepointHandlers JavaDoc
[ https://issues.apache.org/jira/browse/FLINK-9130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425626#comment-16425626 ] Gary Yao commented on FLINK-9130: - [~fhueske], after FLINK-9104 the FLIP-6 REST API docs should include the option. > Add cancel-job option to SavepointHandlers JavaDoc > -- > > Key: FLINK-9130 > URL: https://issues.apache.org/jira/browse/FLINK-9130 > Project: Flink > Issue Type: Improvement > Components: Documentation >Affects Versions: 1.5.0, 1.6.0 >Reporter: Fabian Hueske >Priority: Minor > > The Savepoint JavaDocs are missing the {{cancel-job}} option. > See discussion on ML here: > [https://lists.apache.org/thread.html/dc05751fa6507388dcefc0c845facef6b36b086e256b52c3e37d71dc@%3Cuser.flink.apache.org%3E] -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (FLINK-8507) Upgrade Calcite dependency to 1.16
[ https://issues.apache.org/jira/browse/FLINK-8507?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Timo Walther resolved FLINK-8507. - Resolution: Fixed Fix Version/s: 1.5.0 Fixed in 1.6.0: 3e21f0f8ea1d1e5d6c4a34fda1f8fa821ffd6a40 Fixed in 1.5.0: 7e240edca02d03648ccec471f141bca70b8c79db > Upgrade Calcite dependency to 1.16 > -- > > Key: FLINK-8507 > URL: https://issues.apache.org/jira/browse/FLINK-8507 > Project: Flink > Issue Type: Task > Components: Table API & SQL >Reporter: Shuyi Chen >Assignee: Shuyi Chen >Priority: Major > Fix For: 1.5.0 > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-8507) Upgrade Calcite dependency to 1.16
[ https://issues.apache.org/jira/browse/FLINK-8507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425623#comment-16425623 ] ASF GitHub Bot commented on FLINK-8507: --- Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5791 > Upgrade Calcite dependency to 1.16 > -- > > Key: FLINK-8507 > URL: https://issues.apache.org/jira/browse/FLINK-8507 > Project: Flink > Issue Type: Task > Components: Table API & SQL >Reporter: Shuyi Chen >Assignee: Shuyi Chen >Priority: Major > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5791: [FLINK-8507][Table API & SQL] upgrade calcite depe...
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5791 ---
[jira] [Commented] (FLINK-9108) invalid ProcessWindowFunction link in Document
[ https://issues.apache.org/jira/browse/FLINK-9108?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425620#comment-16425620 ] ASF GitHub Bot commented on FLINK-9108: --- Github user Matrix42 commented on the issue: https://github.com/apache/flink/pull/5785 @zentol Is this closed by accident? > invalid ProcessWindowFunction link in Document > --- > > Key: FLINK-9108 > URL: https://issues.apache.org/jira/browse/FLINK-9108 > Project: Flink > Issue Type: Bug > Components: Documentation >Reporter: Matrix42 >Assignee: Matrix42 >Priority: Trivial > Attachments: QQ截图20180329184203.png > > > !QQ截图20180329184203.png! -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink issue #5785: [FLINK-9108][docs] Fix invalid link
Github user Matrix42 commented on the issue: https://github.com/apache/flink/pull/5785 @zentol Is this closed by accident? ---
[jira] [Updated] (FLINK-9130) Add cancel-job option to SavepointHandlers JavaDoc
[ https://issues.apache.org/jira/browse/FLINK-9130?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fabian Hueske updated FLINK-9130: - Priority: Minor (was: Critical) > Add cancel-job option to SavepointHandlers JavaDoc > -- > > Key: FLINK-9130 > URL: https://issues.apache.org/jira/browse/FLINK-9130 > Project: Flink > Issue Type: Improvement > Components: Documentation >Affects Versions: 1.5.0, 1.6.0 >Reporter: Fabian Hueske >Priority: Minor > > The Savepoint JavaDocs are missing the {{cancel-job}} option. > See discussion on ML here: > [https://lists.apache.org/thread.html/dc05751fa6507388dcefc0c845facef6b36b086e256b52c3e37d71dc@%3Cuser.flink.apache.org%3E] -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9130) Add cancel-job option to SavepointHandlers JavaDoc
[ https://issues.apache.org/jira/browse/FLINK-9130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425612#comment-16425612 ] Fabian Hueske commented on FLINK-9130: -- Thanks for the correction [~gjy]. I've changed the Jira title and description. Does the REST API documentation include the {{cancel-job}} option? > Add cancel-job option to SavepointHandlers JavaDoc > -- > > Key: FLINK-9130 > URL: https://issues.apache.org/jira/browse/FLINK-9130 > Project: Flink > Issue Type: Improvement > Components: Documentation >Affects Versions: 1.5.0, 1.6.0 >Reporter: Fabian Hueske >Priority: Critical > > The Savepoint JavaDocs are missing the {{cancel-job}} option. > See discussion on ML here: > [https://lists.apache.org/thread.html/dc05751fa6507388dcefc0c845facef6b36b086e256b52c3e37d71dc@%3Cuser.flink.apache.org%3E] -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-9130) Add cancel-job option to SavepointHandlers JavaDoc
[ https://issues.apache.org/jira/browse/FLINK-9130?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fabian Hueske updated FLINK-9130: - Description: The Savepoint JavaDocs are missing the {{cancel-job}} option. See discussion on ML here: [https://lists.apache.org/thread.html/dc05751fa6507388dcefc0c845facef6b36b086e256b52c3e37d71dc@%3Cuser.flink.apache.org%3E] was: The Savepoint REST documentation is missing the {{cancel-job}} option. See discussion on ML here: https://lists.apache.org/thread.html/dc05751fa6507388dcefc0c845facef6b36b086e256b52c3e37d71dc@%3Cuser.flink.apache.org%3E > Add cancel-job option to SavepointHandlers JavaDoc > -- > > Key: FLINK-9130 > URL: https://issues.apache.org/jira/browse/FLINK-9130 > Project: Flink > Issue Type: Improvement > Components: Documentation >Affects Versions: 1.5.0, 1.6.0 >Reporter: Fabian Hueske >Priority: Critical > > The Savepoint JavaDocs are missing the {{cancel-job}} option. > See discussion on ML here: > [https://lists.apache.org/thread.html/dc05751fa6507388dcefc0c845facef6b36b086e256b52c3e37d71dc@%3Cuser.flink.apache.org%3E] -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (FLINK-9130) Add cancel-job option to SavepointHandlers JavaDoc
[ https://issues.apache.org/jira/browse/FLINK-9130?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fabian Hueske updated FLINK-9130: - Summary: Add cancel-job option to SavepointHandlers JavaDoc (was: Add cancel-job option to SavepointHandlers JavaDoc and regenerate REST docs) > Add cancel-job option to SavepointHandlers JavaDoc > -- > > Key: FLINK-9130 > URL: https://issues.apache.org/jira/browse/FLINK-9130 > Project: Flink > Issue Type: Improvement > Components: Documentation >Affects Versions: 1.5.0, 1.6.0 >Reporter: Fabian Hueske >Priority: Critical > > The Savepoint REST documentation is missing the {{cancel-job}} option. > See discussion on ML here: > https://lists.apache.org/thread.html/dc05751fa6507388dcefc0c845facef6b36b086e256b52c3e37d71dc@%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9130) Add cancel-job option to SavepointHandlers JavaDoc and regenerate REST docs
[ https://issues.apache.org/jira/browse/FLINK-9130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425603#comment-16425603 ] Gary Yao commented on FLINK-9130: - The Javadocs for {{org.apache.flink.runtime.rest.handler.job.savepoints.SavepointHandlers}} is missing the {{cancel-job}} option. The REST documentation is not generated from Javadocs. > Add cancel-job option to SavepointHandlers JavaDoc and regenerate REST docs > --- > > Key: FLINK-9130 > URL: https://issues.apache.org/jira/browse/FLINK-9130 > Project: Flink > Issue Type: Improvement > Components: Documentation >Affects Versions: 1.5.0, 1.6.0 >Reporter: Fabian Hueske >Priority: Critical > > The Savepoint REST documentation is missing the {{cancel-job}} option. > See discussion on ML here: > https://lists.apache.org/thread.html/dc05751fa6507388dcefc0c845facef6b36b086e256b52c3e37d71dc@%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-7726) Move marshalling testbases out of legacy namespace
[ https://issues.apache.org/jira/browse/FLINK-7726?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler closed FLINK-7726. --- Resolution: Fixed Fix Version/s: (was: 1.5.0) 1.4.0 master: bc4638a3c96049de3ef615159cf83bbd88019575 > Move marshalling testbases out of legacy namespace > -- > > Key: FLINK-7726 > URL: https://issues.apache.org/jira/browse/FLINK-7726 > Project: Flink > Issue Type: Improvement > Components: REST, Tests >Affects Versions: 1.4.0 >Reporter: Chesnay Schepler >Assignee: Chesnay Schepler >Priority: Major > Fix For: 1.4.0 > > > The marshalling test bases currently reside under > {{org.apache.flink.runtime.rest.handler.legacy.messages}} which doesn't make > sense as this isn't legacy code. > We should do this once the port of all handlers has been finished to avoid > merge conflicts. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Closed] (FLINK-8254) REST API documentation wonky due to shading
[ https://issues.apache.org/jira/browse/FLINK-8254?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler closed FLINK-8254. --- Resolution: Fixed master: 37fe082bd594c5c092362077e0329f138c84e544 1.5: 69b9515b47b11be70dfb4d6176883c91086609b3 > REST API documentation wonky due to shading > --- > > Key: FLINK-8254 > URL: https://issues.apache.org/jira/browse/FLINK-8254 > Project: Flink > Issue Type: Improvement > Components: Documentation >Affects Versions: 1.5.0 >Reporter: Chesnay Schepler >Assignee: Chesnay Schepler >Priority: Major > Fix For: 1.5.0 > > > The REST API documentation isn't quite correct as all jackson annotations are > being ignored. Our annotations come from flink-shaded-jackson, but the tool > we use (jackson-module-jsonSchema) checks against vanilla jackson. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-8507) Upgrade Calcite dependency to 1.16
[ https://issues.apache.org/jira/browse/FLINK-8507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425569#comment-16425569 ] ASF GitHub Bot commented on FLINK-8507: --- Github user twalthr commented on the issue: https://github.com/apache/flink/pull/5791 LGTM, will merge in next batch... > Upgrade Calcite dependency to 1.16 > -- > > Key: FLINK-8507 > URL: https://issues.apache.org/jira/browse/FLINK-8507 > Project: Flink > Issue Type: Task > Components: Table API & SQL >Reporter: Shuyi Chen >Assignee: Shuyi Chen >Priority: Major > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink issue #5791: [FLINK-8507][Table API & SQL] upgrade calcite dependency ...
Github user twalthr commented on the issue: https://github.com/apache/flink/pull/5791 LGTM, will merge in next batch... ---
[jira] [Created] (FLINK-9130) Add cancel-job option to SavepointHandlers JavaDoc and regenerate REST docs
Fabian Hueske created FLINK-9130: Summary: Add cancel-job option to SavepointHandlers JavaDoc and regenerate REST docs Key: FLINK-9130 URL: https://issues.apache.org/jira/browse/FLINK-9130 Project: Flink Issue Type: Improvement Components: Documentation Affects Versions: 1.5.0, 1.6.0 Reporter: Fabian Hueske The Savepoint REST documentation is missing the {{cancel-job}} option. See discussion on ML here: https://lists.apache.org/thread.html/dc05751fa6507388dcefc0c845facef6b36b086e256b52c3e37d71dc@%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (FLINK-9059) Add support for unified table source and sink declaration in environment file
[ https://issues.apache.org/jira/browse/FLINK-9059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425558#comment-16425558 ] ASF GitHub Bot commented on FLINK-9059: --- Github user twalthr commented on a diff in the pull request: https://github.com/apache/flink/pull/5758#discussion_r179149794 --- Diff: flink-libraries/flink-sql-client/src/main/java/org/apache/flink/table/client/config/Environment.java --- @@ -29,38 +30,47 @@ /** * Environment configuration that represents the content of an environment file. Environment files - * define sources, execution, and deployment behavior. An environment might be defined by default or + * define tables, execution, and deployment behavior. An environment might be defined by default or * as part of a session. Environments can be merged or enriched with properties (e.g. from CLI command). * * In future versions, we might restrict the merging or enrichment of deployment properties to not * allow overwriting of a deployment by a session. */ public class Environment { - private Map sources; + private Map tables; --- End diff -- Why not maintaining two separate maps for sources and sinks? Then we don't need instance of checks. If a table is both we can simply add it to both maps. > Add support for unified table source and sink declaration in environment file > - > > Key: FLINK-9059 > URL: https://issues.apache.org/jira/browse/FLINK-9059 > Project: Flink > Issue Type: Task > Components: Table API & SQL >Reporter: Shuyi Chen >Assignee: Shuyi Chen >Priority: Major > Fix For: 1.5.0 > > > 1) Add a common property called "type" with single value 'source'. > 2) in yaml file, replace "sources" with "tables". -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5758: [FLINK-9059][Table API & SQL] add table type attri...
Github user twalthr commented on a diff in the pull request: https://github.com/apache/flink/pull/5758#discussion_r179149794 --- Diff: flink-libraries/flink-sql-client/src/main/java/org/apache/flink/table/client/config/Environment.java --- @@ -29,38 +30,47 @@ /** * Environment configuration that represents the content of an environment file. Environment files - * define sources, execution, and deployment behavior. An environment might be defined by default or + * define tables, execution, and deployment behavior. An environment might be defined by default or * as part of a session. Environments can be merged or enriched with properties (e.g. from CLI command). * * In future versions, we might restrict the merging or enrichment of deployment properties to not * allow overwriting of a deployment by a session. */ public class Environment { - private Map sources; + private Map tables; --- End diff -- Why not maintaining two separate maps for sources and sinks? Then we don't need instance of checks. If a table is both we can simply add it to both maps. ---
[jira] [Commented] (FLINK-9059) Add support for unified table source and sink declaration in environment file
[ https://issues.apache.org/jira/browse/FLINK-9059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16425559#comment-16425559 ] ASF GitHub Bot commented on FLINK-9059: --- Github user twalthr commented on a diff in the pull request: https://github.com/apache/flink/pull/5758#discussion_r179150731 --- Diff: flink-libraries/flink-table/src/main/scala/org/apache/flink/table/descriptors/TableDescriptor.scala --- @@ -0,0 +1,75 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.table.descriptors + +import org.apache.flink.table.descriptors.DescriptorProperties.toScala +import org.apache.flink.table.descriptors.StatisticsValidator.{STATISTICS_COLUMNS, STATISTICS_ROW_COUNT, readColumnStats} +import org.apache.flink.table.plan.stats.TableStats + +import scala.collection.JavaConverters._ + +/** + * Common class for all descriptors describing table sources and sinks. + */ +abstract class TableDescriptor extends Descriptor { + + protected var connectorDescriptor: Option[ConnectorDescriptor] = None + protected var formatDescriptor: Option[FormatDescriptor] = None + protected var schemaDescriptor: Option[Schema] = None + protected var statisticsDescriptor: Option[Statistics] = None --- End diff -- I'm wondering if we really need these changes. A sink will never have a schema or statistics. > Add support for unified table source and sink declaration in environment file > - > > Key: FLINK-9059 > URL: https://issues.apache.org/jira/browse/FLINK-9059 > Project: Flink > Issue Type: Task > Components: Table API & SQL >Reporter: Shuyi Chen >Assignee: Shuyi Chen >Priority: Major > Fix For: 1.5.0 > > > 1) Add a common property called "type" with single value 'source'. > 2) in yaml file, replace "sources" with "tables". -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[GitHub] flink pull request #5758: [FLINK-9059][Table API & SQL] add table type attri...
Github user twalthr commented on a diff in the pull request: https://github.com/apache/flink/pull/5758#discussion_r179150731 --- Diff: flink-libraries/flink-table/src/main/scala/org/apache/flink/table/descriptors/TableDescriptor.scala --- @@ -0,0 +1,75 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.table.descriptors + +import org.apache.flink.table.descriptors.DescriptorProperties.toScala +import org.apache.flink.table.descriptors.StatisticsValidator.{STATISTICS_COLUMNS, STATISTICS_ROW_COUNT, readColumnStats} +import org.apache.flink.table.plan.stats.TableStats + +import scala.collection.JavaConverters._ + +/** + * Common class for all descriptors describing table sources and sinks. + */ +abstract class TableDescriptor extends Descriptor { + + protected var connectorDescriptor: Option[ConnectorDescriptor] = None + protected var formatDescriptor: Option[FormatDescriptor] = None + protected var schemaDescriptor: Option[Schema] = None + protected var statisticsDescriptor: Option[Statistics] = None --- End diff -- I'm wondering if we really need these changes. A sink will never have a schema or statistics. ---
[GitHub] flink pull request #5812: [FLINK-9128] [flip6] Add support for scheduleRunAs...
Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/5812 ---