GitHub user paulzwu opened a pull request: https://github.com/apache/flink/pull/5247
[FLINK-8356] Need to add the commit in flush() in JDBCOutputFormat *Thank you very much for contributing to Apache Flink - we are happy that you want to help us improve Flink. To help the community review your contribution in the best possible way, please go through the checklist below, which will get the contribution into a shape in which it can be best reviewed.* *Please understand that we do not do this to make contributions to Flink a hassle. In order to uphold a high standard of quality for code contributions, while at the same time managing a large number of contributions, we need contributors to prepare the contributions well, and give reviewers enough contextual information for the review. Please also understand that contributions that do not follow this guide will take longer to review and thus typically be picked up with lower priority by the community.* ## Contribution Checklist - Make sure that the pull request corresponds to a [JIRA issue](https://issues.apache.org/jira/projects/FLINK/issues). Exceptions are made for typos in JavaDoc or documentation files, which need no JIRA issue. - Name the pull request in the form "[FLINK-XXXX] [component] Title of the pull request", where *FLINK-XXXX* should be replaced by the actual issue number. Skip *component* if you are unsure about which is the best component. Typo fixes that have no associated JIRA issue should be named following this pattern: `[hotfix] [docs] Fix typo in event time introduction` or `[hotfix] [javadocs] Expand JavaDoc for PuncuatedWatermarkGenerator`. - Fill out the template below to describe the changes contributed by the pull request. That will give reviewers the context they need to do the review. - Make sure that the change passes the automated tests, i.e., `mvn clean verify` passes. You can set up Travis CI to do that following [this guide](http://flink.apache.org/contribute-code.html#best-practices). - Each pull request should address only one issue, not mix up code from multiple issues. - Each commit in the pull request has a meaningful commit message (including the JIRA id) - Once all items of the checklist are addressed, remove the above text and this checklist, leaving only the filled out template below. **(The sections below can be removed for hotfixes of typos)** ## What is the purpose of the change *(For example: This pull request makes task deployment go through the blob server, rather than through RPC. That way we avoid re-transferring them on each deployment (during recovery).)* ## Brief change log *(for example:)* - *The TaskInfo is stored in the blob store on job creation time as a persistent artifact* - *Deployments RPC transmits only the blob storage reference* - *TaskManagers retrieve the TaskInfo from the blob cache* ## Verifying this change *(Please pick either of the following options)* This change is a trivial rework / code cleanup without any test coverage. *(or)* This change is already covered by existing tests, such as *(please describe tests)*. *(or)* This change added tests and can be verified as follows: *(example:)* - *Added integration tests for end-to-end deployment with large payloads (100MB)* - *Extended integration test for recovery after master (JobManager) failure* - *Added test that validates that TaskInfo is transferred only once across recoveries* - *Manually verified the change by running a 4 node cluser with 2 JobManagers and 4 TaskManagers, a stateful streaming program, and killing one JobManager and two TaskManagers during the execution, verifying that recovery happens correctly.* ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (yes / no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (yes / no) - The serializers: (yes / no / don't know) - The runtime per-record code paths (performance sensitive): (yes / no / don't know) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: (yes / no / don't know) - The S3 file system connector: (yes / no / don't know) ## Documentation - Does this pull request introduce a new feature? (yes / no) - If yes, how is the feature documented? (not applicable / docs / JavaDocs / not documented) You can merge this pull request into a Git repository by running: $ git pull https://github.com/apache/flink master Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/5247.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #5247 ---- commit d98ba08a7a73dc93c839c77d11245aba70869ab5 Author: Aljoscha Krettek <aljoscha.krettek@...> Date: 2017-11-10T09:54:16Z [FLINK-6163] Document per-window state in ProcessWindowFunction commit 1cd3ba3f2af454bc33f2c880163c01dddd4d1738 Author: Piotr Nowojski <piotr.nowojski@...> Date: 2017-11-10T14:15:11Z [hotfix][docs] Change mailing list link in quickstart to flink-user Previously it was pointing to flink-dev commit 07af8c125bba88912dd39664532b1ec8ee035439 Author: Aljoscha Krettek <aljoscha.krettek@...> Date: 2017-11-10T15:28:46Z [FLINK-7702] Remove Javadoc aggregation for Scala code genjavadoc generated some Java code that was making Javadoc fail. commit 2929eda7e244c1cd7250645702a808af967fae84 Author: Aljoscha Krettek <aljoscha.krettek@...> Date: 2017-11-10T17:13:26Z [FLINK-7702] Add maven-bundle-plugin to root pom Before, we had it in places that require it. This doesn't work when running mvn javadoc:aggregate because this will only run for the root pom and can then not find the "bundle" dependencies. commit ad8ef6d01b23c74fcfb1f39a151dd00bfbd78ca3 Author: Stefan Richter <s.richter@...> Date: 2017-11-13T10:50:07Z [FLINK-8040] [tests] Fix test instability in ResourceGuardTest This closes #5004. commit 3b333b289cbbe43f722edef2c36c370ff4550128 Author: Kent Murra <kentm@...> Date: 2017-09-27T20:48:55Z [FLINK-7657] [table] Add time types FilterableTableSource push down This closes #4746. commit 32e5194d95742dc0115a2b4d8c34e3f025feebd0 Author: twalthr <twalthr@...> Date: 2017-11-13T13:09:45Z [FLINK-7657] [table] Add all basic types to RexProgramExtractor commit 0e5fb0b78cd0a3ccb144071a47579eb6c3d0570a Author: Nico Kruber <nico@...> Date: 2017-11-06T18:53:37Z [FLINK-7973] Fix shading and relocating Hadoop for the S3 filesystems - do not shade everything, especially not JDK classes! -> instead define include patterns explicitly - do not shade core Flink classes (only those imported from flink-hadoop-fs) - hack around Hadoop loading (unshaded/non-relocated) classes based on names in the core-default.xml by overwriting the Configuration class (we may need to extend this for the mapred-default.xml and hdfs-defaults.xml): -> provide a core-default-shaded.xml file with shaded class names and copy and adapt the Configuration class of the respective Hadoop version to load this file instead of core-default.xml. Add checkstyle suppression pattern for the Hadoop Configuration classes Also fix the (integration) tests not working because they tried to load the relocated classes which are apparently not available there Remove minimizeJar from shading of flink-s3-fs-presto because this was causing "java.lang.ClassNotFoundException: org.apache.flink.fs.s3presto.shaded.org.apache.commons.logging.impl.LogFactoryImpl" since these classes are not statically imported and thus removed when minimizing. Fix s3-fs-presto not shading org.HdrHistogram Fix log4j being relocated in the S3 fs implementations Add shading checks to travis commit e9e7c3372189db7e933ff59114b9ec6245838eda Author: Aljoscha Krettek <aljoscha.krettek@...> Date: 2017-11-08T16:49:17Z [FLINK-7973] Add shaded S3 FileSystem end-to-end tests commit 8fa6cf947d0b00095a0e35b3d4545239e340103e Author: Aljoscha Krettek <aljoscha.krettek@...> Date: 2017-11-08T17:57:22Z [hotfix] Make end-to-end test scripts more robust This uses traps to ensure that we properly do cleanups, remove config values and shutdown things. commit 613fc445f0a2669672c90f0718c5fcfe6822b15c Author: Nico Kruber <nico@...> Date: 2017-11-10T12:16:40Z [hotfix] fix presto end-to-end test not cleaning up commit ab2b9f8f9629cf33135b923fd81ee28a4ea8b8bd Author: Nico Kruber <nico@...> Date: 2017-11-10T13:04:29Z [hotfix] ignore a warning from the error check of the S3 e2e tests commit d139e634087aaba7db43429d803d12203b7bc18a Author: Nico Kruber <nico@...> Date: 2017-11-10T15:42:10Z [hotfix] let end-to-end tests check for empty .out files again commit 2906698b4a87f21c6fd099cf8a028f68fc311b1f Author: Stefan Richter <s.richter@...> Date: 2017-11-13T13:31:45Z [FLINK-8053] [checkpoints] Default to asynchronous snapshots for FsStateBackend and MemoryStateBackend. This closes #5005. commit b98a4aa8b8ba1881f938125d0765620a4289a3c7 Author: Alejandro Alcalde <algui91@...> Date: 2017-11-07T12:09:46Z [FLINK-8006] [Startup Shell Scripts] - Fixing $pid This closes #4968. commit 119006752e190334fbf4f90fa53f6dfa9374e81b Author: zentol <chesnay@...> Date: 2017-11-07T16:13:57Z [FLINK-8011][dist] Set flink-python to provided This closes #4973. commit 3f80f8ef334959392ac4a22b064fb6c8b5378462 Author: gyao <gary@...> Date: 2017-11-10T13:54:32Z [hotfix][docs] Fix broken link to FLINK-7811 This closes #4995. commit 49c8686b81eb7cb13912b4fbb6f468840f1cc7ad Author: yew1eb <yew1eb@...> Date: 2017-11-11T12:45:42Z [hotfix][docs][javadocs] Remove double "of" This closes #4999. commit c5371ee6430a36f55001048248b1b2adc901a1d1 Author: yew1eb <yew1eb@...> Date: 2017-11-11T13:08:21Z [hotfix][docs] Fix typos in deployment AWS documentation This closes #5000. commit fcbfbb9126d458feb17527b73be1eb38a2e640f7 Author: Michael Fong <mcfong.open@...> Date: 2017-11-13T02:17:02Z [FLINK-4500][docs] Update cassandra documentation regarding data loss As of FLINK-4500 the Cassandra connector will wait for pending updates to finish upon checkpoint. This closes #5002. commit 16107cf67416391ccf1493bda747a108bbbb3a20 Author: Piotr Nowojski <piotr.nowojski@...> Date: 2017-11-13T15:51:18Z [FLINK-7845][runtime] Make NettyMessage public This a walkaround strange javaassist bug. The issue should go away once we upgrade netty dependency. Please check the ticket for more information. This closes #5007. commit ed90379e3b96b03e24414169f92e6cb8371a6250 Author: zentol <chesnay@...> Date: 2017-11-14T09:58:43Z [FLINK-8056][dist] Use 'web.port' instead of 'jobmanager.web.port' This closes #5010. commit 408e186995f4e6bacc0904b8dd6b6f9b7e28e60e Author: Vetriselvan1187 <vetriblogger1187@...> Date: 2017-11-06T17:11:41Z [FLINK-7998][examples] Fix TPCHQuery3 examples This closes #4959. commit e16472953cef75743791c91f17e8114f2f045054 Author: zentol <chesnay@...> Date: 2017-11-14T13:47:00Z [FLINK-8071][build] Bump shade-plugin asm version to 5.1 This closes #5014. commit fe98cb77c4fcf2ff0e2840c2254fb8b517274917 Author: zentol <chesnay@...> Date: 2017-11-08T10:58:38Z [FLINK-7419][build][avro] Shade jackson dependency in flink-dist This closes #4981. commit 5f00294af4b5527d4e320c2adc66fe9c6c2fd5b6 Author: Nico Kruber <nico@...> Date: 2017-11-14T13:36:22Z [FLINK-7973] disable JNI bridge for relocated hadoop classes in s3-fs-* commit c32fcbc2673233bb1c16a8b96fb2b7e88a55fdca Author: Jark Wu <jark@...> Date: 2017-08-15T11:24:57Z [FLINK-7451] [table] Support non-ascii character literals in Table API and SQL This closes #4544. commit 9d34715747040352f6c0ca13c052763cd8049dad Author: twalthr <twalthr@...> Date: 2017-11-15T13:39:21Z [FLINK-7451] [table] Disable testing of the charset in TableTestBase commit 11218a35dc0fdd7439142a313e6628c51cffe689 Author: twalthr <twalthr@...> Date: 2017-11-14T10:06:54Z [FLINK-8013] [table] Support aggregate functions with generic arrays This closes #5011. commit 54eeccfe18ace636b77ff0226ce1f91d35e4d49d Author: twalthr <twalthr@...> Date: 2017-09-26T10:10:33Z [FLINK-7678] [table] Support composite inputs for user-defined functions This closes #4726. ---- ---