GitHub user medcv opened a pull request:
https://github.com/apache/flink/pull/6083
[FLINK-8983] End-to-end test: Confluent schema registry
## Brief change log
Added an end-to-end test which verifies that Flink is able to work together
with the Confluent schema registry. In order to do that, this test sets up a
Kafka cluster and a Flink job which writes and reads from the Confluent schema
registry producing an Avro type.
## Does this pull request potentially affect one of the following parts:
- Dependencies (does it add or upgrade a dependency): (no)
- The public API, i.e., is any changed class annotated with
`@Public(Evolving)`: (no)
- The serializers: (no)
- The runtime per-record code paths (performance sensitive): (no)
- Anything that affects deployment or recovery: JobManager (and its
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
- The S3 file system connector: (no)
## Documentation
- Does this pull request introduce a new feature? (no)
- If yes, how is the feature documented? (not applicable / docs /
JavaDocs / not documented)
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/medcv/flink FLINK-8983
Alternatively you can review and apply these changes as the patch at:
https://github.com/apache/flink/pull/6083.patch
To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:
This closes #6083
----
commit 8bcaee1a6d8b32e10888e46e608a1478b4a66e9b
Author: Yadan.JS <y_shirvany@...>
Date: 2018-05-21T02:31:26Z
[FLINK-8983] End-to-end test: Confluent schema registry
----
---