twalthr commented on a change in pull request #14111:
URL: https://github.com/apache/flink/pull/14111#discussion_r528611200
##########
File path: docs/dev/table/connectors/kafka.md
##########
@@ -249,6 +249,15 @@ Besides enabling Flink's checkpointing, you can also
choose three different mode
Please refer to [Kafka documentation]({% link dev/connectors/kafka.md
%}#kafka-producers-and-fault-tolerance) for more caveats about delivery
guarantees.
+### Per-partition-watermark Source
+
+Flink supports to emit per-partition-watermark for Kafka. Using this feature,
watermarks are generated inside the Kafka consumer. The per-partition-watermark
are merged in
+the same way as watermarks are merged on stream shuffles. The output watermark
of the source is determined by the minimum watermark among the partitions it
reads. Considering a watermark assigner
+advance the watermark according to the event-time on the records. If some
partitions in the topics are idle, the watermark assigner will not advance. You
can alleviate this problem by
+setting appropriate [idelness timeouts]({{ site.baseurl
}}/dev/event_timestamps_watermarks.html#dealing-with-idle-sources).
Review comment:
This documentation makes no sense for Table/SQL users. We don't have a
possibility to set this via options, right?
##########
File path: docs/dev/table/connectors/kafka.md
##########
@@ -249,6 +249,15 @@ Besides enabling Flink's checkpointing, you can also
choose three different mode
Please refer to [Kafka documentation]({% link dev/connectors/kafka.md
%}#kafka-producers-and-fault-tolerance) for more caveats about delivery
guarantees.
+### Per-partition-watermark Source
+
+Flink supports to emit per-partition-watermark for Kafka. Using this feature,
watermarks are generated inside the Kafka consumer. The per-partition-watermark
are merged in
+the same way as watermarks are merged on stream shuffles. The output watermark
of the source is determined by the minimum watermark among the partitions it
reads. Considering a watermark assigner
+advance the watermark according to the event-time on the records. If some
partitions in the topics are idle, the watermark assigner will not advance. You
can alleviate this problem by
+setting appropriate [idelness timeouts]({{ site.baseurl
}}/dev/event_timestamps_watermarks.html#dealing-with-idle-sources).
Review comment:
We should actually introduce an option for it.
##########
File path: docs/dev/table/sourceSinks.zh.md
##########
@@ -31,16 +31,15 @@ of a dynamic table is stored in external systems (such as
databases, key-value s
_Dynamic sources_ and _dynamic sinks_ can be used to read and write data from
and to an external system. In
the documentation, sources and sinks are often summarized under the term
_connector_.
-Flink provides pre-defined connectors for Kafka, Hive, and different file
systems. See the [connector section]({% link dev/table/connectors/index.zh.md
%})
+Flink provides pre-defined connectors for Kafka, Hive, and different file
systems. See the [connector section]({% link dev/table/connectors/index.md %})
Review comment:
undo this change?
##########
File path: docs/dev/table/sourceSinks.zh.md
##########
@@ -31,16 +31,15 @@ of a dynamic table is stored in external systems (such as
databases, key-value s
_Dynamic sources_ and _dynamic sinks_ can be used to read and write data from
and to an external system. In
the documentation, sources and sinks are often summarized under the term
_connector_.
-Flink provides pre-defined connectors for Kafka, Hive, and different file
systems. See the [connector section]({% link dev/table/connectors/index.zh.md
%})
+Flink provides pre-defined connectors for Kafka, Hive, and different file
systems. See the [connector section]({% link dev/table/connectors/index.md %})
for more information about built-in table sources and sinks.
This page focuses on how to develop a custom, user-defined connector.
<span class="label label-danger">Attention</span> New table source and table
sink interfaces have been
introduced in Flink 1.11 as part of
[FLIP-95](https://cwiki.apache.org/confluence/display/FLINK/FLIP-95%3A+New+TableSource+and+TableSink+interfaces).
-Also the factory interfaces have been reworked. FLIP-95 is not fully
implemented yet. Many ability interfaces
-are not supported yet (e.g. for filter or partition push down). If necessary,
please also have a look
-at the [old table sources and sinks page]({% link
dev/table/legacySourceSinks.zh.md %}). Those interfaces
+Also the factory interfaces have been reworked. Many ability interfaces has
been supported (e.g. for filter or partition push down).
Review comment:
Remove `Many ability interfaces has been supported (e.g. for filter or
partition push down).` entirely?
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]