twalthr commented on a change in pull request #14111:
URL: https://github.com/apache/flink/pull/14111#discussion_r528611200
##
File path: docs/dev/table/connectors/kafka.md
##
@@ -249,6 +249,15 @@ Besides enabling Flink's checkpointing, you can also
choose three different mode
Please refer to [Kafka documentation]({% link dev/connectors/kafka.md
%}#kafka-producers-and-fault-tolerance) for more caveats about delivery
guarantees.
+### Per-partition-watermark Source
+
+Flink supports to emit per-partition-watermark for Kafka. Using this feature,
watermarks are generated inside the Kafka consumer. The per-partition-watermark
are merged in
+the same way as watermarks are merged on stream shuffles. The output watermark
of the source is determined by the minimum watermark among the partitions it
reads. Considering a watermark assigner
+advance the watermark according to the event-time on the records. If some
partitions in the topics are idle, the watermark assigner will not advance. You
can alleviate this problem by
+setting appropriate [idelness timeouts]({{ site.baseurl
}}/dev/event_timestamps_watermarks.html#dealing-with-idle-sources).
Review comment:
This documentation makes no sense for Table/SQL users. We don't have a
possibility to set this via options, right?
##
File path: docs/dev/table/connectors/kafka.md
##
@@ -249,6 +249,15 @@ Besides enabling Flink's checkpointing, you can also
choose three different mode
Please refer to [Kafka documentation]({% link dev/connectors/kafka.md
%}#kafka-producers-and-fault-tolerance) for more caveats about delivery
guarantees.
+### Per-partition-watermark Source
+
+Flink supports to emit per-partition-watermark for Kafka. Using this feature,
watermarks are generated inside the Kafka consumer. The per-partition-watermark
are merged in
+the same way as watermarks are merged on stream shuffles. The output watermark
of the source is determined by the minimum watermark among the partitions it
reads. Considering a watermark assigner
+advance the watermark according to the event-time on the records. If some
partitions in the topics are idle, the watermark assigner will not advance. You
can alleviate this problem by
+setting appropriate [idelness timeouts]({{ site.baseurl
}}/dev/event_timestamps_watermarks.html#dealing-with-idle-sources).
Review comment:
We should actually introduce an option for it.
##
File path: docs/dev/table/sourceSinks.zh.md
##
@@ -31,16 +31,15 @@ of a dynamic table is stored in external systems (such as
databases, key-value s
_Dynamic sources_ and _dynamic sinks_ can be used to read and write data from
and to an external system. In
the documentation, sources and sinks are often summarized under the term
_connector_.
-Flink provides pre-defined connectors for Kafka, Hive, and different file
systems. See the [connector section]({% link dev/table/connectors/index.zh.md
%})
+Flink provides pre-defined connectors for Kafka, Hive, and different file
systems. See the [connector section]({% link dev/table/connectors/index.md %})
Review comment:
undo this change?
##
File path: docs/dev/table/sourceSinks.zh.md
##
@@ -31,16 +31,15 @@ of a dynamic table is stored in external systems (such as
databases, key-value s
_Dynamic sources_ and _dynamic sinks_ can be used to read and write data from
and to an external system. In
the documentation, sources and sinks are often summarized under the term
_connector_.
-Flink provides pre-defined connectors for Kafka, Hive, and different file
systems. See the [connector section]({% link dev/table/connectors/index.zh.md
%})
+Flink provides pre-defined connectors for Kafka, Hive, and different file
systems. See the [connector section]({% link dev/table/connectors/index.md %})
for more information about built-in table sources and sinks.
This page focuses on how to develop a custom, user-defined connector.
Attention New table source and table
sink interfaces have been
introduced in Flink 1.11 as part of
[FLIP-95](https://cwiki.apache.org/confluence/display/FLINK/FLIP-95%3A+New+TableSource+and+TableSink+interfaces).
-Also the factory interfaces have been reworked. FLIP-95 is not fully
implemented yet. Many ability interfaces
-are not supported yet (e.g. for filter or partition push down). If necessary,
please also have a look
-at the [old table sources and sinks page]({% link
dev/table/legacySourceSinks.zh.md %}). Those interfaces
+Also the factory interfaces have been reworked. Many ability interfaces has
been supported (e.g. for filter or partition push down).
Review comment:
Remove `Many ability interfaces has been supported (e.g. for filter or
partition push down).` entirely?
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL