This is an automated email from the ASF dual-hosted git repository. cegerton pushed a commit to branch asf-site in repository https://gitbox.apache.org/repos/asf/kafka-site.git
The following commit(s) were added to refs/heads/asf-site by this push: new f3f93333 MINOR: Add missing entries for Kafka Connect to the documentation's table of contents (#538) f3f93333 is described below commit f3f933339cd38831be1db8f5e821194eb810164f Author: Yash Mayya <yash.ma...@gmail.com> AuthorDate: Tue Sep 5 19:41:48 2023 +0100 MINOR: Add missing entries for Kafka Connect to the documentation's table of contents (#538) Reviewers: Chris Egerton <chr...@aiven.io> --- 35/connect.html | 4 ++-- 35/toc.html | 8 ++++++++ 2 files changed, 10 insertions(+), 2 deletions(-) diff --git a/35/connect.html b/35/connect.html index e397aa79..20872b13 100644 --- a/35/connect.html +++ b/35/connect.html @@ -822,7 +822,7 @@ if (inputsChanged()) <p><code>SinkConnectors</code> usually only have to handle the addition of streams, which may translate to new entries in their outputs (e.g., a new database table). The framework manages any changes to the Kafka input, such as when the set of input topics changes because of a regex subscription. <code>SinkTasks</code> should expect new input streams, which may require creating new resources in the downstream system, such as a new table in a database. The trickiest situation to handl [...] - <h4><a id="connect_configs" href="#connect_configs">Connect Configuration Validation</a></h4> + <h4><a id="connect_configs" href="#connect_configs">Configuration Validation</a></h4> <p>Kafka Connect allows you to validate connector configurations before submitting a connector to be executed and can provide feedback about errors and recommended values. To take advantage of this, connector developers need to provide an implementation of <code>config()</code> to expose the configuration definition to the framework.</p> @@ -868,7 +868,7 @@ Struct struct = new Struct(schema) <p>Sink connectors are usually simpler because they are consuming data and therefore do not need to create schemas. However, they should take just as much care to validate that the schemas they receive have the expected format. When the schema does not match -- usually indicating the upstream producer is generating invalid data that cannot be correctly translated to the destination system -- sink connectors should throw an exception to indicate this error to the system.</p> - <h4><a id="connect_administration" href="#connect_administration">Kafka Connect Administration</a></h4> + <h3><a id="connect_administration" href="#connect_administration">8.4 Administration</a></h3> <p> Kafka Connect's <a href="#connect_rest">REST layer</a> provides a set of APIs to enable administration of the cluster. This includes APIs to view the configuration of connectors and the status of their tasks, as well as to alter their current behavior (e.g. changing configuration and restarting tasks). diff --git a/35/toc.html b/35/toc.html index ce444871..93fcbcd7 100644 --- a/35/toc.html +++ b/35/toc.html @@ -203,6 +203,14 @@ <li><a href="#connect_exactlyonce">Exactly-once support</a></li> </ul> <li><a href="#connect_development">8.3 Connector Development Guide</a></li> + <ul> + <li><a href="#connect_concepts">Core Concepts and APIs</a></li> + <li><a href="#connect_developing">Developing a Simple Connector</a></li> + <li><a href="#connect_dynamicio">Dynamic Input/Output Streams</a></li> + <li><a href="#connect_configs">Configuration Validation</a></li> + <li><a href="#connect_schemas">Working with Schemas</a></li> + </ul> + <li><a href="#connect_administration">8.4 Administration</a></li> </ul> </li> <li><a href="/{{version}}/documentation/streams">9. Kafka Streams</a>