[ 
https://issues.apache.org/jira/browse/KAFKA-5827?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17692413#comment-17692413
 ] 

Greg Harris commented on KAFKA-5827:
------------------------------------

This is controllable via the Client Override feature KIP-458 
[https://cwiki.apache.org/confluence/display/KAFKA/KIP-458%3A+Connector+Client+Config+Override+Policy]
 .
You can configure the `consumer.override.auto.offset.reset` configuration 
property in a connector configuration to have the consumer begin reading from 
the latest record in a partition. After the connector commits offsets, further 
restarts will pick up where the previous commit finished, avoiding data loss 
while not re-reading previously committed messages.

> Allow configuring Kafka sink connectors to start processing records from the 
> end of topics
> ------------------------------------------------------------------------------------------
>
>                 Key: KAFKA-5827
>                 URL: https://issues.apache.org/jira/browse/KAFKA-5827
>             Project: Kafka
>          Issue Type: Improvement
>          Components: KafkaConnect
>            Reporter: Behrang Saeedzadeh
>            Priority: Major
>
> As far as I can see, Kafka connectors start exporting data of a topic from 
> the beginning of its partitions. We have a topic that contains a few million 
> old records that we don't need but we would like to start exporting new 
> records that are added to the topic.
> Basically:
> * When the connector is started for the first time and it does not have a 
> current offset stored, it should start consuming data from the end of topic 
> partitions
> * When the connector is restarted and has a current offset for partitions 
> stored somewhere, it should start from those offsets



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to