[ https://issues.apache.org/jira/browse/KAFKA-2758?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17013271#comment-17013271 ]
Guozhang Wang commented on KAFKA-2758: -------------------------------------- We used to put it on hold especially for 1) since KIP-211 is not merged yet, however even now after KIP-211 is merged we should be careful since a newer versioned client may talk to an older versioned broker (2.0-) which does not have KIP-211 yet. We have some plans for automatically detecting broker versions so I'd suggest before that we do not pick up this ticket yet. > Improve Offset Commit Behavior > ------------------------------ > > Key: KAFKA-2758 > URL: https://issues.apache.org/jira/browse/KAFKA-2758 > Project: Kafka > Issue Type: Improvement > Components: consumer > Reporter: Guozhang Wang > Priority: Major > Labels: newbie, reliability > > There are two scenarios of offset committing that we can improve: > 1) we can filter the partitions whose committed offset is equal to the > consumed offset, meaning there is no new consumed messages from this > partition and hence we do not need to include this partition in the commit > request. > 2) we can make a commit request right after resetting to a fetch / consume > position either according to the reset policy (e.g. on consumer starting up, > or handling of out of range offset, etc), or through the {code} seek {code} > so that if the consumer fails right after these event, upon recovery it can > restarts from the reset position instead of resetting again: this can lead > to, for example, data loss if we use "largest" as reset policy while there > are new messages coming to the fetching partitions. -- This message was sent by Atlassian Jira (v8.3.4#803005)