[
https://issues.apache.org/jira/browse/KAFKA-198?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13147848#comment-13147848
]
Jun Rao commented on KAFKA-198:
-------------------------------
This is a bit tricky to get right. The following is one possible design:
1. In ConsumerIterator, add a method clearCurrentChunk(). This method will
clear the current chunk being iterated. This method has to by synchronized with
makeNext().
2. In KafaMessageStream, add a method clear() which calls
consumerIterator.clearCurrentChunk
3. In Fetcher, break initConnections() into 2 methods: stopConnections() and
startConnections().
4. In ZookeeperConsumerConnector.updateFetcher. Do the following:
a. fetcher.stopConnections
b. for each new Fetcher to be created
b1. clear fetcher queue
b2. call KafkaMessageStream.clear
c. call commitOffsets
d. fetcher.startConnections
> Avoid duplicated message during consumer rebalance
> --------------------------------------------------
>
> Key: KAFKA-198
> URL: https://issues.apache.org/jira/browse/KAFKA-198
> Project: Kafka
> Issue Type: Improvement
> Affects Versions: 0.7
> Reporter: Jun Rao
>
> Currently, a consumer can get duplicated messages when a rebalance is
> triggered. It would be good if we can eliminate those duplicated messages.
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators:
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira