Thanks, actually found this out per: http://www.confluent.io/blog/tutorial-getting-started-with-the-new-apache-kafka-0.9-consumer-client <https://mailtrack.io/trace/link/f2e80a9ef7bfbabfc3e6f8951266d07b52051751?url=http%3A%2F%2Fwww.confluent.io%2Fblog%2Ftutorial-getting-started-with-the-new-apache-kafka-0.9-consumer-client&signature=44a0b68933da863e>
from TFA: consumer.commitSync(Collections.singletonMap(record.partition(), new OffsetAndMetadata(record.offset() + 1))); The committed offset should always be the offset of the next message that your application will read. Wish this was a bit clearer in the API docs :) On Mon, Feb 1, 2016 at 1:52 PM, Dana Powers <dana.pow...@gmail.com> wrote: > The committed offset is actually the next message to consume, not the last > message consumed. So that sounds like expected behavior to me. The consumer > code handles this internally, but if you write code to commit offsets > manually, it can be a gotcha. > > -Dana > > On Mon, Feb 1, 2016 at 1:35 PM, Adam Kunicki <a...@streamsets.com> wrote: > > > Hi, > > > > I've been noticing that a restarted consumer in 0.9 will start consuming > > from the last committed offset (inclusive). This means that any restarted > > consumer will get the last read (and committed) message causing a > duplicate > > each time the consumer is restarted from the same position if there have > > been no new messages. > > > > Per: > > > > > http://www.confluent.io/blog/tutorial-getting-started-with-the-new-apache-kafka-0.9-consumer-client > <https://mailtrack.io/trace/link/f2e80a9ef7bfbabfc3e6f8951266d07b52051751?url=http%3A%2F%2Fwww.confluent.io%2Fblog%2Ftutorial-getting-started-with-the-new-apache-kafka-0.9-consumer-client&signature=44a0b68933da863e> > > < > > > https://mailtrack.io/trace/link/9853c5856f2b5862212148c1a969575c970a3dcc?url=http%3A%2F%2Fwww.confluent.io%2Fblog%2Ftutorial-getting-started-with-the-new-apache-kafka-0.9-consumer-client&signature=63a1a40b88347844 > > > > > this seems like that is the intended behavior. > > > > Can anyone confirm this? If this is the case how are we expected to > handle > > these duplicated messages? > > > > -Adam > > > -- Adam Kunicki StreamSets | Field Engineer mobile: 415.890.DATA (3282) | linkedin <https://mailtrack.io/trace/link/50832933390e909694a7f2157c5d640476609cd1?url=http%3A%2F%2Fwww.adamkunicki.com&signature=c5598df83da6c7fa>