Thanks Anna, you are right on both points. I updated the KIP. -Jason
On Thu, Jul 26, 2018 at 2:08 PM, Anna Povzner <a...@confluent.io> wrote: > Hi Jason, > > Thanks for the update. I agree with the current proposal. > > Two minor comments: > 1) In “API Changes” section, first paragraph says that “users can catch the > more specific exception type and use the new `seekToNearest()` API defined > below.”. Since LogTruncationException “will include the partitions that > were truncated and the offset of divergence”., shouldn’t the client use > seek(offset) to seek to the offset of divergence in response to the > exception? > 2) In “Protocol Changes” section, OffsetsForLeaderEpoch subsection says > “Note > that consumers will send a sentinel value (-1) for the current epoch and > the broker will simply disregard that validation.”. Is that still true with > MetadataResponse containing leader epoch? > > Thanks, > Anna > > On Wed, Jul 25, 2018 at 1:44 PM Jason Gustafson <ja...@confluent.io> > wrote: > > > Hi All, > > > > I have made some updates to the KIP. As many of you know, a side project > of > > mine has been specifying the Kafka replication protocol in TLA. You can > > check out the code here if you are interested: > > https://github.com/hachikuji/kafka-specification. In addition to > > uncovering > > a couple unknown bugs in the replication protocol (e.g. > > https://issues.apache.org/jira/browse/KAFKA-7128), this has helped me > > validate the behavior in this KIP. In fact, the original version I > proposed > > had a weakness. I initially suggested letting the leader validate the > > expected epoch at the fetch offset. This made sense for the consumer in > the > > handling of unclean leader election, but it was not strong enough to > > protect the follower in all cases. In order to make advancement of the > high > > watermark safe, for example, the leader actually needs to be sure that > > every follower in the ISR matches its own epoch. > > > > I attempted to fix this problem by treating the epoch in the fetch > request > > slightly differently for consumers and followers. For consumers, it would > > be the expected epoch of the record at the fetch offset, and the leader > > would raise a LOG_TRUNCATION error if the expectation failed. For > > followers, it would be the current epoch and the leader would require > that > > it match its own epoch. This was unsatisfying both because of the > > inconsistency in behavior and because the consumer was left with the > weaker > > fencing that we already knew was insufficient for the replicas. > Ultimately > > I decided that we should make the behavior consistent and that meant that > > the consumer needed to act more like a following replica. Instead of > > checking for truncation while fetching, the consumer should check for > > truncation after leader changes. After checking for truncation, the > > consumer can then use the current epoch when fetching and get the > stronger > > protection that it provides. What this means is that the Metadata API > must > > include the current leader epoch. Given the problems we have had around > > stale metadata and how challenging they have been to debug, I'm convinced > > that this is a good idea in any case and it resolves the inconsistent > > behavior in the Fetch API. The downside is that there will be some > > additional overhead upon leader changes, but I don't think it is a major > > concern since leader changes are rare and the OffsetForLeaderEpoch > request > > is cheap. > > > > This approach leaves the door open for some interesting follow up > > improvements. For example, now that we have the leader epoch in the > > Metadata request, we can implement similar fencing for the Produce API. > And > > now that the consumer can reason about truncation, we could consider > having > > a configuration to expose records beyond the high watermark. This would > let > > users trade lower end-to-end latency for weaker durability semantics. It > is > > sort of like having an acks=0 option for the consumer. Neither of these > > options are included in this KIP, I am just mentioning them as potential > > work for the future. > > > > Finally, based on the discussion in this thread, I have added the > > seekToCommitted API for the consumer. Please take a look and let me know > > what you think. > > > > Thanks, > > Jason > > > > On Fri, Jul 20, 2018 at 2:34 PM, Guozhang Wang <wangg...@gmail.com> > wrote: > > > > > Hi Jason, > > > > > > The proposed API seems reasonable to me too. Could you please also > update > > > the wiki page ( > > > https://cwiki.apache.org/confluence/display/KAFKA/KIP- > > > 320%3A+Allow+fetchers+to+detect+and+handle+log+truncation) > > > with a section say "workflow" on how the proposed API will be co-used > > with > > > others to: > > > > > > 1. consumer callers handling a LogTruncationException. > > > 2. consumer internals for handling a retriable > > UnknownLeaderEpochException. > > > > > > > > > Guozhang > > > > > > > > > On Tue, Jul 17, 2018 at 10:23 AM, Anna Povzner <a...@confluent.io> > > wrote: > > > > > > > Hi Jason, > > > > > > > > > > > > I also like your proposal and agree that > > KafkaConsumer#seekToCommitted() > > > > is > > > > more intuitive as a way to initialize both consumer's position and > its > > > > fetch state. > > > > > > > > > > > > My understanding that KafkaConsumer#seekToCommitted() is purely for > > > > clients > > > > who store their offsets externally, right? And we are still going to > > > > add KafkaConsumer#findOffsets() > > > > in this KIP as we discussed, so that the client can handle > > > > LogTruncationException? > > > > > > > > > > > > Thanks, > > > > > > > > Anna > > > > > > > > > > > > On Thu, Jul 12, 2018 at 3:57 PM Dong Lin <lindon...@gmail.com> > wrote: > > > > > > > > > Hey Jason, > > > > > > > > > > It is a great summary. The solution sounds good. I might have minor > > > > > comments regarding the method name. But we can discuss that minor > > > points > > > > > later after we reach consensus on the high level API. > > > > > > > > > > Thanks, > > > > > Dong > > > > > > > > > > > > > > > On Thu, Jul 12, 2018 at 11:41 AM, Jason Gustafson < > > ja...@confluent.io> > > > > > wrote: > > > > > > > > > > > Hey Anna and Dong, > > > > > > > > > > > > Thanks a lot for the great discussion. I've been hanging back a > bit > > > > > because > > > > > > honestly the best option hasn't seemed clear. I agree with Anna's > > > > general > > > > > > observation that there is a distinction between the position of > the > > > > > > consumer and its fetch state up to that position. If you think > > about > > > > it, > > > > > a > > > > > > committed offset actually represents both of these. The metadata > is > > > > used > > > > > to > > > > > > initialize the state of the consumer application and the offset > > > > > initializes > > > > > > the position. Additionally, we are extending the offset commit in > > > this > > > > > KIP > > > > > > to also include the last epoch fetched by the consumer, which is > > used > > > > to > > > > > > initialize the internal fetch state. Of course if you do an > > arbitrary > > > > > > `seek` and immediately commit offsets, then there won't be a last > > > epoch > > > > > to > > > > > > commit. This seems intuitive since there is no fetch state in > this > > > > case. > > > > > We > > > > > > only commit fetch state when we have it. > > > > > > > > > > > > So if we think about a committed offset as initializing both the > > > > > consumer's > > > > > > position and its fetch state, then the gap in the API is > evidently > > > that > > > > > we > > > > > > don't have a way to initialize the consumer to a committed > offset. > > We > > > > do > > > > > it > > > > > > implicitly of course for offsets stored in Kafka, but since > > external > > > > > > storage is a use case we support, then we should have an explicit > > API > > > > as > > > > > > well. Perhaps something like this: > > > > > > > > > > > > seekToCommitted(TopicPartition, OffsetAndMetadata) > > > > > > > > > > > > In this KIP, we are proposing to allow the `OffsetAndMetadata` > > object > > > > to > > > > > > include the leader epoch, so I think this would have the same > > effect > > > as > > > > > > Anna's suggested `seekToRecord`. But perhaps it is a more natural > > fit > > > > > given > > > > > > the current API? Furthermore, if we find a need for additional > > > metadata > > > > > in > > > > > > the offset commit API in the future, then we will just need to > > modify > > > > the > > > > > > `OffsetAndMetadata` object and we will not need a new `seek` API. > > > > > > > > > > > > With this approach, I think then we can leave the `position` API > as > > > it > > > > > is. > > > > > > The position of the consumer is still just the next expected > fetch > > > > > offset. > > > > > > If a user needs to record additional state based on previous > fetch > > > > > > progress, then they would use the result of the previous fetch to > > > > obtain > > > > > > it. This makes the dependence on fetch progress explicit. I think > > we > > > > > could > > > > > > make this a little more convenience with a helper in the > > > > > `ConsumerRecords` > > > > > > object, but I think that's more of a nice-to-have. > > > > > > > > > > > > Thoughts? > > > > > > > > > > > > By the way, I have been iterating a little bit on the replica > side > > of > > > > > this > > > > > > KIP. My initial proposal in fact did not have strong enough > fencing > > > to > > > > > > protect all of the edge cases. I believe the current proposal > fixes > > > the > > > > > > problems, but I am still verifying the model. > > > > > > > > > > > > Thanks, > > > > > > Jason > > > > > > > > > > > > > > > > > > On Wed, Jul 11, 2018 at 10:45 AM, Dong Lin <lindon...@gmail.com> > > > > wrote: > > > > > > > > > > > > > Hey Anna, > > > > > > > > > > > > > > Thanks much for the explanation. Approach 1 also sounds good to > > > me. I > > > > > > think > > > > > > > findOffsets() is useful for users who don't use automatic > offset > > > > reset > > > > > > > policy. > > > > > > > > > > > > > > Just one more question. Since users who store offsets > externally > > > need > > > > > to > > > > > > > provide leaderEpoch to findOffsets(...), do we need an extra > API > > > for > > > > > user > > > > > > > to get both offset and leaderEpoch, e.g. recordPosition()? > > > > > > > > > > > > > > Thanks, > > > > > > > Dong > > > > > > > > > > > > > > On Wed, Jul 11, 2018 at 10:12 AM, Anna Povzner < > > a...@confluent.io> > > > > > > wrote: > > > > > > > > > > > > > > > Hi Dong, > > > > > > > > > > > > > > > > > > > > > > > > What I called “not covering all use cases” is what you call > > > > > best-effort > > > > > > > > (not guaranteeing some corner cases). I think we are on the > > same > > > > page > > > > > > > here. > > > > > > > > > > > > > > > > > > > > > > > > I wanted to be clear in the API whether the consumer seeks > to a > > > > > > position > > > > > > > > (offset) or to a record (offset, leader epoch). The only > > use-case > > > > of > > > > > > > > seeking to a record is seeking to a committed offset for a > user > > > who > > > > > > > stores > > > > > > > > committed offsets externally. (Unless users find some other > > > reason > > > > to > > > > > > > seek > > > > > > > > to a record.) I thought it was possible to provide this > > > > functionality > > > > > > > with > > > > > > > > findOffset(offset, leader epoch) followed by a seek(offset). > > > > However, > > > > > > you > > > > > > > > are right that this will not handle the race condition where > > > > > > > non-divergent > > > > > > > > offset found by findOffset() could change again before the > > > consumer > > > > > > does > > > > > > > > the first fetch. > > > > > > > > > > > > > > > > > > > > > > > > Regarding position() — if we add position that returns > (offset, > > > > > leader > > > > > > > > epoch), this is specifically a position after a record that > was > > > > > > actually > > > > > > > > consumed or position of a committed record. In which case, I > > > still > > > > > > think > > > > > > > > it’s cleaner to get a record position of consumed message > from > > a > > > > new > > > > > > > helper > > > > > > > > method in ConsumerRecords() or from committed offsets. > > > > > > > > > > > > > > > > > > > > > > > > I think all the use-cases could be then covered with: > > > > > > > > > > > > > > > > (Approach 1) > > > > > > > > > > > > > > > > seekToRecord(offset, leaderEpoch) — this will just > > initialize/set > > > > the > > > > > > > > consumer state; > > > > > > > > > > > > > > > > findOffsets(offset, leaderEpoch) returns {offset, > leaderEpoch} > > > > > > > > > > > > > > > > > > > > > > > > If we agree that the race condition is also a corner case, > > then I > > > > > think > > > > > > > we > > > > > > > > can cover use-cases with: > > > > > > > > > > > > > > > > (Approach 2) > > > > > > > > > > > > > > > > findOffsets(offset, leaderEpoch) returns offset — we still > want > > > > > leader > > > > > > > > epoch as a parameter for the users who store their committed > > > > offsets > > > > > > > > externally. > > > > > > > > > > > > > > > > > > > > > > > > I am actually now leaning more to approach 1, since it is > more > > > > > > explicit, > > > > > > > > and maybe there are more use cases for it. > > > > > > > > > > > > > > > > > > > > > > > > Thanks, > > > > > > > > > > > > > > > > Anna > > > > > > > > > > > > > > > > > > > > > > > > On Tue, Jul 10, 2018 at 3:47 PM Dong Lin < > lindon...@gmail.com> > > > > > wrote: > > > > > > > > > > > > > > > > > Hey Anna, > > > > > > > > > > > > > > > > > > Thanks for the comment. To answer your question, it seems > > that > > > we > > > > > can > > > > > > > > cover > > > > > > > > > all case in this KIP. As stated in "Consumer Handling" > > section, > > > > > > KIP-101 > > > > > > > > > based approach will be used to derive the truncation offset > > > from > > > > > the > > > > > > > > > 2-tuple (offset, leaderEpoch). This approach is best effort > > and > > > > it > > > > > is > > > > > > > > > inaccurate only in very rare scenarios (as described in > > > KIP-279). > > > > > > > > > > > > > > > > > > By using seek(offset, leaderEpoch), consumer will still be > > able > > > > to > > > > > > > follow > > > > > > > > > this best-effort approach to detect log truncation and > > > determine > > > > > the > > > > > > > > > truncation offset. On the other hand, if we use > seek(offset), > > > > > > consumer > > > > > > > > will > > > > > > > > > not detect log truncation in some cases which weakens the > > > > guarantee > > > > > > of > > > > > > > > this > > > > > > > > > KIP. Does this make sense? > > > > > > > > > > > > > > > > > > Thanks, > > > > > > > > > Dong > > > > > > > > > > > > > > > > > > On Tue, Jul 10, 2018 at 3:14 PM, Anna Povzner < > > > a...@confluent.io > > > > > > > > > > > > wrote: > > > > > > > > > > > > > > > > > > > Sorry, I hit "send" before finishing. Continuing... > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > 2) Hiding most of the consumer handling log truncation > > logic > > > > with > > > > > > > > minimal > > > > > > > > > > exposure in KafkaConsumer API. I was proposing this > path. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Before answering your specific questions… I want to > answer > > to > > > > > your > > > > > > > > > comment > > > > > > > > > > “In general, maybe we should discuss the final solution > > that > > > > > covers > > > > > > > all > > > > > > > > > > cases?”. With current KIP, we don’t cover all cases of > > > consumer > > > > > > > > detecting > > > > > > > > > > log truncation because the KIP proposes a leader epoch > > cache > > > in > > > > > > > > consumer > > > > > > > > > > that does not persist across restarts. Plus, we only > store > > > last > > > > > > > > committed > > > > > > > > > > offset (either internally or users can store externally). > > > This > > > > > has > > > > > > a > > > > > > > > > > limitation that the consumer will not always be able to > > find > > > > > point > > > > > > of > > > > > > > > > > truncation just because we have a limited history (just > one > > > > data > > > > > > > > point). > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > So, maybe we should first agree on whether we accept that > > > > storing > > > > > > > last > > > > > > > > > > committed offset/leader epoch has a limitation that the > > > > consumer > > > > > > will > > > > > > > > not > > > > > > > > > > be able to detect log truncation in all cases? > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Thanks, > > > > > > > > > > > > > > > > > > > > Anna > > > > > > > > > > > > > > > > > > > > On Tue, Jul 10, 2018 at 2:20 PM Anna Povzner < > > > > a...@confluent.io> > > > > > > > > wrote: > > > > > > > > > > > > > > > > > > > > > Hi Dong, > > > > > > > > > > > > > > > > > > > > > > Thanks for the follow up! I finally have much more > clear > > > > > > > > understanding > > > > > > > > > of > > > > > > > > > > > where you are coming from. > > > > > > > > > > > > > > > > > > > > > > You are right. The success of findOffsets()/finding a > > point > > > > of > > > > > > > > > > > non-divergence depends on whether we have enough > entries > > in > > > > the > > > > > > > > > > consumer's > > > > > > > > > > > leader epoch cache. However, I think this is a > > fundamental > > > > > > > limitation > > > > > > > > > of > > > > > > > > > > > having a leader epoch cache that does not persist > across > > > > > consumer > > > > > > > > > > restarts. > > > > > > > > > > > > > > > > > > > > > > If we consider the general case where consumer may or > may > > > not > > > > > > have > > > > > > > > this > > > > > > > > > > > cache, then I see two paths: > > > > > > > > > > > 1) Letting the user to track the leader epoch history > > > > > externally, > > > > > > > and > > > > > > > > > > have > > > > > > > > > > > more exposure to leader epoch and finding point of > > > > > non-divergence > > > > > > > in > > > > > > > > > > > KafkaConsumer API. I understand this is the case you > were > > > > > talking > > > > > > > > > about. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > On Tue, Jul 10, 2018 at 12:16 PM Dong Lin < > > > > lindon...@gmail.com > > > > > > > > > > > > > > wrote: > > > > > > > > > > > > > > > > > > > > > >> Hey Anna, > > > > > > > > > > >> > > > > > > > > > > >> Thanks much for your detailed explanation and example! > > It > > > > does > > > > > > > help > > > > > > > > me > > > > > > > > > > >> understand the difference between our understanding. > > > > > > > > > > >> > > > > > > > > > > >> So it seems that the solution based on findOffsets() > > > > currently > > > > > > > > focuses > > > > > > > > > > >> mainly on the scenario that consumer has cached > > > leaderEpoch > > > > -> > > > > > > > > offset > > > > > > > > > > >> mapping whereas I was thinking about the general case > > > where > > > > > > > consumer > > > > > > > > > may > > > > > > > > > > >> or > > > > > > > > > > >> may not have this cache. I guess that is why we have > > > > different > > > > > > > > > > >> understanding here. I have some comments below. > > > > > > > > > > >> > > > > > > > > > > >> > > > > > > > > > > >> 3) The proposed solution using findOffsets(offset, > > > > > leaderEpoch) > > > > > > > > > followed > > > > > > > > > > >> by > > > > > > > > > > >> seek(offset) works if consumer has the cached > > leaderEpoch > > > -> > > > > > > > offset > > > > > > > > > > >> mapping. But if we assume consumer has this cache, do > we > > > > need > > > > > to > > > > > > > > have > > > > > > > > > > >> leaderEpoch in the findOffsets(...)? Intuitively, the > > > > > > > > > > findOffsets(offset) > > > > > > > > > > >> can also derive the leaderEpoch using offset just like > > the > > > > > > > proposed > > > > > > > > > > >> solution does with seek(offset). > > > > > > > > > > >> > > > > > > > > > > >> > > > > > > > > > > >> 4) If consumer does not have cached leaderEpoch -> > > offset > > > > > > mapping, > > > > > > > > > which > > > > > > > > > > >> is > > > > > > > > > > >> the case if consumer is restarted on a new machine, > then > > > it > > > > is > > > > > > not > > > > > > > > > clear > > > > > > > > > > >> what leaderEpoch would be included in the FetchRequest > > if > > > > > > consumer > > > > > > > > > does > > > > > > > > > > >> seek(offset). This is the case that motivates the > first > > > > > question > > > > > > > of > > > > > > > > > the > > > > > > > > > > >> previous email. In general, maybe we should discuss > the > > > > final > > > > > > > > solution > > > > > > > > > > >> that > > > > > > > > > > >> covers all cases? > > > > > > > > > > >> > > > > > > > > > > >> > > > > > > > > > > >> 5) The second question in my previous email is related > > to > > > > the > > > > > > > > > following > > > > > > > > > > >> paragraph: > > > > > > > > > > >> > > > > > > > > > > >> "... In some cases, offsets returned from position() > > could > > > > be > > > > > > > actual > > > > > > > > > > >> consumed messages by this consumer identified by > > {offset, > > > > > leader > > > > > > > > > epoch}. > > > > > > > > > > >> In > > > > > > > > > > >> other cases, position() returns offset that was not > > > actually > > > > > > > > consumed. > > > > > > > > > > >> Suppose, the user calls position() for the last > > > offset...". > > > > > > > > > > >> > > > > > > > > > > >> I guess my point is that, if user calls position() for > > the > > > > > last > > > > > > > > offset > > > > > > > > > > and > > > > > > > > > > >> uses that offset in seek(...), then user can probably > > just > > > > > call > > > > > > > > > > >> Consumer#seekToEnd() without calling position() and > > > > seek(...). > > > > > > > > > Similarly > > > > > > > > > > >> user can call Consumer#seekToBeginning() to the seek > to > > > the > > > > > > > earliest > > > > > > > > > > >> position without calling position() and seek(...). > Thus > > > > > > position() > > > > > > > > > only > > > > > > > > > > >> needs to return the actual consumed messages > identified > > by > > > > > > > {offset, > > > > > > > > > > leader > > > > > > > > > > >> epoch}. Does this make sense? > > > > > > > > > > >> > > > > > > > > > > >> > > > > > > > > > > >> Thanks, > > > > > > > > > > >> Dong > > > > > > > > > > >> > > > > > > > > > > >> > > > > > > > > > > >> On Mon, Jul 9, 2018 at 6:47 PM, Anna Povzner < > > > > > a...@confluent.io > > > > > > > > > > > > > > > > wrote: > > > > > > > > > > >> > > > > > > > > > > >> > Hi Dong, > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > Thanks for considering my suggestions. > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > Based on your comments, I realized that my > suggestion > > > was > > > > > not > > > > > > > > > complete > > > > > > > > > > >> with > > > > > > > > > > >> > regard to KafkaConsumer API vs. consumer-broker > > > protocol. > > > > > > While > > > > > > > I > > > > > > > > > > >> propose > > > > > > > > > > >> > to keep KafkaConsumer#seek() unchanged and take > offset > > > > only, > > > > > > the > > > > > > > > > > >> underlying > > > > > > > > > > >> > consumer will send the next FetchRequest() to broker > > > with > > > > > > offset > > > > > > > > and > > > > > > > > > > >> > leaderEpoch if it is known (based on leader epoch > > cache > > > in > > > > > > > > > consumer) — > > > > > > > > > > >> note > > > > > > > > > > >> > that this is different from the current KIP, which > > > > suggests > > > > > to > > > > > > > > > always > > > > > > > > > > >> send > > > > > > > > > > >> > unknown leader epoch after seek(). This way, if the > > > > consumer > > > > > > > and a > > > > > > > > > > >> broker > > > > > > > > > > >> > agreed on the point of non-divergence, which is some > > > > > {offset, > > > > > > > > > > >> leaderEpoch} > > > > > > > > > > >> > pair, the new leader which causes another truncation > > > (even > > > > > > > further > > > > > > > > > > back) > > > > > > > > > > >> > will be able to detect new divergence and restart > the > > > > > process > > > > > > of > > > > > > > > > > finding > > > > > > > > > > >> > the new point of non-divergence. So, to answer your > > > > > question, > > > > > > If > > > > > > > > the > > > > > > > > > > >> > truncation happens just after the user calls > > > > > > > > > > >> > KafkaConsumer#findOffsets(offset, leaderEpoch) > > followed > > > > by > > > > > > > > > > seek(offset), > > > > > > > > > > >> > the user will not seek to the wrong position without > > > > knowing > > > > > > > that > > > > > > > > > > >> > truncation has happened, because the consumer will > get > > > > > another > > > > > > > > > > >> truncation > > > > > > > > > > >> > error, and seek again. > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > I am afraid, I did not understand your second > > question. > > > > Let > > > > > me > > > > > > > > > > >> summarize my > > > > > > > > > > >> > suggestions again, and then give an example to > > hopefully > > > > > make > > > > > > my > > > > > > > > > > >> > suggestions more clear. Also, the last part of my > > > example > > > > > > shows > > > > > > > > how > > > > > > > > > > the > > > > > > > > > > >> > use-case in your first question will work. If it > does > > > not > > > > > > answer > > > > > > > > > your > > > > > > > > > > >> > second question, would you mind clarifying? I am > also > > > > > focusing > > > > > > > on > > > > > > > > > the > > > > > > > > > > >> case > > > > > > > > > > >> > of a consumer having enough entries in the cache. > The > > > case > > > > > of > > > > > > > > > > restarting > > > > > > > > > > >> > from committed offset either stored externally or > > > > internally > > > > > > > will > > > > > > > > > > >> probably > > > > > > > > > > >> > need to be discussed more. > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > Let me summarize my suggestion again: > > > > > > > > > > >> > > > > > > > > > > > >> > 1) KafkaConsumer#seek() and KafkaConsumer#position() > > > > remains > > > > > > > > > unchanged > > > > > > > > > > >> > > > > > > > > > > > >> > 2) New KafkaConsumer#findOffsets() takes {offset, > > > > > leaderEpoch} > > > > > > > > pair > > > > > > > > > > per > > > > > > > > > > >> > topic partition and returns offset per topic > > partition. > > > > > > > > > > >> > > > > > > > > > > > >> > 3) FetchRequest() to broker after > KafkaConsumer#seek() > > > > will > > > > > > > > contain > > > > > > > > > > the > > > > > > > > > > >> > offset set by seek and leaderEpoch that corresponds > to > > > the > > > > > > > offset > > > > > > > > > > based > > > > > > > > > > >> on > > > > > > > > > > >> > leader epoch cache in the consumer. > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > The rest of this e-mail is a long and contrived > > example > > > > with > > > > > > > > several > > > > > > > > > > log > > > > > > > > > > >> > truncations and unclean leader elections to > illustrate > > > the > > > > > API > > > > > > > and > > > > > > > > > > your > > > > > > > > > > >> > first use-case. Suppose we have three brokers. > > > Initially, > > > > > > Broker > > > > > > > > A, > > > > > > > > > B, > > > > > > > > > > >> and > > > > > > > > > > >> > C has one message at offset 0 with leader epoch 0. > > Then, > > > > > > Broker > > > > > > > A > > > > > > > > > goes > > > > > > > > > > >> down > > > > > > > > > > >> > for some time. Broker B becomes a leader with epoch > 1, > > > and > > > > > > > writes > > > > > > > > > > >> messages > > > > > > > > > > >> > to offsets 1 and 2. Broker C fetches offset 1, but > > > before > > > > > > > fetching > > > > > > > > > > >> offset > > > > > > > > > > >> > 2, becomes a leader with leader epoch 2 and writes a > > > > message > > > > > > at > > > > > > > > > offset > > > > > > > > > > >> 2. > > > > > > > > > > >> > Here is the state of brokers at this point: > > > > > > > > > > >> > > > > > > > > > > > >> > > Broker A: > > > > > > > > > > >> > > offset 0, epoch 0 <— leader > > > > > > > > > > >> > > goes down… > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > > Broker B: > > > > > > > > > > >> > > offset 0, epoch 0 > > > > > > > > > > >> > > offset 1, epoch 1 <- leader > > > > > > > > > > >> > > offset 2, epoch 1 > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > Broker C: > > > > > > > > > > >> > > offset 0, epoch 0 > > > > > > > > > > >> > > offset 1, epoch 1 > > > > > > > > > > >> > > offset 2, epoch 2 <— leader > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > Before Broker C becomes a leader with leader epoch > 2, > > > the > > > > > > > consumer > > > > > > > > > > >> consumed > > > > > > > > > > >> > the following messages from broker A and broker B: > > > > > > > > > > >> > > > > > > > > > > > >> > {offset=0, leaderEpoch=0}, {offset=1, > leaderEpoch=1}, > > > > > > {offset=2, > > > > > > > > > > >> > leaderEpoch=1}. > > > > > > > > > > >> > > > > > > > > > > > >> > Consumer’s leader epoch cache at this point contains > > the > > > > > > > following > > > > > > > > > > >> entries: > > > > > > > > > > >> > > > > > > > > > > > >> > (leaderEpoch=0, startOffset=0) > > > > > > > > > > >> > > > > > > > > > > > >> > (leaderEpoch=1, startOffset=1) > > > > > > > > > > >> > > > > > > > > > > > >> > endOffset = 3 > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > Then, broker B becomes the follower of broker C, > > > truncates > > > > > and > > > > > > > > > starts > > > > > > > > > > >> > fetching from offset 2. > > > > > > > > > > >> > > > > > > > > > > > >> > Consumer sends fetchRequest(offset=3, leaderEpoch=1) > > and > > > > > gets > > > > > > > > > > >> > LOG_TRUNCATION > > > > > > > > > > >> > error from broker C. > > > > > > > > > > >> > > > > > > > > > > > >> > In response, the client calls > > KafkaConsumer#findOffsets( > > > > > > > offset=3, > > > > > > > > > > >> > leaderEpoch=1). The underlying consumer sends > > > > > > > > > > >> > OffsetsForLeaderEpoch(leaderEpoch=1), broker C > > responds > > > > with > > > > > > > > > > >> > {leaderEpoch=1, endOffset=2}. So, > > > > > > > > > KafkaConsumer#findOffsets(offset=3, > > > > > > > > > > >> > leaderEpoch=1) returns offset=2. > > > > > > > > > > >> > > > > > > > > > > > >> > In response, consumer calls KafkaConsumer@seek > > > (offset=2) > > > > > > > followed > > > > > > > > > by > > > > > > > > > > >> > poll(), which results in FetchRequest(offset=2, > > > > > leaderEpoch=1) > > > > > > > to > > > > > > > > > > >> broker C. > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > I will continue with this example with the goal to > > > answer > > > > > your > > > > > > > > first > > > > > > > > > > >> > question about truncation just after findOffsets() > > > > followed > > > > > by > > > > > > > > > seek(): > > > > > > > > > > >> > > > > > > > > > > > >> > Suppose, brokers B and C go down, and broker A comes > > up > > > > and > > > > > > > > becomes > > > > > > > > > a > > > > > > > > > > >> > leader with leader epoch 3, and writes a message to > > > offset > > > > > 1. > > > > > > > > > Suppose, > > > > > > > > > > >> this > > > > > > > > > > >> > happens before the consumer gets response from > broker > > C > > > to > > > > > the > > > > > > > > > > previous > > > > > > > > > > >> > fetch request: FetchRequest(offset=2, > leaderEpoch=1). > > > > > > > > > > >> > > > > > > > > > > > >> > Consumer re-sends FetchRequest(offset=2, > > leaderEpoch=1) > > > to > > > > > > > broker > > > > > > > > A, > > > > > > > > > > >> which > > > > > > > > > > >> > returns LOG_TRUNCATION error, because broker A has > > > leader > > > > > > epoch > > > > > > > 3 > > > > > > > > > > > > > > > > > > > >> leader > > > > > > > > > > >> > epoch in FetchRequest with starting offset = 1 < > > offset > > > 2 > > > > in > > > > > > > > > > >> > FetchRequest(). > > > > > > > > > > >> > > > > > > > > > > > >> > In response, the user calls > KafkaConsumer#findOffsets( > > > > > > offset=2, > > > > > > > > > > >> > leaderEpoch=1). The underlying consumer sends > > > > > > > > > > >> > OffsetsForLeaderEpoch(leaderEpoch=1), broker A > > responds > > > > with > > > > > > > > > > >> > {leaderEpoch=0, endOffset=1}; the underlying > consumer > > > > finds > > > > > > > > > > leaderEpoch > > > > > > > > > > >> = 0 > > > > > > > > > > >> > in its cache with end offset == 1, which results in > > > > > > > > > > >> > KafkaConsumer#findOffsets(offset=2, leaderEpoch=1) > > > > returning > > > > > > > > offset > > > > > > > > > > = 1. > > > > > > > > > > >> > > > > > > > > > > > >> > In response, the user calls KafkaConsumer@seek > > > (offset=1) > > > > > > > followed > > > > > > > > > by > > > > > > > > > > >> > poll(), which results in FetchRequest(offset=1, > > > > > leaderEpoch=0) > > > > > > > to > > > > > > > > > > >> broker A, > > > > > > > > > > >> > which responds with message at offset 1, leader > epoch > > 3. > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > I will think some more about consumers restarting > from > > > > > > committed > > > > > > > > > > >> offsets, > > > > > > > > > > >> > and send a follow up. > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > Thanks, > > > > > > > > > > >> > > > > > > > > > > > >> > Anna > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > On Sat, Jul 7, 2018 at 1:36 AM Dong Lin < > > > > > lindon...@gmail.com> > > > > > > > > > wrote: > > > > > > > > > > >> > > > > > > > > > > > >> > > Hey Anna, > > > > > > > > > > >> > > > > > > > > > > > > >> > > Thanks much for the thoughtful reply. It makes > sense > > > to > > > > > > > > different > > > > > > > > > > >> between > > > > > > > > > > >> > > "seeking to a message" and "seeking to a > position". > > I > > > > have > > > > > > to > > > > > > > > > > >> questions > > > > > > > > > > >> > > here: > > > > > > > > > > >> > > > > > > > > > > > > >> > > - For "seeking to a message" use-case, with the > > > proposed > > > > > > > > approach > > > > > > > > > > user > > > > > > > > > > >> > > needs to call findOffset(offset, leaderEpoch) > > followed > > > > by > > > > > > > > > > >> seek(offset). > > > > > > > > > > >> > If > > > > > > > > > > >> > > message truncation and message append happen > > > immediately > > > > > > after > > > > > > > > > > >> > > findOffset(offset, > > > > > > > > > > >> > > leaderEpoch) but before seek(offset), it seems > that > > > user > > > > > > will > > > > > > > > seek > > > > > > > > > > to > > > > > > > > > > >> the > > > > > > > > > > >> > > wrong message without knowing the truncation has > > > > happened. > > > > > > > Would > > > > > > > > > > this > > > > > > > > > > >> be > > > > > > > > > > >> > a > > > > > > > > > > >> > > problem? > > > > > > > > > > >> > > > > > > > > > > > > >> > > - For "seeking to a position" use-case, it seems > > that > > > > > there > > > > > > > can > > > > > > > > be > > > > > > > > > > two > > > > > > > > > > >> > > positions, i.e. earliest and latest. So these two > > > cases > > > > > can > > > > > > be > > > > > > > > > > >> > > Consumer.fulfilled by seekToBeginning() and > > > > > > > > Consumer.seekToEnd(). > > > > > > > > > > >> Then it > > > > > > > > > > >> > > seems that user will only need to call position() > > and > > > > > seek() > > > > > > > for > > > > > > > > > > >> "seeking > > > > > > > > > > >> > > to a message" use-case? > > > > > > > > > > >> > > > > > > > > > > > > >> > > Thanks, > > > > > > > > > > >> > > Dong > > > > > > > > > > >> > > > > > > > > > > > > >> > > > > > > > > > > > > >> > > On Wed, Jul 4, 2018 at 12:33 PM, Anna Povzner < > > > > > > > > a...@confluent.io> > > > > > > > > > > >> wrote: > > > > > > > > > > >> > > > > > > > > > > > > >> > > > Hi Jason and Dong, > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > I’ve been thinking about your suggestions and > > > > discussion > > > > > > > > > regarding > > > > > > > > > > >> > > > position(), seek(), and new proposed API. > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > Here is my thought process why we should keep > > > > position() > > > > > > and > > > > > > > > > > seek() > > > > > > > > > > >> API > > > > > > > > > > >> > > > unchanged. > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > I think we should separate {offset, leader > epoch} > > > that > > > > > > > > uniquely > > > > > > > > > > >> > > identifies > > > > > > > > > > >> > > > a message from an offset that is a position. In > > some > > > > > > cases, > > > > > > > > > > offsets > > > > > > > > > > >> > > > returned from position() could be actual > consumed > > > > > messages > > > > > > > by > > > > > > > > > this > > > > > > > > > > >> > > consumer > > > > > > > > > > >> > > > identified by {offset, leader epoch}. In other > > > cases, > > > > > > > > position() > > > > > > > > > > >> > returns > > > > > > > > > > >> > > > offset that was not actually consumed. Suppose, > > the > > > > user > > > > > > > calls > > > > > > > > > > >> > position() > > > > > > > > > > >> > > > for the last offset. Suppose we return {offset, > > > leader > > > > > > > epoch} > > > > > > > > of > > > > > > > > > > the > > > > > > > > > > >> > > > message currently in the log. Then, the message > > gets > > > > > > > truncated > > > > > > > > > > >> before > > > > > > > > > > >> > > > consumer’s first poll(). It does not make sense > > for > > > > > poll() > > > > > > > to > > > > > > > > > fail > > > > > > > > > > >> in > > > > > > > > > > >> > > this > > > > > > > > > > >> > > > case, because the log truncation did not > actually > > > > happen > > > > > > > from > > > > > > > > > the > > > > > > > > > > >> > > consumer > > > > > > > > > > >> > > > perspective. On the other hand, as the KIP > > proposes, > > > > it > > > > > > > makes > > > > > > > > > > sense > > > > > > > > > > >> for > > > > > > > > > > >> > > the > > > > > > > > > > >> > > > committed() method to return {offset, leader > > epoch} > > > > > > because > > > > > > > > > those > > > > > > > > > > >> > offsets > > > > > > > > > > >> > > > represent actual consumed messages. > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > The same argument applies to the seek() method — > > we > > > > are > > > > > > not > > > > > > > > > > seeking > > > > > > > > > > >> to > > > > > > > > > > >> > a > > > > > > > > > > >> > > > message, we are seeking to a position. > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > I like the proposal to add > > > KafkaConsumer#findOffsets() > > > > > > API. > > > > > > > I > > > > > > > > am > > > > > > > > > > >> > assuming > > > > > > > > > > >> > > > something like: > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > Map<TopicPartition, Long> > > > > > findOffsets(Map<TopicPartition, > > > > > > > > > > >> > OffsetAndEpoch> > > > > > > > > > > >> > > > offsetsToSearch) > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > Similar to seek() and position(), I think > > > > findOffsets() > > > > > > > should > > > > > > > > > > >> return > > > > > > > > > > >> > > > offset without leader epoch, because what we > want > > is > > > > the > > > > > > > > offset > > > > > > > > > > >> that we > > > > > > > > > > >> > > > think is closest to the not divergent message > from > > > the > > > > > > given > > > > > > > > > > >> consumed > > > > > > > > > > >> > > > message. Until the consumer actually fetches the > > > > > message, > > > > > > we > > > > > > > > > > should > > > > > > > > > > >> not > > > > > > > > > > >> > > let > > > > > > > > > > >> > > > the consumer store the leader epoch for a > message > > it > > > > did > > > > > > not > > > > > > > > > > >> consume. > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > So, the workflow will be: > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > 1) The user gets LogTruncationException with > > > {offset, > > > > > > leader > > > > > > > > > epoch > > > > > > > > > > >> of > > > > > > > > > > >> > the > > > > > > > > > > >> > > > previous message} (whatever we send with new > > > > > FetchRecords > > > > > > > > > > request). > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > 2) offset = findOffsets(tp -> {offset, leader > > > epoch}) > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > 3) seek(offset) > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > For the use-case where the users store committed > > > > offsets > > > > > > > > > > externally: > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > 1) Such users would have to track the leader > epoch > > > > > > together > > > > > > > > with > > > > > > > > > > an > > > > > > > > > > >> > > offset. > > > > > > > > > > >> > > > Otherwise, there is no way to detect later what > > > leader > > > > > > epoch > > > > > > > > was > > > > > > > > > > >> > > associated > > > > > > > > > > >> > > > with the message. I think it’s reasonable to ask > > > that > > > > > from > > > > > > > > users > > > > > > > > > > if > > > > > > > > > > >> > they > > > > > > > > > > >> > > > want to detect log truncation. Otherwise, they > > will > > > > get > > > > > > the > > > > > > > > > > current > > > > > > > > > > >> > > > behavior. > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > If the users currently get an offset to be > stored > > > > using > > > > > > > > > > position(), > > > > > > > > > > >> I > > > > > > > > > > >> > see > > > > > > > > > > >> > > > two possibilities. First, they call save offset > > > > returned > > > > > > > from > > > > > > > > > > >> > position() > > > > > > > > > > >> > > > that they call before poll(). In that case, it > > would > > > > not > > > > > > be > > > > > > > > > > correct > > > > > > > > > > >> to > > > > > > > > > > >> > > > store {offset, leader epoch} if we would have > > > changed > > > > > > > > position() > > > > > > > > > > to > > > > > > > > > > >> > > return > > > > > > > > > > >> > > > {offset, leader epoch} since actual fetched > > message > > > > > could > > > > > > be > > > > > > > > > > >> different > > > > > > > > > > >> > > > (from the example I described earlier). So, it > > would > > > > be > > > > > > more > > > > > > > > > > >> correct to > > > > > > > > > > >> > > > call position() after poll(). However, the user > > > > already > > > > > > gets > > > > > > > > > > >> > > > ConsumerRecords at this point, from which the > user > > > can > > > > > > > extract > > > > > > > > > > >> {offset, > > > > > > > > > > >> > > > leader epoch} of the last message. > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > So, I like the idea of adding a helper method to > > > > > > > > > ConsumerRecords, > > > > > > > > > > as > > > > > > > > > > >> > > Jason > > > > > > > > > > >> > > > proposed, something like: > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > public OffsetAndEpoch > lastOffsetWithLeaderEpoch(), > > > > where > > > > > > > > > > >> OffsetAndEpoch > > > > > > > > > > >> > > is > > > > > > > > > > >> > > > a data struct holding {offset, leader epoch}. > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > In this case, we would advise the user to follow > > the > > > > > > > workflow: > > > > > > > > > > >> poll(), > > > > > > > > > > >> > > get > > > > > > > > > > >> > > > {offset, leader epoch} from > > > > > ConsumerRecords#lastOffsetWith > > > > > > > > > > >> > LeaderEpoch(), > > > > > > > > > > >> > > > save offset and leader epoch, process records. > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > 2) When the user needs to seek to the last > > committed > > > > > > offset, > > > > > > > > > they > > > > > > > > > > >> call > > > > > > > > > > >> > > new > > > > > > > > > > >> > > > findOffsets(saved offset, leader epoch), and > then > > > > > > > > seek(offset). > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > What do you think? > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > Thanks, > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > Anna > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > On Tue, Jul 3, 2018 at 4:06 PM Dong Lin < > > > > > > > lindon...@gmail.com> > > > > > > > > > > >> wrote: > > > > > > > > > > >> > > > > > > > > > > > > > >> > > > > Hey Jason, > > > > > > > > > > >> > > > > > > > > > > > > > > >> > > > > Thanks much for your thoughtful explanation. > > > > > > > > > > >> > > > > > > > > > > > > > > >> > > > > Yes the solution using findOffsets(offset, > > > > > leaderEpoch) > > > > > > > also > > > > > > > > > > >> works. > > > > > > > > > > >> > The > > > > > > > > > > >> > > > > advantage of this solution it adds only one > API > > > > > instead > > > > > > of > > > > > > > > two > > > > > > > > > > >> APIs. > > > > > > > > > > >> > > The > > > > > > > > > > >> > > > > concern is that its usage seems a bit more > > clumsy > > > > for > > > > > > > > advanced > > > > > > > > > > >> users. > > > > > > > > > > >> > > > More > > > > > > > > > > >> > > > > specifically, advanced users who store offsets > > > > > > externally > > > > > > > > will > > > > > > > > > > >> always > > > > > > > > > > >> > > > need > > > > > > > > > > >> > > > > to call findOffsets() before calling > > seek(offset) > > > > > during > > > > > > > > > > consumer > > > > > > > > > > >> > > > > initialization. And those advanced users will > > need > > > > to > > > > > > > > manually > > > > > > > > > > >> keep > > > > > > > > > > >> > > track > > > > > > > > > > >> > > > > of the leaderEpoch of the last ConsumerRecord. > > > > > > > > > > >> > > > > > > > > > > > > > > >> > > > > The other solution may be more user-friendly > for > > > > > > advanced > > > > > > > > > users > > > > > > > > > > >> is to > > > > > > > > > > >> > > add > > > > > > > > > > >> > > > > two APIs, `void seek(offset, leaderEpoch)` and > > > > > `(offset, > > > > > > > > > epoch) > > > > > > > > > > = > > > > > > > > > > >> > > > > offsetEpochs(topicPartition)`. > > > > > > > > > > >> > > > > > > > > > > > > > > >> > > > > I kind of prefer the second solution because > it > > is > > > > > > easier > > > > > > > to > > > > > > > > > use > > > > > > > > > > >> for > > > > > > > > > > >> > > > > advanced users. If we need to expose > leaderEpoch > > > > > anyway > > > > > > to > > > > > > > > > > safely > > > > > > > > > > >> > > > identify > > > > > > > > > > >> > > > > a message, it may be conceptually simpler to > > > expose > > > > it > > > > > > > > > directly > > > > > > > > > > in > > > > > > > > > > >> > > > > seek(...) rather than requiring one more > > > translation > > > > > > using > > > > > > > > > > >> > > > > findOffsets(...). But I am also OK with the > > first > > > > > > solution > > > > > > > > if > > > > > > > > > > >> other > > > > > > > > > > >> > > > > developers also favor that one :) > > > > > > > > > > >> > > > > > > > > > > > > > > >> > > > > Thanks, > > > > > > > > > > >> > > > > Dong > > > > > > > > > > >> > > > > > > > > > > > > > > >> > > > > > > > > > > > > > > >> > > > > On Mon, Jul 2, 2018 at 11:10 AM, Jason > > Gustafson < > > > > > > > > > > >> ja...@confluent.io > > > > > > > > > > >> > > > > > > > > > > > > >> > > > > wrote: > > > > > > > > > > >> > > > > > > > > > > > > > > >> > > > > > Hi Dong, > > > > > > > > > > >> > > > > > > > > > > > > > > > >> > > > > > Thanks, I've been thinking about your > > > suggestions > > > > a > > > > > > bit. > > > > > > > > It > > > > > > > > > is > > > > > > > > > > >> > > > > challenging > > > > > > > > > > >> > > > > > to make this work given the current APIs. > One > > of > > > > the > > > > > > > > > > >> difficulties > > > > > > > > > > >> > is > > > > > > > > > > >> > > > that > > > > > > > > > > >> > > > > > we don't have an API to find the leader > epoch > > > for > > > > a > > > > > > > given > > > > > > > > > > >> offset at > > > > > > > > > > >> > > the > > > > > > > > > > >> > > > > > moment. So if the user does a seek to offset > > 5, > > > > then > > > > > > > we'll > > > > > > > > > > need > > > > > > > > > > >> a > > > > > > > > > > >> > new > > > > > > > > > > >> > > > API > > > > > > > > > > >> > > > > > to find the corresponding epoch in order to > > > > fulfill > > > > > > the > > > > > > > > new > > > > > > > > > > >> > > position() > > > > > > > > > > >> > > > > API. > > > > > > > > > > >> > > > > > Potentially we could modify ListOffsets to > > > enable > > > > > > > finding > > > > > > > > > the > > > > > > > > > > >> > leader > > > > > > > > > > >> > > > > epoch, > > > > > > > > > > >> > > > > > but I am not sure it is worthwhile. Perhaps > it > > > is > > > > > > > > reasonable > > > > > > > > > > for > > > > > > > > > > >> > > > advanced > > > > > > > > > > >> > > > > > usage to expect that the epoch information, > if > > > > > needed, > > > > > > > > will > > > > > > > > > be > > > > > > > > > > >> > > > extracted > > > > > > > > > > >> > > > > > from the records directly? It might make > sense > > > to > > > > > > > expose a > > > > > > > > > > >> helper > > > > > > > > > > >> > in > > > > > > > > > > >> > > > > > `ConsumerRecords` to make this a little > easier > > > > > though. > > > > > > > > > > >> > > > > > > > > > > > > > > > >> > > > > > Alternatively, if we think it is important > to > > > have > > > > > > this > > > > > > > > > > >> information > > > > > > > > > > >> > > > > exposed > > > > > > > > > > >> > > > > > directly, we could create batch APIs to > solve > > > the > > > > > > naming > > > > > > > > > > >> problem. > > > > > > > > > > >> > For > > > > > > > > > > >> > > > > > example: > > > > > > > > > > >> > > > > > > > > > > > > > > > >> > > > > > Map<TopicPartition, OffsetAndEpoch> > > positions(); > > > > > > > > > > >> > > > > > void seek(Map<TopicPartition, > OffsetAndEpoch> > > > > > > > positions); > > > > > > > > > > >> > > > > > > > > > > > > > > > >> > > > > > However, I'm actually leaning toward leaving > > the > > > > > > seek() > > > > > > > > and > > > > > > > > > > >> > > position() > > > > > > > > > > >> > > > > APIs > > > > > > > > > > >> > > > > > unchanged. Instead, we can add a new API to > > > search > > > > > for > > > > > > > > > offset > > > > > > > > > > by > > > > > > > > > > >> > > > > timestamp > > > > > > > > > > >> > > > > > or by offset/leader epoch. Let's say we call > > it > > > > > > > > > `findOffsets`. > > > > > > > > > > >> If > > > > > > > > > > >> > the > > > > > > > > > > >> > > > > user > > > > > > > > > > >> > > > > > hits a log truncation error, they can use > this > > > API > > > > > to > > > > > > > find > > > > > > > > > the > > > > > > > > > > >> > > closest > > > > > > > > > > >> > > > > > offset and then do a seek(). At the same > time, > > > we > > > > > > > > deprecate > > > > > > > > > > the > > > > > > > > > > >> > > > > > `offsetsForTimes` APIs. We now have two use > > > cases > > > > > > which > > > > > > > > > > require > > > > > > > > > > >> > > finding > > > > > > > > > > >> > > > > > offsets, so I think we should make this API > > > > general > > > > > > and > > > > > > > > > leave > > > > > > > > > > >> the > > > > > > > > > > >> > > door > > > > > > > > > > >> > > > > open > > > > > > > > > > >> > > > > > for future extensions. > > > > > > > > > > >> > > > > > > > > > > > > > > > >> > > > > > By the way, I'm unclear about the desire to > > move > > > > > part > > > > > > of > > > > > > > > > this > > > > > > > > > > >> > > > > functionality > > > > > > > > > > >> > > > > > to AdminClient. Guozhang suggested this > > > > previously, > > > > > > but > > > > > > > I > > > > > > > > > > think > > > > > > > > > > >> it > > > > > > > > > > >> > > only > > > > > > > > > > >> > > > > > makes sense for cross-cutting capabilities > > such > > > as > > > > > > topic > > > > > > > > > > >> creation. > > > > > > > > > > >> > If > > > > > > > > > > >> > > > we > > > > > > > > > > >> > > > > > have an API which is primarily useful by > > > > consumers, > > > > > > > then I > > > > > > > > > > think > > > > > > > > > > >> > > that's > > > > > > > > > > >> > > > > > where it should be exposed. The AdminClient > > also > > > > has > > > > > > its > > > > > > > > own > > > > > > > > > > API > > > > > > > > > > >> > > > > integrity > > > > > > > > > > >> > > > > > and should not become a dumping ground for > > > > advanced > > > > > > use > > > > > > > > > cases. > > > > > > > > > > >> I'll > > > > > > > > > > >> > > > > update > > > > > > > > > > >> > > > > > the KIP with the `findOffsets` API > suggested > > > > above > > > > > > and > > > > > > > we > > > > > > > > > can > > > > > > > > > > >> see > > > > > > > > > > >> > if > > > > > > > > > > >> > > > it > > > > > > > > > > >> > > > > > does a good enough job of keeping the API > > simple > > > > for > > > > > > > > common > > > > > > > > > > >> cases. > > > > > > > > > > >> > > > > > > > > > > > > > > > >> > > > > > Thanks, > > > > > > > > > > >> > > > > > Jason > > > > > > > > > > >> > > > > > > > > > > > > > > > >> > > > > > > > > > > > > > > > >> > > > > > On Sat, Jun 30, 2018 at 4:39 AM, Dong Lin < > > > > > > > > > > lindon...@gmail.com> > > > > > > > > > > >> > > wrote: > > > > > > > > > > >> > > > > > > > > > > > > > > > >> > > > > > > Hey Jason, > > > > > > > > > > >> > > > > > > > > > > > > > > > > >> > > > > > > Regarding seek(...), it seems that we want > > an > > > > API > > > > > > for > > > > > > > > user > > > > > > > > > > to > > > > > > > > > > >> > > > > initialize > > > > > > > > > > >> > > > > > > consumer with (offset, leaderEpoch) and > that > > > API > > > > > > > should > > > > > > > > > > allow > > > > > > > > > > >> > > > throwing > > > > > > > > > > >> > > > > > > PartitionTruncationException. Suppose we > > agree > > > > on > > > > > > > this, > > > > > > > > > then > > > > > > > > > > >> > > > > > > seekToNearest() is not sufficient because > it > > > > will > > > > > > > always > > > > > > > > > > >> swallow > > > > > > > > > > >> > > > > > > PartitionTruncationException. Here we have > > two > > > > > > > options. > > > > > > > > > The > > > > > > > > > > >> first > > > > > > > > > > >> > > > > option > > > > > > > > > > >> > > > > > is > > > > > > > > > > >> > > > > > > to add API offsetsForLeaderEpochs() to > > > translate > > > > > > > > > > (leaderEpoch, > > > > > > > > > > >> > > > offset) > > > > > > > > > > >> > > > > to > > > > > > > > > > >> > > > > > > offset. The second option is to have add > > > > > > seek(offset, > > > > > > > > > > >> > leaderEpoch). > > > > > > > > > > >> > > > It > > > > > > > > > > >> > > > > > > seems that second option may be more > simpler > > > > > because > > > > > > > it > > > > > > > > > > makes > > > > > > > > > > >> it > > > > > > > > > > >> > > > clear > > > > > > > > > > >> > > > > > that > > > > > > > > > > >> > > > > > > (offset, leaderEpoch) will be used to > > identify > > > > > > > > consumer's > > > > > > > > > > >> > position > > > > > > > > > > >> > > > in a > > > > > > > > > > >> > > > > > > partition. And user only needs to handle > > > > > > > > > > >> > > PartitionTruncationException > > > > > > > > > > >> > > > > > from > > > > > > > > > > >> > > > > > > the poll(). In comparison the first option > > > > seems a > > > > > > bit > > > > > > > > > > harder > > > > > > > > > > >> to > > > > > > > > > > >> > > use > > > > > > > > > > >> > > > > > > because user have to also handle the > > > > > > > > > > >> PartitionTruncationException > > > > > > > > > > >> > > if > > > > > > > > > > >> > > > > > > offsetsForLeaderEpochs() returns different > > > > offset > > > > > > from > > > > > > > > > > >> > > user-provided > > > > > > > > > > >> > > > > > > offset. What do you think? > > > > > > > > > > >> > > > > > > > > > > > > > > > > >> > > > > > > If we decide to add API seek(offset, > > > > leaderEpoch), > > > > > > > then > > > > > > > > we > > > > > > > > > > can > > > > > > > > > > >> > > decide > > > > > > > > > > >> > > > > > > whether and how to add API to translate > > > (offset, > > > > > > > > > > leaderEpoch) > > > > > > > > > > >> to > > > > > > > > > > >> > > > > offset. > > > > > > > > > > >> > > > > > It > > > > > > > > > > >> > > > > > > seems that this API will be needed by > > advanced > > > > > user > > > > > > to > > > > > > > > > don't > > > > > > > > > > >> want > > > > > > > > > > >> > > > auto > > > > > > > > > > >> > > > > > > offset reset (so that it can be notified) > > but > > > > > still > > > > > > > > wants > > > > > > > > > to > > > > > > > > > > >> > reset > > > > > > > > > > >> > > > > offset > > > > > > > > > > >> > > > > > > to closest. For those users if probably > > makes > > > > > sense > > > > > > to > > > > > > > > > only > > > > > > > > > > >> have > > > > > > > > > > >> > > the > > > > > > > > > > >> > > > > API > > > > > > > > > > >> > > > > > in > > > > > > > > > > >> > > > > > > AdminClient. offsetsForTimes() seems like > a > > > > common > > > > > > API > > > > > > > > > that > > > > > > > > > > >> will > > > > > > > > > > >> > be > > > > > > > > > > >> > > > > > needed > > > > > > > > > > >> > > > > > > by user's of consumer in general, so it > may > > be > > > > > more > > > > > > > > > > >> reasonable to > > > > > > > > > > >> > > > stay > > > > > > > > > > >> > > > > in > > > > > > > > > > >> > > > > > > the consumer API. I don't have a strong > > > opinion > > > > on > > > > > > > > whether > > > > > > > > > > >> > > > > > > offsetsForTimes() should be replaced by > API > > in > > > > > > > > > AdminClient. > > > > > > > > > > >> > > > > > > > > > > > > > > > > >> > > > > > > Though (offset, leaderEpoch) is needed to > > > > uniquely > > > > > > > > > identify > > > > > > > > > > a > > > > > > > > > > >> > > message > > > > > > > > > > >> > > > > in > > > > > > > > > > >> > > > > > > general, it is only needed for advanced > > users > > > > who > > > > > > has > > > > > > > > > turned > > > > > > > > > > >> on > > > > > > > > > > >> > > > unclean > > > > > > > > > > >> > > > > > > leader election, need to use seek(..), and > > > don't > > > > > > want > > > > > > > > auto > > > > > > > > > > >> offset > > > > > > > > > > >> > > > > reset. > > > > > > > > > > >> > > > > > > Most other users probably just want to > > enable > > > > auto > > > > > > > > offset > > > > > > > > > > >> reset > > > > > > > > > > >> > and > > > > > > > > > > >> > > > > store > > > > > > > > > > >> > > > > > > offset in Kafka. Thus we might want to > keep > > > the > > > > > > > existing > > > > > > > > > > >> > > offset-only > > > > > > > > > > >> > > > > APIs > > > > > > > > > > >> > > > > > > (e.g. seek() and position()) for most > users > > > > while > > > > > > > adding > > > > > > > > > new > > > > > > > > > > >> APIs > > > > > > > > > > >> > > for > > > > > > > > > > >> > > > > > > advanced users. And yes, it seems that we > > need > > > > new > > > > > > > name > > > > > > > > > for > > > > > > > > > > >> > > > position(). > > > > > > > > > > >> > > > > > > > > > > > > > > > > >> > > > > > > Though I think we need new APIs to carry > the > > > new > > > > > > > > > information > > > > > > > > > > >> > (e.g. > > > > > > > > > > >> > > > > > > leaderEpoch), I am not very sure how that > > > should > > > > > > look > > > > > > > > > like. > > > > > > > > > > >> One > > > > > > > > > > >> > > > > possible > > > > > > > > > > >> > > > > > > option is those APIs in KIP-232. Another > > > option > > > > is > > > > > > > > > something > > > > > > > > > > >> like > > > > > > > > > > >> > > > this: > > > > > > > > > > >> > > > > > > > > > > > > > > > > >> > > > > > > ````` > > > > > > > > > > >> > > > > > > class OffsetEpochs { > > > > > > > > > > >> > > > > > > long offset; > > > > > > > > > > >> > > > > > > int leaderEpoch; > > > > > > > > > > >> > > > > > > int partitionEpoch; // This may be > > needed > > > > > later > > > > > > as > > > > > > > > > > >> discussed > > > > > > > > > > >> > in > > > > > > > > > > >> > > > > > KIP-232 > > > > > > > > > > >> > > > > > > ... // Hopefully these are all we need > to > > > > > identify > > > > > > > > > message > > > > > > > > > > >> in > > > > > > > > > > >> > > > Kafka. > > > > > > > > > > >> > > > > > But > > > > > > > > > > >> > > > > > > if we need more then we can add new fields > > in > > > > this > > > > > > > > class. > > > > > > > > > > >> > > > > > > } > > > > > > > > > > >> > > > > > > > > > > > > > > > > >> > > > > > > OffsetEpochs offsetEpochs(TopicPartition); > > > > > > > > > > >> > > > > > > > > > > > > > > > > >> > > > > > > void seek(TopicPartition, OffsetEpochs); > > > > > > > > > > >> > > > > > > `````` > > > > > > > > > > >> > > > > > > > > > > > > > > > > >> > > > > > > > > > > > > > > > > >> > > > > > > Thanks, > > > > > > > > > > >> > > > > > > Dong > > > > > > > > > > >> > > > > > > > > > > > > > > > > >> > > > > > > > > > > > > > > > > >> > > > > > > On Fri, Jun 29, 2018 at 11:13 AM, Jason > > > > Gustafson > > > > > < > > > > > > > > > > >> > > > ja...@confluent.io> > > > > > > > > > > >> > > > > > > wrote: > > > > > > > > > > >> > > > > > > > > > > > > > > > > >> > > > > > > > Hey Dong, > > > > > > > > > > >> > > > > > > > > > > > > > > > > > >> > > > > > > > Thanks for the feedback. The first three > > > > points > > > > > > are > > > > > > > > > easy: > > > > > > > > > > >> > > > > > > > > > > > > > > > > > >> > > > > > > > 1. Yes, we should be consistent. > > > > > > > > > > >> > > > > > > > 2. Yes, I will add this. > > > > > > > > > > >> > > > > > > > 3. Yes, I think we should document the > > > changes > > > > > to > > > > > > > the > > > > > > > > > > >> committed > > > > > > > > > > >> > > > > offset > > > > > > > > > > >> > > > > >