[ https://issues.apache.org/jira/browse/KAFKA-10522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Chris Lambertus updated KAFKA-10522: ------------------------------------ Reporter: Luigi Berrettini (was: Luigi Berrettini) > Duplicate detection and max.in.flight.requests.per.connection details > --------------------------------------------------------------------- > > Key: KAFKA-10522 > URL: https://issues.apache.org/jira/browse/KAFKA-10522 > Project: Kafka > Issue Type: Wish > Reporter: Luigi Berrettini > Priority: Major > > I was looking at [https://github.com/apache/kafka/pull/3743] and I was > wondering if you could help me understand it better. > I saw that the > [Sender|https://github.com/apache/kafka/blob/2.6.0/clients/src/main/java/org/apache/kafka/clients/producer/internals/Sender.java#L602] > checks for a {{Errors.DUPLICATE_SEQUENCE_NUMBER}} but I was not able to find > where this error is triggered on the server side. > It seems to me that duplicate detection relies on checking if the sequence > number is more than {{lastPersistedSeq + 1}}. > If this is the case: > * why storing the metadata for the last batches and not just relying on the > sequence number of the last message persisted in the log? > * why limiting {{max.in.flight.requests.per.connection}} to a maximun value > of 5 if duplicates are still detected when metadata is not found (and > therefore with any number of max in flights)? -- This message was sent by Atlassian Jira (v8.3.4#803005)