Apurva Mehta created KAFKA-5793:
-----------------------------------

             Summary: Tighten up situations where OutOfOrderSequence may be 
returned
                 Key: KAFKA-5793
                 URL: https://issues.apache.org/jira/browse/KAFKA-5793
             Project: Kafka
          Issue Type: Bug
    Affects Versions: 0.11.0.0
            Reporter: Apurva Mehta
            Assignee: Apurva Mehta
             Fix For: 1.0.0


Details of the problem are provided here: 
https://cwiki.apache.org/confluence/display/KAFKA/Kafka+Exactly+Once+-+Solving+the+problem+of+spurious+OutOfOrderSequence+errors

A quick summary follows:

In the discussion of KIP-185: Make exactly once in order delivery per partition 
the default producer setting, the following point regarding the 
OutOfOrderSequenceException was raised:

1. The OutOfOrderSequenceException indicates that there has been data loss on 
the broker.. ie. a previously acknowledged message no longer exists. For most 
part, this should only occur in rare situations (simultaneous power outages, 
multiple disk losses, software bugs resulting in data corruption, etc.).
2. However, there is another perfectly normal scenario where data is removed: 
in particular, data could be deleted because it is old and crosses the 
retention threshold.
Hence, if a producer remains inactive for longer than a topic's retention 
period, we could get an OutOfOrderSequence which is a false positive: the data 
is removed through valid processes, and this isn't an error.
3. We would like to eliminate the possibility of getting spurious 
OutOfOrderSequenceExceptions – when you get it, it should always mean data loss 
and should be taken very seriously. 




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to