Finbarr Naughton created KAFKA-7716:
---------------------------------------
Summary: Unprocessed messages when Broker fails
Key: KAFKA-7716
URL: https://issues.apache.org/jira/browse/KAFKA-7716
Project: Kafka
Issue Type: Bug
Components: core, streams
Affects Versions: 2.0.1, 1.0.0
Reporter: Finbarr Naughton
This occurs when running on Kubernetes on bare metal.
A Streams application with a single topology listening to two input topics A
and B. A is read as a GlobalKTable, B as a KStream. The topology joins the
stream to the GKTable and writes an updated message to topic A. The application
is configured to use exactly_once processing.
There are three worker nodes. Kafka brokers are deployed as a statefulset on
the three nodes using the helm chart from here
-[https://github.com/helm/charts/tree/master/incubator/kafka]
The application has three instances spread across the three nodes.
During a test, topic A is pre-populated with 50k messages over 5 minutes. Then
50k messages with the same key-set are sent to topic B over 5 minutes. The
expected behaviour is that Topic A will contain 50k updated messages
afterwards. While all brokers are available this is the case, even when one of
the application pods is deleted.
When a broker fails, however, a few expected updated messages fail to appear on
topic A despite their existence on topic B.
More complete description here -
[https://stackoverflow.com/questions/53557247/some-events-unprocessed-by-kafka-streams-application-on-kubernetes-on-bare-metal]
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)