Hi Yoshimo, Since the topic name is "price-publisher-price-events-changelog" I think it is an internal Streams topic indeed, more concretely it is a changelog topic associated with a store named "price events" (assuming your application.id is set to "price-publisher"). The changelog topic usage is documented here:
https://kafka.apache.org/21/documentation/streams/architecture#streams_architecture_recovery Guozhang On Tue, Feb 26, 2019 at 2:01 AM Yoshimo <[email protected]> wrote: > Hello, > > during our resilience tests, a message is logged with an error. The log > line is the following: > > 2019-02-25 14:18:35.100+0000 [kafka-producer-network-thread | > str1-StreamThread-1-producer] ERROR o.a.k.s.p.i.RecordCollectorImpl - task > [0_0] Error sending record (key \x003\x9F\xF0 value [10, 7, 51, 51, 56, 51, > 50, 56, 48, 18, 28, 50, 48, 49, 50.... -110, 45, 32, -61, -56, 25] > timestamp 1551104224970) to topic price-publisher-price-events-changelog > due to org.apache.kafka.common.errors.TimeoutException: Expiring 3 > record(s) for price-publisher-price-events-changelog-0: 30034 ms has passed > since last attempt plus backoff time; No more records will be sent and no > more offsets will be recorded for this task. > > The message has a huge list of numbers which is cropped in the above > example. The Timeout error is probably OK however the thing is that we > don't publish this message in our application code. Is it this message > published by Kafka Streams internally? > > Our application uses a Transformer, with three distinct stores, a > punctuator that can possibly delete the values from the above stores and it > has exactly once enabled. > > Thank you for the help, > Yiannis -- -- Guozhang
