Hi Garrett, Since your error message says "offset X" is out of range, it means that the offset was reset to because there was no data any more on topic partition "foo-0". I suspect that is because all the log segments got truncated and the topic partition contains empty list. It is less likely caused by KAFKA-5510 and hence offsets.retention.minutes may not help here.
Since you mentioned setting log.retention.hours=48 does not help, and that the input sample data may be a day or two before the new build goes out, I suspect there may be some messages with timestamps older than 48 hours published to the log, causing it to roll new segments and get deleted immediately: note that the Kafka brokers use the current system time to determine the diffs with the message timestamps. If that is the case it is not a Streams issue, not even a general Consumer issue, but a Kafka broker side log retention operation. What I'm not clear is that in your error message "X" is actually 0: this is quite weird that a consumer may auto-reset its position to 0, did you run some tools periodically to reset the offset to 0? Guozhang On Wed, Aug 9, 2017 at 7:16 AM, Garrett Barton <garrett.bar...@gmail.com> wrote: > I have a small test setup with a local zk/kafka server and a streams app > that loads sample data. The test setup is usually up for a day or two > before a new build goes out and its blown away and loaded from scratch. > > Lately I've seen that after a few hours the stream app will stop processing > and start spamming the logs with: > > org.apache.kafka.clients.consumer.internals.Fetcher: Fetch Offset 0 is out > of range for partition foo-0, resetting offset > org.apache.kafka.clients.consumer.internals.Fetcher: Fetch Offset 0 is out > of range for partition foo-0, resetting offset > org.apache.kafka.clients.consumer.internals.Fetcher: Fetch Offset 0 is out > of range for partition foo-0, resetting offset > > Pretty much sinks a core into spamming the logs. > > Restarting the application puts it right back in that broke state. > > I thought it was because of this: > https://issues.apache.org/jira/browse/KAFKA-5510 > So I set my log.retention.hours=48, and offsets.retention.minutes=10081, > which is huge compared to the total data retention time. Yet same error > occurred. > > Any ideas? > -- -- Guozhang