Hi Garrett,

Since your error message says "offset X" is out of range, it means that the
offset was reset to because there was no data any more on topic partition
"foo-0". I suspect that is because all the log segments got truncated and
the topic partition contains empty list. It is less likely caused by
KAFKA-5510 and hence offsets.retention.minutes may not help here.

Since you mentioned setting log.retention.hours=48 does not help, and that
the input sample data may be a day or two before the new build goes out, I
suspect there may be some messages with timestamps older than 48 hours
published to the log, causing it to roll new segments and get deleted
immediately: note that the Kafka brokers use the current system time to
determine the diffs with the message timestamps. If that is the case it is
not a Streams issue, not even a general Consumer issue, but a Kafka broker
side log retention operation.

What I'm not clear is that in your error message "X" is actually 0: this is
quite weird that a consumer may auto-reset its position to 0, did you run
some tools periodically to reset the offset to 0?


Guozhang





On Wed, Aug 9, 2017 at 7:16 AM, Garrett Barton <garrett.bar...@gmail.com>
wrote:

> I have a small test setup with a local zk/kafka server and a streams app
> that loads sample data.  The test setup is usually up for a day or two
> before a new build goes out and its blown away and loaded from scratch.
>
> Lately I've seen that after a few hours the stream app will stop processing
> and start spamming the logs with:
>
> org.apache.kafka.clients.consumer.internals.Fetcher: Fetch Offset 0 is out
> of range for partition foo-0, resetting offset
> org.apache.kafka.clients.consumer.internals.Fetcher: Fetch Offset 0 is out
> of range for partition foo-0, resetting offset
> org.apache.kafka.clients.consumer.internals.Fetcher: Fetch Offset 0 is out
> of range for partition foo-0, resetting offset
>
> Pretty much sinks a core into spamming the logs.
>
> Restarting the application puts it right back in that broke state.
>
> I thought it was because of this:
> https://issues.apache.org/jira/browse/KAFKA-5510
> So I set my log.retention.hours=48, and offsets.retention.minutes=10081,
> which is huge compared to the total data retention time.  Yet same error
> occurred.
>
> Any ideas?
>



-- 
-- Guozhang

Reply via email to