[
https://issues.apache.org/jira/browse/KAFKA-2477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14952127#comment-14952127
]
Chinmay Soman edited comment on KAFKA-2477 at 10/11/15 1:49 AM:
----------------------------------------------------------------
[~ewencp] I'm attaching the screenshot of Max lag observed for the different
brokers which describes the behaviour.
Also here's the pertinent log:
========
[2015-10-10 22:17:17,759] 3171793337 [kafka-request-handler-0] WARN
kafka.server.ReplicaManager - [Replica Manager on Broker 70]: Error when
processing fetch request for partition [...topic...,62] offset 91963211 from
follower with correlation id 176614372. Possible cause: Request for offset
91963211 but we only have log segments in the range 55923986 to 91963210.
[2015-10-10 22:17:17,759] 3171793337 [kafka-request-handler-4] WARN
kafka.server.ReplicaManager - [Replica Manager on Broker 70]: Error when
processing fetch request for partition [...topic...,62] offset 91963211 from
follower with correlation id 152788081. Possible cause: Request for offset
91963211 but we only have log segments in the range 55923986 to 91963210.
[2015-10-10 22:17:20,256] 3171795834 [kafka-scheduler-4] INFO
kafka.cluster.Partition - Partition [...topic...,62] on broker 70: Shrinking
ISR for partition [hp.event.user.driver_app.experiment,62] from 70,69,71 to 70
========
was (Author: cpsoman):
Max lag observed for the different brokers
> Replicas spuriously deleting all segments in partition
> ------------------------------------------------------
>
> Key: KAFKA-2477
> URL: https://issues.apache.org/jira/browse/KAFKA-2477
> Project: Kafka
> Issue Type: Bug
> Affects Versions: 0.8.2.1
> Reporter: HÃ¥kon Hitland
> Assignee: Jiangjie Qin
> Fix For: 0.9.0.0
>
> Attachments: Screen Shot 2015-10-10 at 6.54.44 PM.png, kafka_log.txt,
> kafka_log_trace.txt
>
>
> We're seeing some strange behaviour in brokers: a replica will sometimes
> schedule all segments in a partition for deletion, and then immediately start
> replicating them back, triggering our check for under-replicating topics.
> This happens on average a couple of times a week, for different brokers and
> topics.
> We have per-topic retention.ms and retention.bytes configuration, the topics
> where we've seen this happen are hitting the size limit.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)