Amit Khandelwal created KAFKA-9613:
--------------------------------------

             Summary: orruptRecordException: Found record size 0 smaller than 
minimum record overhead
                 Key: KAFKA-9613
                 URL: https://issues.apache.org/jira/browse/KAFKA-9613
             Project: Kafka
          Issue Type: Bug
            Reporter: Amit Khandelwal


20200224;21:01:38: [2020-02-24 21:01:38,615] ERROR [ReplicaManager broker=0] 
Error processing fetch with max size 1048576 from consumer on partition 
SANDBOX.BROKER.NEWORDER-0: (fetchOffset=211886, logStartOffset=-1, 
maxBytes=1048576, currentLeaderEpoch=Optional.empty) 
(kafka.server.ReplicaManager)

20200224;21:01:38: org.apache.kafka.common.errors.CorruptRecordException: Found 
record size 0 smaller than minimum record overhead (14) in file 
/data/tmp/kafka-topic-logs/SANDBOX.BROKER.NEWORDER-0/00000000000000000000.log.

20200224;21:05:48: [2020-02-24 21:05:48,711] INFO [GroupMetadataManager 
brokerId=0] Removed 0 expired offsets in 1 milliseconds. 
(kafka.coordinator.group.GroupMetadataManager)

20200224;21:10:22: [2020-02-24 21:10:22,204] INFO [GroupCoordinator 0]: Member 
xxxxxxxx_011-9e61d2c9-ce5a-4231-bda1-f04e6c260dc0-StreamThread-1-consumer-27768816-ee87-498f-8896-191912282d4f
 in group yyyyyyyyy_011 has failed, removing it from the group 
(kafka.coordinator.group.GroupCoordinator)

 

[https://stackoverflow.com/questions/60404510/kafka-broker-issue-replica-manager-with-max-size#]

 

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to