[ 
https://issues.apache.org/jira/browse/KAFKA-14449?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17644466#comment-17644466
 ] 

Swathi Mocharla commented on KAFKA-14449:
-----------------------------------------

[~jsancio] , could you take a look?

> Brokers not re-joining the ISR list and stuck at started until all the 
> brokers restart
> --------------------------------------------------------------------------------------
>
>                 Key: KAFKA-14449
>                 URL: https://issues.apache.org/jira/browse/KAFKA-14449
>             Project: Kafka
>          Issue Type: Bug
>          Components: core
>    Affects Versions: 3.2.0
>            Reporter: Swathi Mocharla
>            Priority: Critical
>
> hi,
> We are upgrading a 3 broker cluster (1001,1002,1003) from 3.1.0 to 3.2.0.
> During upgrade, it is noticed that when 1003 is restarted, it doesn't join 
> back the ISR list and the broker is stuck. Same is the case with 1002.
> Only when 1001 is restrarted, 1003,1002 re-join the ISR list and start 
> replicating data.
>  
> {code:java}
> {"type":"log", "host":"kf-pl47-me8-2", "level":"INFO", 
> "neid":"kafka-b352b4f8cf4447e9a73d9e7ef3ec746c", "system":"kafka", 
> "time":"2022-12-06T10:07:30.386", "timezone":"UTC", "log":{"message":"main - 
> kafka.server.KafkaServer - [KafkaServer id=1003] started"}}
> {"type":"log", "host":"kf-pl47-me8-2", "level":"INFO", 
> "neid":"kafka-b352b4f8cf4447e9a73d9e7ef3ec746c", "system":"kafka", 
> "time":"2022-12-06T10:07:30.442", "timezone":"UTC", 
> "log":{"message":"data-plane-kafka-request-handler-1 - state.change.logger - 
> [Broker id=1003] Add 397 partitions and deleted 0 partitions from metadata 
> cache in response to UpdateMetadata request sent by controller 1002 epoch 18 
> with correlation id 0"}}
> {"type":"log", "host":"kf-pl47-me8-2", "level":"INFO", 
> "neid":"kafka-b352b4f8cf4447e9a73d9e7ef3ec746c", "system":"kafka", 
> "time":"2022-12-06T10:07:30.448", "timezone":"UTC", 
> "log":{"message":"BrokerToControllerChannelManager broker=1003 name=alterIsr 
> - kafka.server.BrokerToControllerRequestThread - 
> [BrokerToControllerChannelManager broker=1003 name=alterIsr]: Recorded new 
> controller, from now on will use broker 
> kf-pl47-me8-1.kf-pl47-me8-headless.nc0968-admin-ns.svc.cluster.local:9092 
> (id: 1002 rack: null)"}}
> {"type":"log", "host":"kf-pl47-me8-2", "level":"ERROR", 
> "neid":"kafka-b352b4f8cf4447e9a73d9e7ef3ec746c", "system":"kafka", 
> "time":"2022-12-06T10:07:30.451", "timezone":"UTC", 
> "log":{"message":"data-plane-kafka-network-thread-1003-ListenerName(PLAINTEXT)-PLAINTEXT-1
>  - kafka.network.Processor - Closing socket for 
> 192.168.216.11:9092-192.168.199.100:53778-0 because of error"}}
> org.apache.kafka.common.errors.InvalidRequestException: Error getting request 
> for apiKey: LEADER_AND_ISR, apiVersion: 6, connectionId: 
> 192.168.216.11:9092-192.168.199.100:53778-0, listenerName: 
> ListenerName(PLAINTEXT), principal: User:ANONYMOUS
> org.apache.kafka.common.errors.InvalidRequestException: Error getting request 
> for apiKey: LEADER_AND_ISR, apiVersion: 6, connectionId: 
> 192.168.216.11:9092-192.168.235.153:46282-461, listenerName: 
> ListenerName(PLAINTEXT), principal: User:ANONYMOUS
> Caused by: org.apache.kafka.common.errors.UnsupportedVersionException: Can't 
> read version 6 of LeaderAndIsrTopicState
> {"type":"log", "host":"kf-pl47-me8-2", "level":"INFO", 
> "neid":"kafka-b352b4f8cf4447e9a73d9e7ef3ec746c", "system":"kafka", 
> "time":"2022-12-06T10:12:50.916", "timezone":"UTC", 
> "log":{"message":"controller-event-thread - kafka.controller.KafkaController 
> - [Controller id=1003] 1003 successfully elected as the controller. Epoch 
> incremented to 20 and epoch zk version is now 20"}}
> {"type":"log", "host":"kf-pl47-me8-2", "level":"INFO", 
> "neid":"kafka-b352b4f8cf4447e9a73d9e7ef3ec746c", "system":"kafka", 
> "time":"2022-12-06T10:12:50.917", "timezone":"UTC", 
> "log":{"message":"controller-event-thread - kafka.controller.KafkaController 
> - [Controller id=1003] Registering handlers"}}
> {code}
>  
> This possibly was introduced by KAFKA-13587.
> In the below snapshot during the upgrade, at 16:05:15 UTC 2022, 1001 was 
> restarting and both 1002 and 1003 were already up and running (after the 
> upgrade from 3.1.0 to 3.2.0), but did not manage to re-join the ISRs. 
> {code:java}
> Wed Dec  7 16:05:15 UTC 2022
> Topic: test     TopicId: L6Yj_Nf9RrirNhFQzvXODw PartitionCount: 2       
> ReplicationFactor: 3    Configs: 
> compression.type=producer,min.insync.replicas=1,cleanup.policy=delete,flush.ms=1000,segment.bytes=100000000,flush.messages=10000,max.message.bytes=1000012,index.interval.bytes=4096,unclean.leader.election.enable=false,retention.bytes=1000000000,segment.index.bytes=10485760
>         Topic: test     Partition: 0    Leader: none    Replicas: 
> 1002,1003,1001        Isr: 1001
>         Topic: test     Partition: 1    Leader: none    Replicas: 
> 1001,1002,1003        Isr: 1001
> Wed Dec  7 16:05:33 UTC 2022
> Topic: test     TopicId: L6Yj_Nf9RrirNhFQzvXODw PartitionCount: 2       
> ReplicationFactor: 3    Configs: 
> compression.type=producer,min.insync.replicas=1,cleanup.policy=delete,flush.ms=1000,segment.bytes=100000000,flush.messages=10000,max.message.bytes=1000012,index.interval.bytes=4096,unclean.leader.election.enable=false,retention.bytes=1000000000,segment.index.bytes=10485760
>         Topic: test     Partition: 0    Leader: 1001    Replicas: 
> 1002,1003,1001        Isr: 1001,1002,1003
>         Topic: test     Partition: 1    Leader: 1001    Replicas: 
> 1001,1002,1003        Isr: 1001,1002,1003{code}
> Is there anything the user needs to do explicitly to work around this issue?



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to