Hi, team. I’m using 0.8.1. I found some strange log repeatedly on server.log in one of my brokers and it keeps logging until now.
server.log ====================================================================================== ... [2014-06-09 10:41:47,402] ERROR Conditional update of path /brokers/topics/topicTRACE/partitions/1/state with data {"controller_epoch":19,"leader":2,"version":1,"leader_epoch":43,"isr":[4,2]} and expected version 439 failed due to org.apache.zookeeper.KeeperException$BadVersionException: KeeperErrorCode = BadVersion for /brokers/topics/topicTRACE/partitions/1/state (kafka.utils.ZkUtils$) [2014-06-09 10:41:47,402] INFO Partition [topicTRACE,1] on broker 2: Cached zkVersion [439] not equal to that in zookeeper, skip updating ISR (kafka.cluster.Partition) [2014-06-09 10:41:47,402] INFO Partition [topicDEBUG,0] on broker 2: Shrinking ISR for partition [topicDEBUG,0] from 1,3,2 to 2 (kafka.cluster.Partition) [2014-06-09 10:41:47,416] ERROR Conditional update of path /brokers/topics/topicDEBUG/partitions/0/state with data {"controller_epoch":19,"leader":2,"version":1,"leader_epoch":43,"isr":[2]} and expected version 1424 failed due to org.apache.zookeeper.KeeperException$BadVersionException: KeeperErrorCode = BadVersion for /brokers/topics/topicDEBUG/partitions/0/state (kafka.utils.ZkUtils$) [2014-06-09 10:41:47,432] INFO Partition [topicDEBUG,0] on broker 2: Cached zkVersion [1424] not equal to that in zookeeper, skip updating ISR (kafka.cluster.Partition) [2014-06-09 10:41:47,432] INFO Partition [topicCDR,3] on broker 2: Shrinking ISR for partition [topicCDR,3] from 4,1,2 to 2 (kafka.cluster.Partition) [2014-06-09 10:41:47,435] ERROR Conditional update of path /brokers/topics/topicCDR/partitions/3/state with data {"controller_epoch":19,"leader":2,"version":1,"leader_epoch":46,"isr":[2]} and expected version 541 failed due to org.apache.zookeeper.KeeperException$BadVersionException: KeeperErrorCode = BadVersion for /brokers/topics/topicCDR/partitions/3/state (kafka.utils.ZkUtils$) [2014-06-09 10:41:47,435] INFO Partition [topicCDR,3] on broker 2: Cached zkVersion [541] not equal to that in zookeeper, skip updating ISR (kafka.cluster.Partition) [2014-06-09 10:41:48,426] INFO Partition [topicTRACE,1] on broker 2: Shrinking ISR for partition [topicTRACE,1] from 4,3,2 to 4,2 (kafka.cluster.Partition) ... ================================================================================================= and found some error and warning in controller.log controller.log ====================================================================================== ... [2014-06-09 10:42:03,962] WARN [Controller-3-to-broker-1-send-thread], Controller 3 fails to send a request to broker id:1,host:c-ccp-tk1-a58,port:9091 (kafka.controller.RequestSendThread) java.net.SocketTimeoutException at sun.nio.ch.SocketAdaptor$SocketInputStream.read(SocketAdaptor.java:229) at sun.nio.ch.ChannelInputStream.read(ChannelInputStream.java:103) at java.nio.channels.Channels$ReadableByteChannelImpl.read(Channels.java:385) at kafka.utils.Utils$.read(Utils.scala:375) at kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.scala:54) at kafka.network.Receive$class.readCompletely(Transmission.scala:56) at kafka.network.BoundedByteBufferReceive.readCompletely(BoundedByteBufferReceive.scala:29) at kafka.network.BlockingChannel.receive(BlockingChannel.scala:100) at kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:146) at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:51) [2014-06-09 10:42:03,964] ERROR [Controller-3-to-broker-1-send-thread], Controller 3 epoch 21 failed to send UpdateMetadata request with correlation id 1 to broker id:1,host:c-ccp-tk1-a58,port:9091. Reconnecting to broker. (kafka.controller.RequestSendThread) java.nio.channels.ClosedChannelException at kafka.network.BlockingChannel.send(BlockingChannel.scala:89) at kafka.controller.RequestSendThread.liftedTree1$1(ControllerChannelManager.scala:132) at kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:131) at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:51) ... [2014-06-09 10:42:38,064] WARN [OfflinePartitionLeaderSelector]: No broker in ISR is alive for [topicTRACE,0]. Elect leader 3 from live brokers 3. There's potential data loss. (kafka.controller.OfflinePartitionLeaderSelector) ... ================================================================================================= Why is this happen? Is there any possibilities data loss? To normalize my brokers, What I have to do? Do I have to restart this broker? Thanks in advance.