Re: kafka cluster question

2024-01-19 Thread Haruki Okada
Hi.

Which server did you shutdown in testing?
If it was 192.168.20.223, that is natural kafka-consumer-groups script
fails because you passed only 192.168.20.223 to the bootstrap-server arg.

In HA setup, you have to pass multiple brokers (as the comma separated
string) to bootstrap-server so that the client can fetch initial metadata
from other servers even when one fails.

2024年1月20日(土) 0:30 Yavuz Sert :

> Hi all,
>
> I'm trying to do some tests about high availability on kafka v2.8.2
> I have 3 kafka brokers and 3 zookeeper instances.
> when i shutdown one of the kafka service only in one server i got this
> error:
>
> [root@node-223 ~]# /root/kafka_2.12-2.8.2/bin/kafka-consumer-groups.sh
> --bootstrap-server 192.168.20.223:9092 --group app2 --describe
>
> Error: Executing consumer group command failed due to
> org.apache.kafka.common.errors.TimeoutException:
> Call(callName=findCoordinator, deadlineMs=1705677946526, tries=47,
> nextAllowedTryMs=1705677946627) timed out at 1705677946527 after 47
> attempt(s)
> java.util.concurrent.ExecutionException:
> org.apache.kafka.common.errors.TimeoutException:
> Call(callName=findCoordinator, deadlineMs=1705677946526, tries=47,
> nextAllowedTryMs=1705677946627) timed out at 1705677946527 after 47
> attempt(s)
> at
>
> org.apache.kafka.common.internals.KafkaFutureImpl.wrapAndThrow(KafkaFutureImpl.java:45)
> at
>
> org.apache.kafka.common.internals.KafkaFutureImpl.access$000(KafkaFutureImpl.java:32)
> at
>
> org.apache.kafka.common.internals.KafkaFutureImpl$SingleWaiter.await(KafkaFutureImpl.java:89)
> at
>
> org.apache.kafka.common.internals.KafkaFutureImpl.get(KafkaFutureImpl.java:260)
> at
>
> kafka.admin.ConsumerGroupCommand$ConsumerGroupService.$anonfun$describeConsumerGroups$1(ConsumerGroupCommand.scala:550)
> at
> scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:286)
> at scala.collection.Iterator.foreach(Iterator.scala:943)
> at scala.collection.Iterator.foreach$(Iterator.scala:943)
> at scala.collection.AbstractIterator.foreach(Iterator.scala:1431)
> at scala.collection.IterableLike.foreach(IterableLike.scala:74)
> at scala.collection.IterableLike.foreach$(IterableLike.scala:73)
> at scala.collection.AbstractIterable.foreach(Iterable.scala:56)
> at scala.collection.TraversableLike.map(TraversableLike.scala:286)
> at scala.collection.TraversableLike.map$(TraversableLike.scala:279)
> at scala.collection.AbstractTraversable.map(Traversable.scala:108)
> at
>
> kafka.admin.ConsumerGroupCommand$ConsumerGroupService.describeConsumerGroups(ConsumerGroupCommand.scala:549)
>   at
>
> kafka.admin.ConsumerGroupCommand$ConsumerGroupService.collectGroupsOffsets(ConsumerGroupCommand.scala:565)
>   at
>
> kafka.admin.ConsumerGroupCommand$ConsumerGroupService.describeGroups(ConsumerGroupCommand.scala:368)
> at
> kafka.admin.ConsumerGroupCommand$.run(ConsumerGroupCommand.scala:73)
> at
> kafka.admin.ConsumerGroupCommand$.main(ConsumerGroupCommand.scala:60)
> at
> kafka.admin.ConsumerGroupCommand.main(ConsumerGroupCommand.scala)
> Caused by: org.apache.kafka.common.errors.TimeoutException:
> Call(callName=findCoordinator, deadlineMs=1705677946526, tries=47,
> nextAllowedTryMs=1705677946627) timed out at 1705677946527 after 47
> attempt(s)
> *Caused by: org.apache.kafka.common.errors.TimeoutException: Timed out
> waiting for a node assignment. Call: findCoordinator*
>
> kafka conf (for 1 server)
> broker.id=0
> listeners=PLAINTEXT://0.0.0.0:9092
> advertised.listeners=PLAINTEXT://192.168.20.223:9092
> num.network.threads=3
> num.io.threads=8
> socket.send.buffer.bytes=102400
> socket.receive.buffer.bytes=102400
> socket.request.max.bytes=104857600
> log.dirs=/root/kafkadir
> num.partitions=1
> num.recovery.threads.per.data.dir=1
> offsets.topic.replication.factor=1
> transaction.state.log.replication.factor=1
> transaction.state.log.min.isr=1
> log.retention.hours=1
> log.segment.bytes=104857600
> log.retention.check.interval.ms=30
> delete.topic.enable=true
> zookeeper.connection.timeout.ms=18000
> zookeeper.connect=192.168.20.223:2181,192.168.20.224:2181,
> 192.168.20.225:2181
> group.initial.rebalance.delay.ms=0
> max.request.size=104857600
> message.max.bytes=104857600
>
> How can i fix or troubleshoot the error?
>
> Thanks
>
> Yavuz
>


-- 

Okada Haruki
ocadar...@gmail.com



kafka cluster question

2024-01-19 Thread Yavuz Sert
Hi all,

I'm trying to do some tests about high availability on kafka v2.8.2
I have 3 kafka brokers and 3 zookeeper instances.
when i shutdown one of the kafka service only in one server i got this
error:

[root@node-223 ~]# /root/kafka_2.12-2.8.2/bin/kafka-consumer-groups.sh
--bootstrap-server 192.168.20.223:9092 --group app2 --describe

Error: Executing consumer group command failed due to
org.apache.kafka.common.errors.TimeoutException:
Call(callName=findCoordinator, deadlineMs=1705677946526, tries=47,
nextAllowedTryMs=1705677946627) timed out at 1705677946527 after 47
attempt(s)
java.util.concurrent.ExecutionException:
org.apache.kafka.common.errors.TimeoutException:
Call(callName=findCoordinator, deadlineMs=1705677946526, tries=47,
nextAllowedTryMs=1705677946627) timed out at 1705677946527 after 47
attempt(s)
at
org.apache.kafka.common.internals.KafkaFutureImpl.wrapAndThrow(KafkaFutureImpl.java:45)
at
org.apache.kafka.common.internals.KafkaFutureImpl.access$000(KafkaFutureImpl.java:32)
at
org.apache.kafka.common.internals.KafkaFutureImpl$SingleWaiter.await(KafkaFutureImpl.java:89)
at
org.apache.kafka.common.internals.KafkaFutureImpl.get(KafkaFutureImpl.java:260)
at
kafka.admin.ConsumerGroupCommand$ConsumerGroupService.$anonfun$describeConsumerGroups$1(ConsumerGroupCommand.scala:550)
at
scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:286)
at scala.collection.Iterator.foreach(Iterator.scala:943)
at scala.collection.Iterator.foreach$(Iterator.scala:943)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1431)
at scala.collection.IterableLike.foreach(IterableLike.scala:74)
at scala.collection.IterableLike.foreach$(IterableLike.scala:73)
at scala.collection.AbstractIterable.foreach(Iterable.scala:56)
at scala.collection.TraversableLike.map(TraversableLike.scala:286)
at scala.collection.TraversableLike.map$(TraversableLike.scala:279)
at scala.collection.AbstractTraversable.map(Traversable.scala:108)
at
kafka.admin.ConsumerGroupCommand$ConsumerGroupService.describeConsumerGroups(ConsumerGroupCommand.scala:549)
  at
kafka.admin.ConsumerGroupCommand$ConsumerGroupService.collectGroupsOffsets(ConsumerGroupCommand.scala:565)
  at
kafka.admin.ConsumerGroupCommand$ConsumerGroupService.describeGroups(ConsumerGroupCommand.scala:368)
at
kafka.admin.ConsumerGroupCommand$.run(ConsumerGroupCommand.scala:73)
at
kafka.admin.ConsumerGroupCommand$.main(ConsumerGroupCommand.scala:60)
at kafka.admin.ConsumerGroupCommand.main(ConsumerGroupCommand.scala)
Caused by: org.apache.kafka.common.errors.TimeoutException:
Call(callName=findCoordinator, deadlineMs=1705677946526, tries=47,
nextAllowedTryMs=1705677946627) timed out at 1705677946527 after 47
attempt(s)
*Caused by: org.apache.kafka.common.errors.TimeoutException: Timed out
waiting for a node assignment. Call: findCoordinator*

kafka conf (for 1 server)
broker.id=0
listeners=PLAINTEXT://0.0.0.0:9092
advertised.listeners=PLAINTEXT://192.168.20.223:9092
num.network.threads=3
num.io.threads=8
socket.send.buffer.bytes=102400
socket.receive.buffer.bytes=102400
socket.request.max.bytes=104857600
log.dirs=/root/kafkadir
num.partitions=1
num.recovery.threads.per.data.dir=1
offsets.topic.replication.factor=1
transaction.state.log.replication.factor=1
transaction.state.log.min.isr=1
log.retention.hours=1
log.segment.bytes=104857600
log.retention.check.interval.ms=30
delete.topic.enable=true
zookeeper.connection.timeout.ms=18000
zookeeper.connect=192.168.20.223:2181,192.168.20.224:2181,
192.168.20.225:2181
group.initial.rebalance.delay.ms=0
max.request.size=104857600
message.max.bytes=104857600

How can i fix or troubleshoot the error?

Thanks

Yavuz