Hi,

I am using flink 12.1 to consume data from kafka in a streaming job. Using
the flink-connector-kafka_2.12:1.12.1. Kafka broker version is 2.2.1
 In logs I see warnings like this -

2021-08-26 13:36:49,903 WARN
org.apache.flink.streaming.connectors.kafka.internals.KafkaFetcher [] -
Committing offsets to Kafka failed. This does not compromise Flink's
checkpoints.
org.apache.kafka.clients.consumer.CommitFailedException: Commit cannot be
completed since the group has already rebalanced and assigned the
partitions to another member.
This means that the time between subsequent calls to poll() was longer than
the configured max.poll.interval.ms, which typically implies that the poll
loop is spending too much time message processing.
You can address this either by increasing max.poll.interval.ms or by
reducing the maximum size of batches returned in poll() with
max.poll.records.
    at
org.apache.kafka.clients.consumer.internals.ConsumerCoordinator$OffsetCommitResponseHandler.handle(ConsumerCoordinator.java:840)
    at
org.apache.kafka.clients.consumer.internals.ConsumerCoordinator$OffsetCommitResponseHandler.handle(ConsumerCoordinator.java:790)

    at
org.apache.kafka.clients.consumer.internals.AbstractCoordinator$CoordinatorResponseHandler.onSuccess(AbstractCoordinator.java:910)

    at
org.apache.kafka.clients.consumer.internals.AbstractCoordinator$CoordinatorResponseHandler.onSuccess(AbstractCoordinator.java:890)

    at
org.apache.kafka.clients.consumer.internals.RequestFuture$1.onSuccess(RequestFuture.java:204)

    at
org.apache.kafka.clients.consumer.internals.RequestFuture.fireSuccess(RequestFuture.java:167)

    at
org.apache.kafka.clients.consumer.internals.RequestFuture.complete(RequestFuture.java:127)

    at
org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient$RequestFutureCompletionHandler.fireCompletion(ConsumerNetworkClient.java:575)
    at
org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.firePendingCompletedRequests(ConsumerNetworkClient.java:389)

    at
org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:297)

    at
org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:236)

    at
org.apache.kafka.clients.consumer.KafkaConsumer.pollForFetches(KafkaConsumer.java:1256)
    at
org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1200)
    at
org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1135)

    at
org.apache.flink.streaming.connectors.kafka.internals.KafkaConsumerThread.run(KafkaConsumerThread.java:258)


I understand that this might not cause an issue as checkpointing is not
impacted, however metrics monitoring might as I am using burrow to monitor
group offsets. I have already tried to change below properties in kafka
producer configs -

        kafkaProps.setProperty("max.poll.interval.ms","900000");
        kafkaProps.setProperty("max.poll.records","200");
        kafkaProps.setProperty("heartbeat.interval.ms","1000");
        kafkaProps.setProperty("request.timeout.ms","40000");
        kafkaProps.setProperty("session.timeout.ms","10000");
But the warnings are still present in the logs.

In addition I see this error just before this warn -
ERROR org.apache.kafka.clients.consumer.internals.ConsumerCoordinator [] -
[Consumer clientId=consumer-3, groupId=xxx] Offset commit failed on
partition xxx-1 at offset 33651:
The coordinator is not aware of this member.

The code uses watermarkstrategy to extract timestamp and emit watermark.

Any clue is much appreciated.

Thanks,
Hemant

Reply via email to