[
https://issues.apache.org/jira/browse/KAFKA-19024?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17939117#comment-17939117
]
Lan Ding commented on KAFKA-19024:
----------------------------------
In the current implementation, the `group.share.max.groups` is only used when
setting the maxEntries for the ShareSessionCache.
Perhaps we could introduce a new error type `MAX_SHARE_GROUP_SIZE_REACHED`.
When processing a Heartbeat request, if the group ID does not exist and the
`group.share.max.groups` limit is reached, this exception would be thrown.
Clients could then catch this exception and handle it accordingly (e.g.,
logging an error message). Do you think this approach is feasible?
> Enhance the client behaviour when it tries to exceed the
> `group.share.max.groups`
> ---------------------------------------------------------------------------------
>
> Key: KAFKA-19024
> URL: https://issues.apache.org/jira/browse/KAFKA-19024
> Project: Kafka
> Issue Type: Sub-task
> Reporter: Sanskar Jhajharia
> Assignee: Lan Ding
> Priority: Minor
>
> For share groups we use the `group.share.max.groups` config to define the
> number of max share groups we allow. However, when we exceed the same, the
> client logs do not specify any such error and simply do not consume. The
> group doesn't get created but the client continues to send Heartbeats hoping
> for one of the existing groups to shut down and allowing it to form a group.
> Having a log or an exception in the client logs will help them debug such
> situations accurately.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)