[ https://issues.apache.org/jira/browse/KAFKA-19519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=18007996#comment-18007996 ]
Lan Ding commented on KAFKA-19519: ---------------------------------- Hi [~showuon], if you're not working on this, may I take it? Thanks. > Introduce a new config for group coordinator max record size > ------------------------------------------------------------ > > Key: KAFKA-19519 > URL: https://issues.apache.org/jira/browse/KAFKA-19519 > Project: Kafka > Issue Type: Improvement > Reporter: Luke Chen > Priority: Major > > In KAFKA-19427, there's a use case that when there is a consumer group > subscribes huge amount of topics/partitions, when this group is rebalanced, > and then the coordinator broker stores the assignment of this group to > __consumer_offsets, it will throw an error RecordTooLargeException . > > Currently, to resolve this issue, the only solution is to increase the > broker-level `message.max.bytes` config. But the side effect of this change > is it potentially allows all topics without override the topic level > {{max.message.bytes}} config, will now allow higher message size. > > We could introduce a new config to drive the value used by the group > coordinator - e.g group.coordinator.append.max.bytes - instead of relying on > the broker message.max.bytes. This would be used to set the max bytes at the > topic level when the topic is created. -- This message was sent by Atlassian Jira (v8.20.10#820010)