[ https://issues.apache.org/jira/browse/KAFKA-3892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15346321#comment-15346321 ]
Noah Sloan commented on KAFKA-3892: ----------------------------------- I can reproduce but I don't have code I can share (and you would need 0.9 brokers with a lot of topics to try it against.) The only way I know it is happening is because: 1. We have brokers with large metadata. 2. We have many producers/consumers in 1 VM. 3. That VM is memory constrained. Then the VM OOMs and the heap dump shows unrequested metadata. This could be happening to all 0.9 producers/consumers, but without all 3 conditions, you would never notice. There isn't any way to see this is happening as a client without a heap dump or a debugger. > Clients retain metadata for non-subscribed topics > ------------------------------------------------- > > Key: KAFKA-3892 > URL: https://issues.apache.org/jira/browse/KAFKA-3892 > Project: Kafka > Issue Type: Bug > Components: clients > Affects Versions: 0.9.0.1 > Reporter: Noah Sloan > > After upgrading to 0.9.0.1 from 0.8.2 (and adopting the new consumer and > producer classes,) we noticed services with small heap crashing due to > OutOfMemoryErrors. These services contained many producers and consumers (~20 > total) and were connected to brokers with >2000 topics and over 10k > partitions. Heap dumps revealed that each client had 3.3MB of Metadata > retained in their Cluster, with references to topics that were not being > produced or subscribed to. While the services were running with 128MB of heap > prior to the upgrade, we to had increased max heap to 200MB to accommodate > all the extra data. > While this is not technically a memory leak, it does impose a significant > overhead on clients when connected to a large cluster. -- This message was sent by Atlassian JIRA (v6.3.4#6332)