[
https://issues.apache.org/jira/browse/KAFKA-3892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15346321#comment-15346321
]
Noah Sloan commented on KAFKA-3892:
-----------------------------------
I can reproduce but I don't have code I can share (and you would need 0.9
brokers with a lot of topics to try it against.) The only way I know it is
happening is because:
1. We have brokers with large metadata.
2. We have many producers/consumers in 1 VM.
3. That VM is memory constrained.
Then the VM OOMs and the heap dump shows unrequested metadata.
This could be happening to all 0.9 producers/consumers, but without all 3
conditions, you would never notice. There isn't any way to see this is
happening as a client without a heap dump or a debugger.
> Clients retain metadata for non-subscribed topics
> -------------------------------------------------
>
> Key: KAFKA-3892
> URL: https://issues.apache.org/jira/browse/KAFKA-3892
> Project: Kafka
> Issue Type: Bug
> Components: clients
> Affects Versions: 0.9.0.1
> Reporter: Noah Sloan
>
> After upgrading to 0.9.0.1 from 0.8.2 (and adopting the new consumer and
> producer classes,) we noticed services with small heap crashing due to
> OutOfMemoryErrors. These services contained many producers and consumers (~20
> total) and were connected to brokers with >2000 topics and over 10k
> partitions. Heap dumps revealed that each client had 3.3MB of Metadata
> retained in their Cluster, with references to topics that were not being
> produced or subscribed to. While the services were running with 128MB of heap
> prior to the upgrade, we to had increased max heap to 200MB to accommodate
> all the extra data.
> While this is not technically a memory leak, it does impose a significant
> overhead on clients when connected to a large cluster.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)