[
https://issues.apache.org/jira/browse/KAFKA-4168?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Eno Thereska updated KAFKA-4168:
--------------------------------
Issue Type: Improvement (was: Sub-task)
Parent: (was: KAFKA-3776)
> More precise accounting of memory usage
> ---------------------------------------
>
> Key: KAFKA-4168
> URL: https://issues.apache.org/jira/browse/KAFKA-4168
> Project: Kafka
> Issue Type: Improvement
> Components: streams
> Affects Versions: 0.10.1.0
> Reporter: Eno Thereska
> Fix For: 0.10.2.0
>
>
> Right now, the cache.max.bytes.buffering parameter controls the size of the
> cache used. Specifically the size includes the size of the values stored in
> the cache plus basic overheads, such as key size, all the LRU entry sizes,
> etc. However, we could be more fine-grained in the memory accounting and add
> up the size of hash sets, hash maps and their entries more precisely. For
> example, currently a dirty entry is placed into a dirty keys set, but we do
> not account for the size of that set in the memory consumption calculation.
> It is likely this falls under "memory management" rather than "buffer cache
> management".
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)