John Lu created KAFKA-6980:
------------------------------
Summary: Recommended MaxDirectMemorySize for consumers
Key: KAFKA-6980
URL: https://issues.apache.org/jira/browse/KAFKA-6980
Project: Kafka
Issue Type: Wish
Components: consumer, documentation
Affects Versions: 0.10.2.0
Environment: CloudFoundry
Reporter: John Lu
We are observing that when MaxDirectMemorySize is set too low, our Kafka
consumer threads are failing and encountering the following exception:
{{java.lang.OutOfMemoryError: Direct buffer memory}}
Is there a way to estimate how much direct memory is required for optimal
performance? In the documentation, it is suggested that the amount of memory
required is [Number of Partitions * max.partition.fetch.bytes].
When we pick a value slightly above that, we no longer encounter the error, but
if we double or triple the number, our throughput improves drastically. So we
are wondering if there is another setting or parameter to consider?
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)