Sunil,
This has to do with Kafka's behavior of being persistent and using the
broker's filesystem as the storage mechanism for the commit log. In
modern operating systems a watermark of *85%* of the available RAM is
dedicated to page cache and therefore, with Kafka running in a machine
with *32GB* of RAM *~28-30GB* will be used to store the data.
Reason why the JVM heap doesn't need to be higher than *~6GB*. All the
data is stored off-heap anyway ¯\_(ツ)_/¯
Thanks,
-- Ricardo
On 6/20/20 4:18 AM, sunil chaudhari wrote:
Hi,
I was going through this document.
https://docs.confluent.io/current/kafka/deployment.html
“ does not require setting heap sizes more than 6 GB. This will result in a
file system cache of up to 28-30 GB on a 32 GB machine.”
Can someone please put focus on above statement? Its bit unclear to me as
why file system cache will reach to 28-30 GB ?
I have 64 GB machine for each broker. Should I stick to 6 GB still? Or I
can assign some more?
Regards,
Sunil.