Juan Olivares created KAFKA-8448:
------------------------------------

             Summary: Too many kafka.log.Log instances (Memory Leak)
                 Key: KAFKA-8448
                 URL: https://issues.apache.org/jira/browse/KAFKA-8448
             Project: Kafka
          Issue Type: Bug
    Affects Versions: 2.2.0
         Environment: Red Hat 4.4.7-16, java version "1.8.0_152", 
kafka_2.12-2.2.0
            Reporter: Juan Olivares


We have a custom Kafka health check which creates a topic add some ACLs 
(read/write topic and group), produce & consume a single message and then 
quickly remove it and all the related ACLs created.

We have observed that # of instances of {{kafka.log.Log}} keep growing, while 
there's no evidence of topics being leaked, neither running 
{{/opt/kafka/bin/kafka-topics.sh --zookeeper localhost:2181 --describe}} , nor 
looking at the disk directory where topics are stored.

After looking at the heapdump we've observed the following
 - None of the {{kafka.log.Log}} references ({{currentLogs}}, {{logsToBeDeleted 
}} and {{logsToBeDeleted}}) in {{kafka.log.LogManager}} is holding the big 
amount of {{kafka.log.Log}} instances.
 - The only reference preventing {{kafka.log.Log}} to be Garbage collected 
seems to be 
{{java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue}} which 
contains schedule tasks created with the name 
{{PeriodicProducerExpirationCheck}}.

I can see in the code that for every {{kafka.log.Log}} a task with this name is 
scheduled.
{code:java}
  scheduler.schedule(name = "PeriodicProducerExpirationCheck", fun = () => {
    lock synchronized {
      producerStateManager.removeExpiredProducers(time.milliseconds)
    }
  }, period = producerIdExpirationCheckIntervalMs, delay = 
producerIdExpirationCheckIntervalMs, unit = TimeUnit.MILLISECONDS)
{code}

However it seems those tasks are never unscheduled/cancelled



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to