[ 
https://issues.apache.org/jira/browse/FLINK-33239?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser updated FLINK-33239:
-----------------------------------
    Priority: Major  (was: Blocker)

> After enabling exactly-once in the Flink Kafka sink, the Kafka broker's 
> memory keeps increasing, eventually causing the Kafka broker to crash.
> ----------------------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: FLINK-33239
>                 URL: https://issues.apache.org/jira/browse/FLINK-33239
>             Project: Flink
>          Issue Type: Bug
>          Components: Connectors / Kafka
>    Affects Versions: 1.17.1
>         Environment: flink 1.17.1
> kafka server 2.8.2
>            Reporter: Darcy Lin
>            Priority: Major
>         Attachments: image-2023-10-11-18-47-32-712.png
>
>
> We are using Flink version 1.17.1 and Kafka server version 2.8.2. After 
> enabling exactly-once, in order to allow downstream consumers to read data 
> from Kafka as soon as possible, we set the checkpoint interval to 5 seconds. 
> Approximately three days after writing to the Kafka cluster, the Kafka JVM's 
> memory is exhausted. We printed the memory consumption and found that the 
> main consumption is on the {{kafka.log.ProducerStateEntry}} object.
> Currently, in the exactly-once Kafka sink, a new producer is created every 
> time a checkpoint is executed. The {{kafka.log.ProducerStateEntry}} object 
> seems to store the producer's state, so it keeps increasing. We'd like to 
> ask: Is this normal? If it's normal, do we need to allocate a large amount of 
> memory for our Kafka cluster? If it's not normal, how should we solve this 
> problem?
> !image-2023-10-11-18-47-32-712.png!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to