Hi, are there any exceptions in the node log that lead to the pod restart?

On Mon, Jan 19, 2026 at 12:22 PM Joel Ferreira (Nokia) via user
<[email protected]> wrote:
>
> Hi,
>
> In our project, we implemented an in-memory cache, based on Apache Ignite 
> 3.1, running on Kubernetes. The Ignite cluster is configured with 3 replicas.
> We are noticing that each of the 3 pods has a memory consumption that 
> progressively increases until it reaches the configured limit, at which point 
> a pod restart occurs.
>
> For your reference, we are using the following configuration:
>
> StatefulSet memory limit: 2Gi
> 3 replicas
> JVM_MAX_MEM = JVM_MIN_MEM: 1Gi
> In-memory table created with:
>
> Zone with high availability
> Profile using "aimem" engine
>
>
> We also created a cronjob that keeps the number of entries in cache within a 
> limit of 6000, but this has no effect on the memory stability of the pods.
> In fact, we even tried clearing all entries from the cache table and did not 
> observe any memory release from the pods. On the contrary, consumption 
> continued to rise slowly.
> Is there any known memory issue that could explain this behavior? Or could 
> there be some setting we missed?
>
> Thanks in advance.
> Kind regards,
> Joel Ferreira

Reply via email to