[
https://issues.apache.org/jira/browse/SPARK-26395?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16772058#comment-16772058
]
Konstantinos Andrikopoulos commented on SPARK-26395:
After setting the property spark.appStateStore.asyncTracking.enable to false
that made the situation a bit better in 2 out of our 3 Thrift server instances.
However according to my understanding after setting this property to false we
should n't face this issue.
> Spark Thrift server memory leak
> ---
>
> Key: SPARK-26395
> URL: https://issues.apache.org/jira/browse/SPARK-26395
> Project: Spark
> Issue Type: Bug
> Components: Spark Core
>Affects Versions: 2.3.2
>Reporter: Konstantinos Andrikopoulos
>Priority: Major
>
> We are running Thrift Server in standalone mode and we have observed that the
> heap of the driver is constantly increasing. After analysing the heap dump
> the issue seems to be that the ElementTrackingStore is constantly increasing
> due to the addition of RDDOperationGraphWrapper objects that are not cleaned
> up.
> The ElementTrackingStore defines the addTrigger method were you are able to
> set thresholds in order to perform cleanup but in practice it is used for
> ExecutorSummaryWrapper, JobDataWrapper and StageDataWrapper classes by using
> the following spark properties
> * spark.ui.retainedDeadExecutors
> * spark.ui.retainedJobs
> * spark.ui.retainedStages
> So the RDDOperationGraphWrapper which is been added using the onJobStart
> method of AppStatusListener class [kvstore.write(uigraph) #line 291]
> in not cleaned up and it constantly increases causing a memory leak
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org