[jira] [Commented] (SPARK-26395) Spark Thrift server memory leak

2019-02-19 Thread Konstantinos Andrikopoulos (JIRA)


[ 
https://issues.apache.org/jira/browse/SPARK-26395?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16772058#comment-16772058
 ] 

Konstantinos Andrikopoulos commented on SPARK-26395:


After setting the property spark.appStateStore.asyncTracking.enable to false 
that made the situation a bit better in 2 out of our 3 Thrift server instances. 
However according to my understanding after setting this property to false we 
should n't face this issue. 

> Spark Thrift server memory leak
> ---
>
> Key: SPARK-26395
> URL: https://issues.apache.org/jira/browse/SPARK-26395
> Project: Spark
>  Issue Type: Bug
>  Components: Spark Core
>Affects Versions: 2.3.2
>Reporter: Konstantinos Andrikopoulos
>Priority: Major
>
> We are running Thrift Server in standalone mode and we have observed that the 
> heap of the driver is constantly increasing. After analysing the heap dump 
> the issue seems to be that the ElementTrackingStore is constantly increasing 
> due to the addition of RDDOperationGraphWrapper objects that are not cleaned 
> up.
> The ElementTrackingStore defines the addTrigger method were you are able to 
> set thresholds in order to perform cleanup but in practice it is used for  
> ExecutorSummaryWrapper, JobDataWrapper and StageDataWrapper classes by using 
> the following spark properties 
>  * spark.ui.retainedDeadExecutors
>  * spark.ui.retainedJobs
>  * spark.ui.retainedStages
> So the  RDDOperationGraphWrapper which is been added using the onJobStart 
> method of  AppStatusListener class [kvstore.write(uigraph) #line 291]
> in not cleaned up and it constantly increases causing a memory leak



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-26395) Spark Thrift server memory leak

2018-12-18 Thread Konstantinos Andrikopoulos (JIRA)
Konstantinos Andrikopoulos created SPARK-26395:
--

 Summary: Spark Thrift server memory leak
 Key: SPARK-26395
 URL: https://issues.apache.org/jira/browse/SPARK-26395
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 2.3.2
Reporter: Konstantinos Andrikopoulos


We are running Thrift Server in standalone mode and we have observed that the 
heap of the driver is constantly increasing. After analysing the heap dump the 
issue seems to be that the ElementTrackingStore is constantly increasing due to 
the addition of RDDOperationGraphWrapper objects that are not cleaned up.

The ElementTrackingStore defines the addTrigger method were you are able to set 
thresholds in order to perform cleanup but in practice it is used for  
ExecutorSummaryWrapper, JobDataWrapper and StageDataWrapper classes by using 
the following spark properties 
 * spark.ui.retainedDeadExecutors
 * spark.ui.retainedJobs
 * spark.ui.retainedStages

So the  RDDOperationGraphWrapper which is been added using the onJobStart 
method of  AppStatusListener class [kvstore.write(uigraph) #line 291]

in not cleaned up and it constantly increases causing a memory leak



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org