[ 
https://issues.apache.org/jira/browse/SPARK-26395?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16772058#comment-16772058
 ] 

Konstantinos Andrikopoulos commented on SPARK-26395:
----------------------------------------------------

After setting the property spark.appStateStore.asyncTracking.enable to false 
that made the situation a bit better in 2 out of our 3 Thrift server instances. 
However according to my understanding after setting this property to false we 
should n't face this issue. 

> Spark Thrift server memory leak
> -------------------------------
>
>                 Key: SPARK-26395
>                 URL: https://issues.apache.org/jira/browse/SPARK-26395
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>    Affects Versions: 2.3.2
>            Reporter: Konstantinos Andrikopoulos
>            Priority: Major
>
> We are running Thrift Server in standalone mode and we have observed that the 
> heap of the driver is constantly increasing. After analysing the heap dump 
> the issue seems to be that the ElementTrackingStore is constantly increasing 
> due to the addition of RDDOperationGraphWrapper objects that are not cleaned 
> up.
> The ElementTrackingStore defines the addTrigger method were you are able to 
> set thresholds in order to perform cleanup but in practice it is used for  
> ExecutorSummaryWrapper, JobDataWrapper and StageDataWrapper classes by using 
> the following spark properties 
>  * spark.ui.retainedDeadExecutors
>  * spark.ui.retainedJobs
>  * spark.ui.retainedStages
> So the  RDDOperationGraphWrapper which is been added using the onJobStart 
> method of  AppStatusListener class [kvstore.write(uigraph) #line 291]
> in not cleaned up and it constantly increases causing a memory leak



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to