[
https://issues.apache.org/jira/browse/SPARK-1042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13970607#comment-13970607
]
OuyangJin commented on SPARK-1042:
----------------------------------
yes is in org.apache.spark.broadcast, "Deleted broadcast file" ,this is a info
log
> spark cleans all java broadcast variables when it hits the spark.cleaner.ttl
> -----------------------------------------------------------------------------
>
> Key: SPARK-1042
> URL: https://issues.apache.org/jira/browse/SPARK-1042
> Project: Spark
> Issue Type: Bug
> Components: Java API, Spark Core
> Affects Versions: 0.8.0, 0.8.1, 0.9.0
> Reporter: Tal Sliwowicz
> Assignee: OuyangJin
> Priority: Critical
> Labels: memory_leak
>
> When setting spark.cleaner.ttl, spark performs the cleanup on time - but it
> cleans all broadcast variables, not just the ones that are older than the
> ttl. This creates an exception when the next mapPartitions runs because it
> cannot find the broadcast variable, even when it was created immediately
> before running the task.
> Our temp workaround - not set the ttl and suffer from an ongoing memory leak
> (forces a restart).
> We are using JavaSparkContext and our broadcast variables are Java HashMaps.
--
This message was sent by Atlassian JIRA
(v6.2#6252)