[
https://issues.apache.org/jira/browse/SPARK-1042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13970587#comment-13970587
]
Tal Sliwowicz commented on SPARK-1042:
--------------------------------------
1. Yes, I'm sure it creates and uses a news instance
2. Just 1 round. I can increase the ttl, but it it still happens.
3. I turned logs to WARN. Can you be more specific about the classes I should
increase logging on? I will move them to info or even debug
> spark cleans all java broadcast variables when it hits the spark.cleaner.ttl
> -----------------------------------------------------------------------------
>
> Key: SPARK-1042
> URL: https://issues.apache.org/jira/browse/SPARK-1042
> Project: Spark
> Issue Type: Bug
> Components: Java API, Spark Core
> Affects Versions: 0.8.0, 0.8.1, 0.9.0
> Reporter: Tal Sliwowicz
> Assignee: OuyangJin
> Priority: Critical
> Labels: memory_leak
>
> When setting spark.cleaner.ttl, spark performs the cleanup on time - but it
> cleans all broadcast variables, not just the ones that are older than the
> ttl. This creates an exception when the next mapPartitions runs because it
> cannot find the broadcast variable, even when it was created immediately
> before running the task.
> Our temp workaround - not set the ttl and suffer from an ongoing memory leak
> (forces a restart).
> We are using JavaSparkContext and our broadcast variables are Java HashMaps.
--
This message was sent by Atlassian JIRA
(v6.2#6252)