[ 
https://issues.apache.org/jira/browse/SPARK-5841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14568312#comment-14568312
 ] 

Cory Nguyen edited comment on SPARK-5841 at 6/2/15 12:58 AM:
-------------------------------------------------------------

We are using Luigi with Spark 1.3.1 to manage our jobs

However we run into a unique rare case with the following conditions that 
trigger this resolved Block Manger Bug:

- Dataset is relatively large ~ 1.5TB
- Spark job is ran with Luigi
- save to local HDFS

The spark job would process data and mappings just fine, until the very end 
when it proceeds to save the files to local hdfs this is when it triggers this 
bug. 

However, the job saves and complete data successfully if it was saved to s3:// 
location.

wondering what might cause this resolved bug to trigger when ran with luigi 
saving to local hdfs but not trigger when saved to s3 with luigi or ran without 
luigi?


was (Author: cqnguyen):
We are using Luigi with Spark to manage our jobs

However we run into a unique rare case with the following conditions that 
trigger this resolved Block Manger Bug:

- Dataset is relatively large ~ 1.5TB
- Spark job is ran with Luigi
- save to local HDFS

The spark job would process data and mappings just fine, until the very end 
when it proceeds to save the files to local hdfs this is when it triggers this 
bug. 

However, the job saves and complete data successfully if it was saved to s3:// 
location.

wondering what might cause this resolved bug to trigger when ran with luigi 
saving to local hdfs but not trigger when saved to s3 with luigi or ran without 
luigi?

> Memory leak in DiskBlockManager
> -------------------------------
>
>                 Key: SPARK-5841
>                 URL: https://issues.apache.org/jira/browse/SPARK-5841
>             Project: Spark
>          Issue Type: Bug
>          Components: Block Manager
>    Affects Versions: 1.2.1
>            Reporter: Matt Whelan
>            Assignee: Matt Whelan
>             Fix For: 1.3.0
>
>
> DiskBlockManager registers a Runtime shutdown hook, which creates a hard 
> reference to the entire Driver ActorSystem.  If a long-running JVM repeatedly 
> creates and destroys SparkContext instances, it leaks memory.  
> I suggest we deregister the shutdown hook if DiskBlockManager.stop is called. 
>  It's redundant at that point.
> PR coming.
> See also 
> http://mail-archives.apache.org/mod_mbox/spark-user/201501.mbox/%3CCA+kjH+w_DDTEBE9XB6NrPxLTUXD=nc_d-3ogxtumk_5v-e0...@mail.gmail.com%3E



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to