[
https://issues.apache.org/jira/browse/SPARK-1762?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14940395#comment-14940395
]
FangzhouXing commented on SPARK-1762:
-------------------------------------
What is the current eviction policy?
Instead of pinning, what if we just make the eviction policy smarter? (from a
quick look, it seems like the current policy is FIFO)
We want developers to think about how much memory the system has less, not more.
> Add functionality to pin RDDs in cache
> --------------------------------------
>
> Key: SPARK-1762
> URL: https://issues.apache.org/jira/browse/SPARK-1762
> Project: Spark
> Issue Type: Improvement
> Components: Spark Core
> Affects Versions: 1.0.0
> Reporter: Andrew Or
>
> Right now, all RDDs are created equal, and there is no mechanism to identify
> a certain RDD to be more important than the rest. This is a problem if the
> RDD fraction is small, because just caching a few RDDs can evict more
> important ones.
> A side effect of this feature is that we can now more safely allocate a
> smaller spark.storage.memoryFraction if we know how large our important RDDs
> are, without having to worry about them being evicted. This allows us to use
> more memory for shuffles, for instance, and avoid disk spills.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]