Okie doke, good to know.
On Wed, Mar 19, 2014 at 7:35 PM, Matei Zaharia <matei.zaha...@gmail.com>wrote: > Yes, Spark automatically removes old RDDs from the cache when you make new > ones. Unpersist forces it to remove them right away. In both cases though, > note that Java doesn’t garbage-collect the objects released until later. > > Matei > > On Mar 19, 2014, at 7:22 PM, Nicholas Chammas <nicholas.cham...@gmail.com> > wrote: > > Related question: > > If I keep creating new RDDs and cache()-ing them, does Spark automatically > unpersist the least recently used RDD when it runs out of memory? Or is an > explicit unpersist the only way to get rid of an RDD (barring the PR > Tathagata mentioned)? > > Also, does unpersist()-ing an RDD immediately free up space, or just allow > that space to be reclaimed when needed? > > > On Wed, Mar 19, 2014 at 7:01 PM, Tathagata Das < > tathagata.das1...@gmail.com> wrote: > >> Just a head's up, there is an active >> <https://github.com/apache/spark/pull/126>*pull requeust* that will >> automatically unpersist RDDs that are not in reference/scope from the >> application any more. >> >> TD >> >> >> On Wed, Mar 19, 2014 at 6:58 PM, hequn cheng <chenghe...@gmail.com>wrote: >> >>> persist and unpersist. >>> unpersist:Mark the RDD as non-persistent, and remove all blocks for it >>> from memory and disk >>> >>> >>> 2014-03-19 16:40 GMT+08:00 林武康 <vboylin1...@gmail.com>: >>> >>> Hi, can any one tell me about the lifecycle of an rdd? I search >>>> through the official website and still can't figure it out. Can I use an >>>> rdd in some stages and destroy it in order to release memory because that >>>> no stages ahead will use this rdd any more. Is it possible? >>>> >>>> Thanks! >>>> >>>> Sincerely >>>> Lin wukang >>>> >>> >>> >> > >