Hi Meisam, Each block manager removes data from the cache in a least-recently-used fashion as space fills up. If you’d like to remove an RDD manually before that, you can call rdd.unpersist().
Matei On Nov 13, 2013, at 8:15 PM, Meisam Fathi <[email protected]> wrote: > Hi Community, > > When an RDD in the application becomes unreachable and gets garbage > collected, how does Spark remove RDD's data from BlockManagers on the > worker nodes? > > Thanks, > Meisam
