Hi Meisam,

Each block manager removes data from the cache in a least-recently-used fashion 
as space fills up. If you’d like to remove an RDD manually before that, you can 
call rdd.unpersist().

Matei

On Nov 13, 2013, at 8:15 PM, Meisam Fathi <[email protected]> wrote:

> Hi Community,
> 
> When an RDD in the application becomes unreachable and gets garbage
> collected, how does Spark remove RDD's data from BlockManagers on the
> worker nodes?
> 
> Thanks,
> Meisam

Reply via email to