Github user tdas commented on a diff in the pull request:

    https://github.com/apache/spark/pull/126#discussion_r10718664
  
    --- Diff: core/src/main/scala/org/apache/spark/rdd/RDD.scala ---
    @@ -1025,6 +1025,14 @@ abstract class RDD[T: ClassTag](
         checkpointData.flatMap(_.getCheckpointFile)
       }
     
    +  def cleanup() {
    --- End diff --
    
    I have not exactly tested this but I dont think it will fail. A cleaned up 
RDD should, for all purposes, same as a RDD that has never been submitted as 
part of a job, so never computed. So it will just be recomputed from scratch.
    
    Nonetheless, I agree, that cleanup should be exposed only with the propers 
semantics and safeguards (throw exception if reused after explicit cleanup). So 
for now we will not expose it. I have removed it and will updated the PR soon.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

Reply via email to