Github user jinxing64 commented on a diff in the pull request:

    https://github.com/apache/spark/pull/19476#discussion_r144768355
  
    --- Diff: core/src/main/scala/org/apache/spark/storage/BlockManager.scala 
---
    @@ -1552,4 +1582,65 @@ private[spark] object BlockManager {
         override val metricRegistry = new MetricRegistry
         metricRegistry.registerAll(metricSet)
       }
    +
    +  class RemoteBlockTempFileManager(blockManager: BlockManager)
    +      extends TempFileManager with Logging {
    +
    +    private class ReferenceWithCleanup(file: File, referenceQueue: 
JReferenceQueue[File])
    +        extends WeakReference[File](file, referenceQueue) {
    +      private val filePath = file.getAbsolutePath
    +
    +      def cleanUp(): Unit = {
    +        logDebug(s"Clean up file $filePath")
    +
    +        if (!new File(filePath).delete()) {
    +          logDebug(s"Fail to delete file $filePath")
    +        }
    +      }
    +    }
    --- End diff --
    
    Yeah, the thing is if we `registerTempFileToClean` after `cleanup`, the tmp 
file will never be cleaned. In #18565, if `registerTempFileToClean` return 
false, it means the `TempFileManager` has already stopped, i.e. `cleanup` has 
already been finished and will not be called again. We can not rely on 
`TempFileManager` to clean the tmp file.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to