[
https://issues.apache.org/jira/browse/SPARK-1860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14001151#comment-14001151
]
Andrew Ash commented on SPARK-1860:
-----------------------------------
[~mkim] is going to take a look at this after discussion at
https://issues.apache.org/jira/browse/SPARK-1154
I think the correct fix as Patrick outlines would be:
{code}
// pseudocode
for folder in onDiskFolders:
if folder is owned by a running application:
continue
if folder contains any folder/file (recursively) that is more recently
touched (mtime) than the TTS:
continue
cleanUp(folder)
{code}
Schedule that to run periodically (interval configured by setting) and this
should be all fixed up.
Is that right?
An alternative approach could be to have executor clean up the application's
work directory when the application terminates, but un-clean executor shutdown
could still leave work directories around so a TTL approach still needs to be
included as well.
> Standalone Worker cleanup should not clean up running applications
> ------------------------------------------------------------------
>
> Key: SPARK-1860
> URL: https://issues.apache.org/jira/browse/SPARK-1860
> Project: Spark
> Issue Type: Bug
> Components: Deploy
> Affects Versions: 1.0.0
> Reporter: Aaron Davidson
> Priority: Critical
> Fix For: 1.1.0
>
>
> The default values of the standalone worker cleanup code cleanup all
> application data every 7 days. This includes jars that were added to any
> applications that happen to be running for longer than 7 days, hitting
> streaming jobs especially hard.
> Applications should not be cleaned up if they're still running. Until then,
> this behavior should not be enabled by default.
--
This message was sent by Atlassian JIRA
(v6.2#6252)