Github user mgaido91 commented on a diff in the pull request:
https://github.com/apache/spark/pull/21895#discussion_r207499147
--- Diff:
core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala ---
@@ -779,6 +808,8 @@ private[history] class FsHistoryProvider(conf:
SparkConf, clock: Clock)
listing.delete(classOf[LogInfo], log.logPath)
}
}
+ // Clean the blacklist from the expired entries.
+ clearBlacklist(CLEAN_INTERVAL_S)
--- End diff --
This is scheduled anyway every CLEAN_INTERVAL_S. So I don't think that
changing the value here helps. We may define another config for the
blacklisting expiration, but this seems an overkill to me. I think it is very
unlikely that a user changes application permissions on this files and when he
does, he can always restart the SHS. Or we can also decide to clean the
blacklist every X amount of time. I don't have a strong opinion on which of
these options is the best honestly.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]