Github user mgaido91 commented on a diff in the pull request:

    https://github.com/apache/spark/pull/21895#discussion_r207499147
  
    --- Diff: 
core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala ---
    @@ -779,6 +808,8 @@ private[history] class FsHistoryProvider(conf: 
SparkConf, clock: Clock)
             listing.delete(classOf[LogInfo], log.logPath)
           }
         }
    +    // Clean the blacklist from the expired entries.
    +    clearBlacklist(CLEAN_INTERVAL_S)
    --- End diff --
    
    This is scheduled anyway every CLEAN_INTERVAL_S. So I don't think that 
changing the value here helps. We may define another config for the 
blacklisting expiration, but this seems an overkill to me. I think it is very 
unlikely that a user changes application permissions on this files and when he 
does, he can always restart the SHS. Or we can also decide to clean the 
blacklist every X amount of time. I don't have a strong opinion on which of 
these options is the best honestly.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to