[
https://issues.apache.org/jira/browse/SPARK-27709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Dongjoon Hyun updated SPARK-27709:
----------------------------------
Affects Version/s: (was: 3.0.0)
3.1.0
> AppStatusListener.cleanupExecutors should remove dead executors in an
> ordering that makes sense, not a random order
> -------------------------------------------------------------------------------------------------------------------
>
> Key: SPARK-27709
> URL: https://issues.apache.org/jira/browse/SPARK-27709
> Project: Spark
> Issue Type: Improvement
> Components: Web UI
> Affects Versions: 3.1.0
> Reporter: Josh Rosen
> Priority: Minor
>
> When AppStatusListener removes dead executors in excess of
> {{spark.ui.retainedDeadExecutors}}, it looks like it does so in an
> essentially random order:
> Based on the [current
> code|https://github.com/apache/spark/blob/fee695d0cf211e4119c7df7a984708628dc9368a/core/src/main/scala/org/apache/spark/status/AppStatusListener.scala#L1112]
> it looks like we only index based on {{"active"}} but don't perform any
> secondary indexing or sorting based on the age / ID of the executor.
> Instead, I think it might make sense to remove the oldest executors first,
> similar to how we order by "completionTime" when cleaning up old stages.
> I think we should also consider making a higher default of
> {{spark.ui.retainedDeadExecutors}}: it currently defaults to 100 but this
> seems really low in comparison to the total number of retained tasks / stages
> / jobs (which collectively take much more space to store). Maybe ~1000 is a
> safe default?
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]