Github user markhamstra commented on a diff in the pull request:

    https://github.com/apache/spark/pull/14737#discussion_r75602737
  
    --- Diff: core/src/main/scala/org/apache/spark/ui/SparkUI.scala ---
    @@ -141,6 +141,7 @@ private[spark] object SparkUI {
       val DEFAULT_POOL_NAME = "default"
       val DEFAULT_RETAINED_STAGES = 1000
       val DEFAULT_RETAINED_JOBS = 1000
    +  val DEFAULT_RETAINED_NODES = 2
    --- End diff --
    
    `NODES`, both here and in `spark.ui.retainedNodes` if far too ambiguous and 
non-specific for this configuration value -- "node" is already overloaded too 
many times in the existing Spark code and documentation; we don't need or want 
to add another overload.
    
    Additionally, the default behavior should be the same as current behavior, 
since the change in behavior would be unexpected and it is far from clear to me 
that the overwhelming majority of users would prefer the proposed new behavior. 


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to