Github user mridulm commented on the pull request:

    https://github.com/apache/spark/pull/981#issuecomment-45318235
  
    What is the impact on memory footprint for this change ?
    We have jobs which result in 40k stages, and we have jobs which have stages 
with 50k tasks.
    I would assume the former would be impacted : but by how much ?
    
    And if the impact is non trivial (the master is already at limits of its 
memory), how can the user go about reducing the memory footprint of this change 
? Possibly disabling it ? Reducing amount of info stored ? Something 
better/different ?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

Reply via email to