Github user pwendell commented on the pull request:

    https://github.com/apache/spark/pull/9051#issuecomment-147493632
  
    I see the underlying problem posed in the JIRA - it's difficult to assess 
duration since it currently includes the time spent waiting on dependent 
stages. However, this patch doesn't seem like the obvious way to fix that. I 
think there are some alternatives that would make more sense:
    
    1. Re-define duration so that it's only defined starting when the first 
task in a stage launches (some concerns here about changing semantics, though).
    2. Add a new field that represents the time spent servicing the stage 
"service time" (?)
    3. Add a new field that represents the time spent queuing before any tasks 
launched "queue time" (?)
    
    Those all seem better ways to address the issue in the JIRA. This way of 
showing the max task time, it seems indirect. And also not always helpful since 
max task time doesn't have a simple relationship with "duration" as desired 
here... for instance the max task could be pretty short but the duration is 
anyways really long for the stage.
    
    /cc @rxin @kayousterhout for any thoughts.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to