Github user pwendell commented on the pull request:

    https://github.com/apache/spark/pull/3029#issuecomment-63244719
  
    Hey @davies - I played with this a bit and I actually found the behavior 
around concurrent stages might not be great. The reason is that the set of 
active stages will change as stages complete, and then it will suddenly change 
the slider significantly once one stage completes. Here is an example workload:
    
    ```
    > ./bin/spark-shell --conf spark.scheduler.mode=FAIR
    scala> val a = sc.makeRDD(1 to 1000, 10000).map(x => (x, x)).reduceByKey(_ 
+ _)
    scala> val b = sc.makeRDD(1 to 1000, 10000).map(x => (x, x)).reduceByKey(_ 
+ _)
    scala> a.union(b).count()
    ```
    
    Probably what we want in the longer term is to have a slider for the entire 
job rather than stages. But anyways, I'd prefer either the "flip flop" behavior 
or have multiple stacked progress bars. @kayousterhout didn't like the "flip 
flop" but I find it more understandable than what is here now. Since this is an 
opt-in feature I think it's fine to have some version that can go in now and 
then refine it later.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to