Github user tgravescs commented on the issue:

    https://github.com/apache/spark/pull/22112
  
    To clarify your last few comments, I think you are saying if you were to 
fail all the reduce tasks, the shuffle write data is still there and doesn't 
get removed and since first write wins on rerun it might still use the older 
already shuffled data?
    
    So in order to fix that we would need a way to tell the executors to remove 
that older committed shuffle data


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to