holdenk opened a new pull request #28708:
URL: https://github.com/apache/spark/pull/28708


   ### Why are the changes needed?
   
   Recomputting shuffle blocks can be expensive, we should take advantage of 
our decommissioning time to migrate these blocks.
   
   ### Does this PR introduce any user-facing change?
   
   This PR introduces two new configs parameters, 
`spark.storage.decommission.shuffle_blocks` & 
`spark.storage.decommission.rdd_blocks` that control which blocks should be 
migrated during storage decommissioning.
   
   ### How was this patch tested?
   
   New unit test & expansion of the Spark on K8s decom test to assert that 
decommisioning with shuffle block migration means that the results are not 
recomputed even when the original executor is terminated.
   
   This PR is a cleaned-up version of the previous WIP PR I made 
https://github.com/apache/spark/pull/28331 (thanks to @attilapiros for his very 
helpful reviewing on it :)).


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to