leixm commented on PR #307:
URL: 
https://github.com/apache/incubator-uniffle/pull/307#issuecomment-1311273243

   > > > We can read the executor configuration of spark through configuration, 
but the actual app running process may not be able to allocate so many 
resources.
   > > 
   > > 
   > > If we use dynamic allocation, we can't know the number of executors. So 
I think we can give a configuration first, user can set that value. Similarly 
ByteDance Shuffle Service give the concurrency tasks through an experience 
formula, you can see 
https://github.com/bytedance/CloudShuffleService/blob/ef0ffb3f43f9f6e96af49629aed2a6ce61a6a2ab/spark-shuffle-manager-2/src/main/scala/org/apache/spark/shuffle/css/CssShuffleManager.scala#L64
   > 
   > Yes. This optimization has been applied in our internal uniffle, it works 
well.
   
   Maybe we can apply this feature to our community and estimate the number of 
ShuffleServers needed according to the number of concurrent tasks.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to