zuston opened a new issue, #2494:
URL: https://github.com/apache/uniffle/issues/2494

   ### Code of Conduct
   
   - [x] I agree to follow this project's [Code of 
Conduct](https://www.apache.org/foundation/policies/conduct)
   
   
   ### Search before asking
   
   - [x] I have searched in the 
[issues](https://github.com/apache/incubator-uniffle/issues?q=is%3Aissue) and 
found no similar issues.
   
   
   ### What would you like to be improved?
   
   From many spark jobs inspection, I found the large part of client shuffle 
write is the compression time. Please see the following log.
   
   `2025-06-04 16:45:45,934 INFO writer.RssShuffleWriter: Finish write shuffle 
for appId[application_1710209318993_131126136_1749024542724], shuffleId[302], 
taskId[28290_0] with write 25645 ms, include checkSendResult[68], commit[0], 
WriteBufferManager cost copyTime[1725], writeTime[25187], serializeTime[950], 
sortTime[0], estimateTime[0], requireMemoryTime[135], 
uncompressedDataLen[8176523436], compressedDataLen[2839134582], 
compressTime[18677], compressRatio[2.8799353]`
   
   From the uniffle code base, maybe the cost of compression is the buffer 
memory allocation. 
   
   
   ### How should we improve?
   
   _No response_
   
   ### Are you willing to submit PR?
   
   - [x] Yes I am willing to submit a PR!


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to