cloud-fan commented on issue #27096: [SPARK-28148][SQL] Repartition after join 
is not optimized away
URL: https://github.com/apache/spark/pull/27096#issuecomment-574507348
 
 
   In general, shuffles are added by Spark and Spark can pick the best # of 
partitions. However, for user-specified shuffles (df.repartition), the semantic 
should be respecting the user request and keep the # of partitions.

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to