Dandandan edited a comment on issue #27:
URL: https://github.com/apache/arrow-datafusion/issues/27#issuecomment-839140660


   I also realize I may have implemented this in an inefficient way. Looking at 
Spark it does first a partial aggregate followed by a hash repartition followed 
by another aggregation. I think this will not slow down the low cardinality 
query that much (as the result set will be small already after first aggregate) 
while being fast - maybe even faster - for high cardinality aggregates.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to