[
https://issues.apache.org/jira/browse/SPARK-55664?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=18060870#comment-18060870
]
Devin Petersohn commented on SPARK-55664:
-----------------------------------------
cc [~holden]
> Explore bounded collection types for pandas aggregations
> --------------------------------------------------------
>
> Key: SPARK-55664
> URL: https://issues.apache.org/jira/browse/SPARK-55664
> Project: Spark
> Issue Type: Bug
> Components: Pandas API on Spark, PySpark
> Affects Versions: 4.1.1
> Reporter: Devin Petersohn
> Priority: Major
>
> From [https://github.com/apache/spark/pull/54370#discussion_r2824804015]
> Right now, pandas API for Spark aggregations (like describe) collect full
> intermediate results back to the driver.
> The idea is to use bounded/fixed-size summary structures during aggregations
> on the executors. Instead of collecting raw data, each executor would produce
> a compact summary (e.g., counts, min/max, approximate quantiles) that has a
> known upper bound on size. These summaries would then be merged (similar to a
> treeReduce pattern), so the final result sent to the driver is smaller.
> This is a more general optimization that could apply beyond describe to other
> Spark aggregations.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]