Github user mridulm commented on the pull request:
https://github.com/apache/spark/pull/2247#issuecomment-54267869
Yes, exactly : with higher block size configured sometimes.
Most tasks will not incur (de)compression overhead except for the very few
which spill due to skew.
On 03-Sep-2014 2:09 pm, "Reynold Xin" <[email protected]> wrote:
> I don't get it. Are you saying compression is turned on for spill files,
> but not for general shuffle files?
>
> And in response to your original question about sort leakage: yes this is
> breaking that. See discussion in #2178 (comment)
> <https://github.com/apache/spark/pull/2178#issuecomment-54214760>
>
> â
> Reply to this email directly or view it on GitHub
> <https://github.com/apache/spark/pull/2247#issuecomment-54267261>.
>
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]