Github user HyukjinKwon commented on the pull request:
https://github.com/apache/spark/pull/11806#issuecomment-198206078
Actually, I did not understand why the overhead of compression at record (I
mean a row in Spark, a key-value in Hadoop output format) level would be very
high. I think it is slightly high and the compression ratio is a bit lower but
allows random accessing at record level.
Maybe I have the lack of knowledge and experience. I would really
appreciate if you help me.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]