Github user HyukjinKwon commented on the pull request:

    https://github.com/apache/spark/pull/11806#issuecomment-198206078
  
    Actually, I did not understand why the overhead of compression at record (I 
mean a row in Spark, a key-value in Hadoop output format) level would be very 
high. I think it is slightly high and the compression ratio is a bit lower but 
allows random accessing at record level.
    
    Maybe I have the lack of knowledge and experience. I would really 
appreciate if you help me.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to