[ 
https://issues.apache.org/jira/browse/HBASE-27049?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17538566#comment-17538566
 ] 

Andrew Kyle Purtell commented on HBASE-27049:
---------------------------------------------

We have to care about file compatibility and so if making the change here any 
changes to the binary format of the file must be analyzed and addressed. 
Unfortunately it will not enough just to do a configuration switch if there is 
a change to the binary format of the compression because our users may do 
rolling upgrades, or rollback, or offline hfile writes and bulk loads. 

So making the change you propose is fine, but there must also be a plan for 
evolution of the hfile format (and WAL format too, for the value compression 
option) if the binary format of the compression will change. Read the impl code 
of the Hadoop block compression streams and you will see what I mean. 

> Decrease memory copy when decompress data
> -----------------------------------------
>
>                 Key: HBASE-27049
>                 URL: https://issues.apache.org/jira/browse/HBASE-27049
>             Project: HBase
>          Issue Type: Improvement
>          Components: regionserver
>            Reporter: chenfengge
>            Priority: Minor
>
> HBase RegionServer use createDecompressionStream in class 
> org.apache.hadoop.hbase.io.compress.Compression, which cause extra memory 
> copy during decompression.We can offer interface for block decompression, 
> like "void decompress(ByteBuff src, ByteBuff dst);", and offer default 
> implementation for all algorithms.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

Reply via email to