Thanks J-D for your information.
At the HMaster WebGUI (http://host:60030/regionserver.jsp) I saw the same size 
of memstoreSizeMB as well as storefileSizeMB  after flushing.
I will check  out the logs for the size as you proposed.

>> (I'm still using 0.90.4-cdh3u2 because upgrading instructions seems quite 
>> tedious to me)

> Stop everything, deploy new version, restart.


I'm using Cloudera Manager Free 3.7 and on cloudera.com there are some 
instructions that require several step especially concerning uninstallation.
Not that I'am too lazy to do that, however I don't want loose my working test 
system due to a an error on upgrading process.
But someday I will have to do that because we need to evaluate coprocessors for 
our use case.

Sorry for this "let me google
that for you situation" but I just added the that point to the post expecting 
that not every feature is listed on jira..but now I know better :-)
Thanks anyway for the jira links.

regards
Chris


________________________________
Von: Jean-Daniel Cryans <[email protected]>
An: [email protected]; Christian Schäfer <[email protected]> 
Gesendet: 23:53 Freitag, 6.Juli 2012
Betreff: Re: Question about compression

Inline.

J-D

On Fri, Jul 6, 2012 at 3:21 AM, Christian Schäfer <[email protected]> wrote:
> a) Where does compression (like snappy) actually occur.
>
> I set snappy to a column family and filled it with some data (30 MB) -> 
> 640x480 array of 11 Bit values.
>
> After flushing the memstore the size of the data kept exactly the same but 
> flushing was 10x faster than flushing of the table without compression.
>
> So it's "only" the transfer that is compressed? Or are there possibilities to 
> apply compression to the HFiles?

The files are compressed on flush/compact and it's done per 64KB
block. I doubt it the file was the same size as the memstore, look at
your log where it gives the numbers for each flush.

>
> (I'm still using 0.90.4-cdh3u2 because upgrading instructions seems quite 
> tedious to me)

Stop everything, deploy new version,
restart.

>
>
> b) Are there some possibilities to apply delta-compression to HBase to 
> minimize disk usage due to duplicated data?
>
> Has it to be added or even built or is it already included in HBase?

The first hit when googling "hbase delta compression" returns this:
https://issues.apache.org/jira/browse/HBASE-4218

As you can see it was included in 0.94 (no clue how that translates
for CDH... CDH5??)

There is also prefix compression in the pipeline:
https://issues.apache.org/jira/browse/HBASE-4676

Hope this helps,

J-D

Reply via email to