Hi,

I'm making a comparison between the following compression methods: gzip
and lzo provided by Hadoop and gzip from package java.util.zip.
The test consists of compression and decompression of approximately 92,000
files with an average size of 2kb, however the decompression time of lzo is
twice the decompression time of gzip provided by Hadoop, it does not seem
right.
The results obtained in the test are:

      Method         |   Bytes   |               Compression
       |                    Decompression
         -           |     -     | Total Time(with i/o)  Time     Speed
        | Total Time(with i/o)  Time      Speed
Gzip (Haddop)        | 200876304 | 121.454s              43.167s
 4,653,424.079 B/s | 332.305s              111.806s   1,796,635.326 B/s
Lzo                  | 200876304 | 120.564s              54.072s
 3,714,914.621 B/s | 509.371s              184.906s   1,086,368.904 B/s
Gzip (java.util.zip) | 200876304 | 148.014s              63.414s
 3,167,647.371 B/s | 483.148s              4.528s    44,360,682.244 B/s

You can see the code I'm using to the test here:
http://www.linux.ime.usp.br/~jvcoletto/compression/

Can anyone explain me why am I getting these results?
Thanks.

Reply via email to