Hi,
I'm making a comparison between the following compression methods: gzip
and lzo provided by Hadoop and gzip from package java.util.zip.
The test consists of compression and decompression of approximately 92,000
files with an average size of 2kb, however the decompression time of lzo is
twice the decompression time of gzip provided by Hadoop, it does not seem
right.
The results obtained in the test are:
Method | Bytes | Compression
| Decompression
- | - | Total Time(with i/o) Time Speed
| Total Time(with i/o) Time Speed
Gzip (Haddop) | 200876304 | 121.454s 43.167s
4,653,424.079 B/s | 332.305s 111.806s 1,796,635.326 B/s
Lzo | 200876304 | 120.564s 54.072s
3,714,914.621 B/s | 509.371s 184.906s 1,086,368.904 B/s
Gzip (java.util.zip) | 200876304 | 148.014s 63.414s
3,167,647.371 B/s | 483.148s 4.528s 44,360,682.244 B/s
You can see the code I'm using to the test here:
http://www.linux.ime.usp.br/~jvcoletto/compression/
Can anyone explain me why am I getting these results?
Thanks.