During this month I refactor the code used for the tests and kept doing them with the same base mentioned above (about 92 000 files with an average size of 2kb), but procedure differently: I ran the compression and decompression 50 times in eight different computers. The results were not different from those previously reported, on average gzip was two times faster than the lzo. As a last resort did profiling with JProfiler, but found nothing to explain to me why the gzip be faster than lzo.
In this address http://www.linux.ime.usp.br/~jvcoletto/compression/ I share the table with the results obtained in the tests, the code used in the tests and the results obtained in JProfiler. Anyone have any ideas to help me? Thank you. -- Jose Vinicius Pimenta Coletto Em 2 de março de 2011 16:32, José Vinícius Pimenta Coletto < [email protected]> escreveu: > Hi, > > I'm making a comparison between the following compression methods: gzip > and lzo provided by Hadoop and gzip from package java.util.zip. > The test consists of compression and decompression of approximately 92,000 > files with an average size of 2kb, however the decompression time of lzo is > twice the decompression time of gzip provided by Hadoop, it does not seem > right. > The results obtained in the test are: > > Method | Bytes | Compression > | Decompression > - | - | Total Time(with i/o) Time Speed > | Total Time(with i/o) Time Speed > Gzip (Haddop) | 200876304 | 121.454s 43.167s > 4,653,424.079 B/s | 332.305s 111.806s 1,796,635.326 B/s > Lzo | 200876304 | 120.564s 54.072s > 3,714,914.621 B/s | 509.371s 184.906s 1,086,368.904 B/s > Gzip (java.util.zip) | 200876304 | 148.014s 63.414s > 3,167,647.371 B/s | 483.148s 4.528s 44,360,682.244 B/s > > You can see the code I'm using to the test here: > http://www.linux.ime.usp.br/~jvcoletto/compression/ > > Can anyone explain me why am I getting these results? > Thanks. >
