[
https://issues.apache.org/jira/browse/HADOOP-17868?focusedWorklogId=644531&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-644531
]
ASF GitHub Bot logged work on HADOOP-17868:
-------------------------------------------
Author: ASF GitHub Bot
Created on: 31/Aug/21 16:58
Start Date: 31/Aug/21 16:58
Worklog Time Spent: 10m
Work Description: sunchao commented on a change in pull request #3336:
URL: https://github.com/apache/hadoop/pull/3336#discussion_r699512553
##########
File path:
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/TestCodec.java
##########
@@ -725,6 +726,172 @@ public void testGzipCompatibility() throws IOException {
assertArrayEquals(b, dflchk);
}
+ @Test
+ public void testGzipCompatibilityWithCompressor() throws IOException {
+ // don't use native libs
+ ZlibFactory.setNativeZlibLoaded(false);
+ Configuration conf = new Configuration();
+ CompressionCodec codec = ReflectionUtils.newInstance(GzipCodec.class,
conf);
+
+ for (int i = 0; i < 100; i++){
+ Compressor compressor = codec.createCompressor();
+ assertThat(compressor).withFailMessage("should be BuiltInGzipCompressor")
+ .isInstanceOf(BuiltInGzipCompressor.class);
+
+ Random r = new Random();
+ long seed = r.nextLong();
+ r.setSeed(seed);
+ LOG.info("seed: {}", seed);
+
+ int inputSize = r.nextInt(256 * 1024 + 1);
+ byte[] b = new byte[inputSize];
+ r.nextBytes(b);
+
+ compressor.setInput(b, 0, b.length);
+ compressor.finish();
+
+ byte[] output = new byte[inputSize + 1024];
+ int outputOff = 0;
+
+ while (!compressor.finished()) {
+ byte[] buf = new byte[r.nextInt(1024)];
+ int compressed = compressor.compress(buf, 0, buf.length);
+ System.arraycopy(buf, 0, output, outputOff, compressed);
+ outputOff += compressed;
+ }
+
+ DataInputBuffer gzbuf = new DataInputBuffer();
+ gzbuf.reset(output, outputOff);
+
+ Decompressor decom = codec.createDecompressor();
+ assertThat(decom).as("decompressor should not be null").isNotNull();
+ assertThat(decom).withFailMessage("should be BuiltInGzipDecompressor")
+ .isInstanceOf(BuiltInGzipDecompressor.class);
+ try (InputStream gzin = codec.createInputStream(gzbuf, decom);
+ DataOutputBuffer dflbuf = new DataOutputBuffer()) {
+ dflbuf.reset();
+ IOUtils.copyBytes(gzin, dflbuf, 4096);
+ final byte[] dflchk = Arrays.copyOf(dflbuf.getData(),
dflbuf.getLength());
+ assertThat(b).as("check decompressed output").isEqualTo(dflchk);
+ }
+ }
+ }
+
+ @Test
+ public void testGzipCompatibilityWithCompressorAndGZIPOutputStream() throws
IOException {
+ // don't use native libs
+ ZlibFactory.setNativeZlibLoaded(false);
+ Configuration conf = new Configuration();
+ CompressionCodec codec = ReflectionUtils.newInstance(GzipCodec.class,
conf);
+
+ for (int i = 0; i < 100; i++){
+ Compressor compressor = codec.createCompressor();
+ assertThat(compressor).withFailMessage("should be BuiltInGzipCompressor")
+ .isInstanceOf(BuiltInGzipCompressor.class);
+
+ Random r = new Random();
Review comment:
nit: we can move this out of the loop
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
Issue Time Tracking
-------------------
Worklog Id: (was: 644531)
Time Spent: 2h 50m (was: 2h 40m)
> Add more test for the BuiltInGzipCompressor
> -------------------------------------------
>
> Key: HADOOP-17868
> URL: https://issues.apache.org/jira/browse/HADOOP-17868
> Project: Hadoop Common
> Issue Type: Test
> Reporter: L. C. Hsieh
> Priority: Major
> Labels: pull-request-available
> Time Spent: 2h 50m
> Remaining Estimate: 0h
>
> We added BuiltInGzipCompressor recently. It is better to add more
> compatibility tests for the compressor.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]