[ https://issues.apache.org/jira/browse/HADOOP-5879?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12713446#action_12713446 ]
He Yongqiang commented on HADOOP-5879: -------------------------------------- test failed test is not related this patch. However, what Chris commented is right. The current patch can not guarantee the Gzip Compressor got from CodecPool is of the settings what users expect. This is kind of global settings, but if we change settings but the CodecPool does not clean its buffered codecs which of old settings. So it may make things work in a wrong way. So one possible way is to let CodecPool do special for Gzip codec, and does either 1) keeps a map for holding gzip codec of different settings. or 2) treats the setting as a global setting, and when the setting is changed, clean all gzip codecs cached in CodecPool. Does the changes for CodecPool sound reasonable/acceptable? > GzipCodec should read compression level etc from configuration > -------------------------------------------------------------- > > Key: HADOOP-5879 > URL: https://issues.apache.org/jira/browse/HADOOP-5879 > Project: Hadoop Core > Issue Type: Improvement > Components: io > Reporter: Zheng Shao > Attachments: hadoop-5879-5-21.patch > > > GzipCodec currently uses the default compression level. We should allow > overriding the default value from Configuration. > {code} > static final class GzipZlibCompressor extends ZlibCompressor { > public GzipZlibCompressor() { > super(ZlibCompressor.CompressionLevel.DEFAULT_COMPRESSION, > ZlibCompressor.CompressionStrategy.DEFAULT_STRATEGY, > ZlibCompressor.CompressionHeader.GZIP_FORMAT, 64*1024); > } > } > {code} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.