Frens Jan Rumph created HADOOP-18799: ----------------------------------------
Summary: zstd compressor fails Key: HADOOP-18799 URL: https://issues.apache.org/jira/browse/HADOOP-18799 Project: Hadoop Common Issue Type: Bug Components: native Affects Versions: 3.3.0 Reporter: Frens Jan Rumph It seems like I've hit an issue similar to https://issues.apache.org/jira/browse/HADOOP-15822. I haven't been able to reproduce the issue though. I did manage to add a little bit of logging to org.apache.hadoop.io.compress.zstd.ZStandardCompressor. I've captured the off and len arguments of compress and the srcOffset and srcLen arguments for deflateBytesDirect: {{ compress 0 131591 deflateBytesDirect 0 131591 compress 0 131591 deflateBytesDirect 0 131591 compress 0 131591 deflateBytesDirect 0 131591 compress 0 131591 deflateBytesDirect 0 131591 compress 0 131591 deflateBytesDirect 0 131591 compress 0 131591 deflateBytesDirect 0 131591 compress 0 131591 deflateBytesDirect 0 131591 compress 0 131591 deflateBytesDirect 0 131591 compress 0 131591 deflateBytesDirect 131072 519 }} Just after that last line the process dies with a java.lang.InternalError: Src size is incorrect: {{org.apache.hadoop.mapred.YarnChild: Error running child : java.lang.InternalError: Src size is incorrect at org.apache.hadoop.io.compress.zstd.ZStandardCompressor.deflateBytesDirect(Native Method) at org.apache.hadoop.io.compress.zstd.ZStandardCompressor.compress(ZStandardCompressor.java:220) at org.apache.hadoop.io.compress.CompressorStream.compress(CompressorStream.java:81) at org.apache.hadoop.io.compress.CompressorStream.write(CompressorStream.java:76) at java.base/java.io.BufferedOutputStream.write(BufferedOutputStream.java:123) at java.base/java.io.DataOutputStream.write(DataOutputStream.java:107) at org.apache.hadoop.io.SequenceFile$BlockCompressWriter.writeBuffer(SequenceFile.java:1569) ...}} I have also seen this error: java.lang.InternalError: Error (generic): {{ java.lang.InternalError: Error (generic) at org.apache.hadoop.io.compress.zstd.ZStandardCompressor.deflateBytesDirect(Native Method) at org.apache.hadoop.io.compress.zstd.ZStandardCompressor.compress(ZStandardCompressor.java:220) at org.apache.hadoop.io.compress.CompressorStream.compress(CompressorStream.java:81) at org.apache.hadoop.io.compress.CompressorStream.write(CompressorStream.java:76) at java.base/java.io.BufferedOutputStream.write(BufferedOutputStream.java:123) at java.base/java.io.DataOutputStream.write(DataOutputStream.java:107) at org.apache.hadoop.io.SequenceFile$BlockCompressWriter.writeBuffer(SequenceFile.java:15 ... }} Note that the arguments `131072 519` are _always_ given to `deflateBytesDirect` in case things go wrong. In other cases the offset argument is zero and the size argument is smaller, but not zero; e.g., 0 and 7772. As for some context: we're using the compression as part of writing sequence files with data serialised with Kryo to Backblaze using the S3A file system / S3 client with a map-reduce job on YARN. The job has no issues with smaller values, but for larger ones this situation happens. I've seen very larges values being written successfully, but at some point this error is raised all over the place (after a few larger values). Perhaps some buffer is filling up? Unfortunately, I'm developing using a Mac with M1 processor. So reproducing the issue locally is not a simple feat. If I can somehow produce more leads to investigate this, I'd be happy to. As an aside: we're considering working around this using the hbase-compression-zstd module. This is an alternative compression codec that uses the zstd-jni library without depending on hadoop native. -- This message was sent by Atlassian Jira (v8.20.10#820010) --------------------------------------------------------------------- To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-dev-h...@hadoop.apache.org