[jira] [Commented] (HADOOP-17763) DistCp job fails when AM is killed
[ https://issues.apache.org/jira/browse/HADOOP-17763?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17391644#comment-17391644 ] Srinivasu Majeti commented on HADOOP-17763: --- HI [~BilwaST] , Is there something pending for committing this fix ? > DistCp job fails when AM is killed > -- > > Key: HADOOP-17763 > URL: https://issues.apache.org/jira/browse/HADOOP-17763 > Project: Hadoop Common > Issue Type: Bug >Reporter: Bilwa S T >Assignee: Bilwa S T >Priority: Major > Attachments: HADOOP-17763.001.patch, HADOOP-17763.002.patch > > > Job fails as tasks fail with below exception > {code:java} > 2021-06-11 18:48:47,047 | ERROR | IPC Server handler 0 on 27101 | Task: > attempt_1623387358383_0006_m_00_1000 - exited : > java.io.FileNotFoundException: File does not exist: > hdfs://hacluster/staging-dir/dsperf/.staging/_distcp-646531269/fileList.seq > at > org.apache.hadoop.hdfs.DistributedFileSystem$29.doCall(DistributedFileSystem.java:1637) > at > org.apache.hadoop.hdfs.DistributedFileSystem$29.doCall(DistributedFileSystem.java:1630) > at > org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) > at > org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1645) > at org.apache.hadoop.io.SequenceFile$Reader.(SequenceFile.java:1863) > at org.apache.hadoop.io.SequenceFile$Reader.(SequenceFile.java:1886) > at > org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader.initialize(SequenceFileRecordReader.java:54) > at > org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.initialize(MapTask.java:560) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:798) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:347) > at org.apache.hadoop.mapred.YarnChild$1.run(YarnChild.java:183) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1761) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:177) > | TaskAttemptListenerImpl.java:304{code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HADOOP-15196) Zlib decompression fails when file having trailing garbage
[ https://issues.apache.org/jira/browse/HADOOP-15196?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17380549#comment-17380549 ] Srinivasu Majeti edited comment on HADOOP-15196 at 7/14/21, 12:22 PM: -- Hi [~brahmareddy], Are you going to get this code changes into production ? Or is there another fix for this situation. We might need to add another condition for message before ignoring trailing garbage. We are also getting a diff msg now {code:java} Caused by: java.io.IOException: incorrect header check at org.apache.hadoop.io.compress.zlib.ZlibDecompressor.inflateBytesDirect(Native Method) {code} So we might need a check for below as well {code:java} message.contains("incorrect header check") along with message.contains("unknown compression method") {code} was (Author: smajeti): Hi [~brahmareddy], Are you going to get this code changes into production ? Or is there another fix for this situation. We might need to add another condition for message before ignoring trailing garbage. {code:java} message.contains("incorrect header check") along with message.contains("unknown compression method") {code} > Zlib decompression fails when file having trailing garbage > -- > > Key: HADOOP-15196 > URL: https://issues.apache.org/jira/browse/HADOOP-15196 > Project: Hadoop Common > Issue Type: Bug >Affects Versions: 2.7.0 >Reporter: Brahma Reddy Battula >Assignee: Brahma Reddy Battula >Priority: Major > Attachments: HADOOP-15196.patch > > > *When file has trailing garbage gzip will ignore.* > {noformat} > gzip -d 2018011309-js.rishenglipin.com.gz > gzip: 2018011309-js.rishenglipin.com.gz: decompression OK, trailing garbage > ignored > {noformat} > *when we use same file and decompress,we got following.* > {noformat} > 2018-01-13 14:23:43,151 | WARN | task-result-getter-3 | Lost task 0.0 in > stage 345.0 (TID 5686, node-core-gyVYT, executor 3): java.io.IOException: > unknown compression method > at > org.apache.hadoop.io.compress.zlib.ZlibDecompressor.inflateBytesDirect(Native > Method) > at > org.apache.hadoop.io.compress.zlib.ZlibDecompressor.decompress(ZlibDecompressor.java:225) > at > org.apache.hadoop.io.compress.DecompressorStream.decompress(DecompressorStream.java:91) > at > org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:85) > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15196) Zlib decompression fails when file having trailing garbage
[ https://issues.apache.org/jira/browse/HADOOP-15196?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17380549#comment-17380549 ] Srinivasu Majeti commented on HADOOP-15196: --- Hi [~brahmareddy], Are you going to get this code changes into production ? Or is there another fix for this situation. We might need to add another condition for message before ignoring trailing garbage. {code:java} message.contains("incorrect header check") along with message.contains("unknown compression method") {code} > Zlib decompression fails when file having trailing garbage > -- > > Key: HADOOP-15196 > URL: https://issues.apache.org/jira/browse/HADOOP-15196 > Project: Hadoop Common > Issue Type: Bug >Affects Versions: 2.7.0 >Reporter: Brahma Reddy Battula >Assignee: Brahma Reddy Battula >Priority: Major > Attachments: HADOOP-15196.patch > > > *When file has trailing garbage gzip will ignore.* > {noformat} > gzip -d 2018011309-js.rishenglipin.com.gz > gzip: 2018011309-js.rishenglipin.com.gz: decompression OK, trailing garbage > ignored > {noformat} > *when we use same file and decompress,we got following.* > {noformat} > 2018-01-13 14:23:43,151 | WARN | task-result-getter-3 | Lost task 0.0 in > stage 345.0 (TID 5686, node-core-gyVYT, executor 3): java.io.IOException: > unknown compression method > at > org.apache.hadoop.io.compress.zlib.ZlibDecompressor.inflateBytesDirect(Native > Method) > at > org.apache.hadoop.io.compress.zlib.ZlibDecompressor.decompress(ZlibDecompressor.java:225) > at > org.apache.hadoop.io.compress.DecompressorStream.decompress(DecompressorStream.java:91) > at > org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:85) > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15524) BytesWritable causes OOME when array size reaches Integer.MAX_VALUE
[ https://issues.apache.org/jira/browse/HADOOP-15524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17102329#comment-17102329 ] Srinivasu Majeti commented on HADOOP-15524: --- Hi [~nanda], [~arpaga] , Can we commit this Jira ? Please let me know if there are any concerns on this fix to be worried or safe to commit ? We have got a customer request on this and might need a backport . > BytesWritable causes OOME when array size reaches Integer.MAX_VALUE > --- > > Key: HADOOP-15524 > URL: https://issues.apache.org/jira/browse/HADOOP-15524 > Project: Hadoop Common > Issue Type: Bug > Components: io >Reporter: Joseph Smith >Priority: Major > > BytesWritable.setSize uses Integer.MAX_VALUE to initialize the internal > array. On my environment, this causes an OOME > {code:java} > Exception in thread "main" java.lang.OutOfMemoryError: Requested array size > exceeds VM limit > {code} > byte[Integer.MAX_VALUE-2] must be used to prevent this error. > Tested on OSX and CentOS 7 using Java version 1.8.0_131. > I noticed that java.util.ArrayList contains the following > {code:java} > /** > * The maximum size of array to allocate. > * Some VMs reserve some header words in an array. > * Attempts to allocate larger arrays may result in > * OutOfMemoryError: Requested array size exceeds VM limit > */ > private static final int MAX_ARRAY_SIZE = Integer.MAX_VALUE - 8; > {code} > > BytesWritable.setSize should use something similar to prevent an OOME from > occurring. > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-16756) Inconsistent Behavior on distcp -update over S3
[ https://issues.apache.org/jira/browse/HADOOP-16756?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16995514#comment-16995514 ] Srinivasu Majeti commented on HADOOP-16756: --- Hi [~ste...@apache.org], [~daisuke.kobayashi], Should we need another option like -skipblocklengthcheck for a copy from on-prem to the cloud? Or skip it always if target file system is different from the source ? > Inconsistent Behavior on distcp -update over S3 > --- > > Key: HADOOP-16756 > URL: https://issues.apache.org/jira/browse/HADOOP-16756 > Project: Hadoop Common > Issue Type: Bug > Components: fs/s3, tools/distcp >Affects Versions: 3.3.0 >Reporter: Daisuke Kobayashi >Priority: Major > > Distcp over S3A always copies all source files no matter the files are > changed or not. This is opposite to the statement in the doc below. > [http://hadoop.apache.org/docs/current/hadoop-distcp/DistCp.html] > {noformat} > And to use -update to only copy changed files. > {noformat} > CopyMapper compares file length as well as block size before copying. While > the file length should match, the block size does not. This is apparently > because the returned block size from S3A is always 32MB. > [https://github.com/apache/hadoop/blob/release-3.2.0-RC1/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/mapred/CopyMapper.java#L348] > I'd suppose we should update the documentation or make code change. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org