[jira] [Updated] (HDDS-1121) Key read failure when data is written parallel in to Ozone
[ https://issues.apache.org/jira/browse/HDDS-1121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anu Engineer updated HDDS-1121: --- Resolution: Fixed Fix Version/s: 0.4.0 Status: Resolved (was: Patch Available) [~linyiqun], [~shashikant] Thanks for the reviews. [~bharatviswa] Thanks for fixiing this issue. [~gopalv] thanks for reporting this issue. I have committed this to the trunk after verifying, checkstyle, findbugs etc. on the local machine. ASF infra seems to have some issues, since it has been 2 days since the patch is posted and jenkins have not picked up. I have verified these things manually. > Key read failure when data is written parallel in to Ozone > -- > > Key: HDDS-1121 > URL: https://issues.apache.org/jira/browse/HDDS-1121 > Project: Hadoop Distributed Data Store > Issue Type: Bug >Reporter: Bharat Viswanadham >Assignee: Bharat Viswanadham >Priority: Major > Fix For: 0.4.0 > > Attachments: HDDS-1121.00.patch, HDDS-1121.01.patch, > HDDS-1121.02.patch > > > When hive is run with multiple threads for data ingestion to ozone. After > ingestion is done, during read we see this below error. > This issue is found during hive testing, and found by [~t3rmin4t0r] > {code:java} > caused by: org.apache.hadoop.ozone.common.OzoneChecksumException: Checksum > mismatch at index 0 > at > org.apache.hadoop.ozone.common.ChecksumData.verifyChecksumDataMatches(ChecksumData.java:143) > at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:239) > at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:217) > at > org.apache.hadoop.hdds.scm.storage.BlockInputStream.readChunkFromContainer(BlockInputStream.java:227) > at > org.apache.hadoop.hdds.scm.storage.BlockInputStream.seek(BlockInputStream.java:259) > at > org.apache.hadoop.ozone.client.io.KeyInputStream$ChunkInputStreamEntry.seek(KeyInputStream.java:249) > at > org.apache.hadoop.ozone.client.io.KeyInputStream.seek(KeyInputStream.java:180) > at > org.apache.hadoop.fs.ozone.OzoneFSInputStream.seek(OzoneFSInputStream.java:62) > at org.apache.hadoop.fs.FSInputStream.read(FSInputStream.java:82) > at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121) > at > org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:111) > at org.apache.orc.impl.ReaderImpl.extractFileTail(ReaderImpl.java:555) > at org.apache.orc.impl.ReaderImpl.(ReaderImpl.java:370) > at org.apache.hadoop.hive.ql.io.orc.ReaderImpl.(ReaderImpl.java:61) > at org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader(OrcFile.java:105) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails(OrcInputFormat.java:1647) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.callInternal(OrcInputFormat.java:1533) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.access$2700(OrcInputFormat.java:1329) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1513) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1510) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1688) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1510) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1329) > at java.util.concurrent.FutureTask.run(FutureTask.java:266){code} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDDS-1121) Key read failure when data is written parallel in to Ozone
[ https://issues.apache.org/jira/browse/HDDS-1121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bharat Viswanadham updated HDDS-1121: - Attachment: HDDS-1121.02.patch > Key read failure when data is written parallel in to Ozone > -- > > Key: HDDS-1121 > URL: https://issues.apache.org/jira/browse/HDDS-1121 > Project: Hadoop Distributed Data Store > Issue Type: Bug >Reporter: Bharat Viswanadham >Assignee: Bharat Viswanadham >Priority: Major > Attachments: HDDS-1121.00.patch, HDDS-1121.01.patch, > HDDS-1121.02.patch > > > When hive is run with multiple threads for data ingestion to ozone. After > ingestion is done, during read we see this below error. > This issue is found during hive testing, and found by [~t3rmin4t0r] > {code:java} > caused by: org.apache.hadoop.ozone.common.OzoneChecksumException: Checksum > mismatch at index 0 > at > org.apache.hadoop.ozone.common.ChecksumData.verifyChecksumDataMatches(ChecksumData.java:143) > at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:239) > at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:217) > at > org.apache.hadoop.hdds.scm.storage.BlockInputStream.readChunkFromContainer(BlockInputStream.java:227) > at > org.apache.hadoop.hdds.scm.storage.BlockInputStream.seek(BlockInputStream.java:259) > at > org.apache.hadoop.ozone.client.io.KeyInputStream$ChunkInputStreamEntry.seek(KeyInputStream.java:249) > at > org.apache.hadoop.ozone.client.io.KeyInputStream.seek(KeyInputStream.java:180) > at > org.apache.hadoop.fs.ozone.OzoneFSInputStream.seek(OzoneFSInputStream.java:62) > at org.apache.hadoop.fs.FSInputStream.read(FSInputStream.java:82) > at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121) > at > org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:111) > at org.apache.orc.impl.ReaderImpl.extractFileTail(ReaderImpl.java:555) > at org.apache.orc.impl.ReaderImpl.(ReaderImpl.java:370) > at org.apache.hadoop.hive.ql.io.orc.ReaderImpl.(ReaderImpl.java:61) > at org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader(OrcFile.java:105) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails(OrcInputFormat.java:1647) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.callInternal(OrcInputFormat.java:1533) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.access$2700(OrcInputFormat.java:1329) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1513) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1510) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1688) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1510) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1329) > at java.util.concurrent.FutureTask.run(FutureTask.java:266){code} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDDS-1121) Key read failure when data is written parallel in to Ozone
[ https://issues.apache.org/jira/browse/HDDS-1121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bharat Viswanadham updated HDDS-1121: - Attachment: HDDS-1121.01.patch > Key read failure when data is written parallel in to Ozone > -- > > Key: HDDS-1121 > URL: https://issues.apache.org/jira/browse/HDDS-1121 > Project: Hadoop Distributed Data Store > Issue Type: Bug >Reporter: Bharat Viswanadham >Assignee: Bharat Viswanadham >Priority: Major > Attachments: HDDS-1121.00.patch, HDDS-1121.01.patch > > > When hive is run with multiple threads for data ingestion to ozone. After > ingestion is done, during read we see this below error. > This issue is found during hive testing, and found by [~t3rmin4t0r] > {code:java} > caused by: org.apache.hadoop.ozone.common.OzoneChecksumException: Checksum > mismatch at index 0 > at > org.apache.hadoop.ozone.common.ChecksumData.verifyChecksumDataMatches(ChecksumData.java:143) > at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:239) > at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:217) > at > org.apache.hadoop.hdds.scm.storage.BlockInputStream.readChunkFromContainer(BlockInputStream.java:227) > at > org.apache.hadoop.hdds.scm.storage.BlockInputStream.seek(BlockInputStream.java:259) > at > org.apache.hadoop.ozone.client.io.KeyInputStream$ChunkInputStreamEntry.seek(KeyInputStream.java:249) > at > org.apache.hadoop.ozone.client.io.KeyInputStream.seek(KeyInputStream.java:180) > at > org.apache.hadoop.fs.ozone.OzoneFSInputStream.seek(OzoneFSInputStream.java:62) > at org.apache.hadoop.fs.FSInputStream.read(FSInputStream.java:82) > at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121) > at > org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:111) > at org.apache.orc.impl.ReaderImpl.extractFileTail(ReaderImpl.java:555) > at org.apache.orc.impl.ReaderImpl.(ReaderImpl.java:370) > at org.apache.hadoop.hive.ql.io.orc.ReaderImpl.(ReaderImpl.java:61) > at org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader(OrcFile.java:105) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails(OrcInputFormat.java:1647) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.callInternal(OrcInputFormat.java:1533) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.access$2700(OrcInputFormat.java:1329) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1513) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1510) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1688) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1510) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1329) > at java.util.concurrent.FutureTask.run(FutureTask.java:266){code} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDDS-1121) Key read failure when data is written parallel in to Ozone
[ https://issues.apache.org/jira/browse/HDDS-1121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bharat Viswanadham updated HDDS-1121: - Status: Patch Available (was: In Progress) > Key read failure when data is written parallel in to Ozone > -- > > Key: HDDS-1121 > URL: https://issues.apache.org/jira/browse/HDDS-1121 > Project: Hadoop Distributed Data Store > Issue Type: Bug >Reporter: Bharat Viswanadham >Assignee: Bharat Viswanadham >Priority: Major > Attachments: HDDS-1121.00.patch > > > When hive is run with multiple threads for data ingestion to ozone. After > ingestion is done, during read we see this below error. > This issue is found during hive testing. > {code:java} > caused by: org.apache.hadoop.ozone.common.OzoneChecksumException: Checksum > mismatch at index 0 > at > org.apache.hadoop.ozone.common.ChecksumData.verifyChecksumDataMatches(ChecksumData.java:143) > at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:239) > at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:217) > at > org.apache.hadoop.hdds.scm.storage.BlockInputStream.readChunkFromContainer(BlockInputStream.java:227) > at > org.apache.hadoop.hdds.scm.storage.BlockInputStream.seek(BlockInputStream.java:259) > at > org.apache.hadoop.ozone.client.io.KeyInputStream$ChunkInputStreamEntry.seek(KeyInputStream.java:249) > at > org.apache.hadoop.ozone.client.io.KeyInputStream.seek(KeyInputStream.java:180) > at > org.apache.hadoop.fs.ozone.OzoneFSInputStream.seek(OzoneFSInputStream.java:62) > at org.apache.hadoop.fs.FSInputStream.read(FSInputStream.java:82) > at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121) > at > org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:111) > at org.apache.orc.impl.ReaderImpl.extractFileTail(ReaderImpl.java:555) > at org.apache.orc.impl.ReaderImpl.(ReaderImpl.java:370) > at org.apache.hadoop.hive.ql.io.orc.ReaderImpl.(ReaderImpl.java:61) > at org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader(OrcFile.java:105) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails(OrcInputFormat.java:1647) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.callInternal(OrcInputFormat.java:1533) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.access$2700(OrcInputFormat.java:1329) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1513) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1510) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1688) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1510) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1329) > at java.util.concurrent.FutureTask.run(FutureTask.java:266){code} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDDS-1121) Key read failure when data is written parallel in to Ozone
[ https://issues.apache.org/jira/browse/HDDS-1121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bharat Viswanadham updated HDDS-1121: - Description: When hive is run with multiple threads for data ingestion to ozone. After ingestion is done, during read we see this below error. This issue is found during hive testing, and found by [~t3rmin4t0r] {code:java} caused by: org.apache.hadoop.ozone.common.OzoneChecksumException: Checksum mismatch at index 0 at org.apache.hadoop.ozone.common.ChecksumData.verifyChecksumDataMatches(ChecksumData.java:143) at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:239) at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:217) at org.apache.hadoop.hdds.scm.storage.BlockInputStream.readChunkFromContainer(BlockInputStream.java:227) at org.apache.hadoop.hdds.scm.storage.BlockInputStream.seek(BlockInputStream.java:259) at org.apache.hadoop.ozone.client.io.KeyInputStream$ChunkInputStreamEntry.seek(KeyInputStream.java:249) at org.apache.hadoop.ozone.client.io.KeyInputStream.seek(KeyInputStream.java:180) at org.apache.hadoop.fs.ozone.OzoneFSInputStream.seek(OzoneFSInputStream.java:62) at org.apache.hadoop.fs.FSInputStream.read(FSInputStream.java:82) at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121) at org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:111) at org.apache.orc.impl.ReaderImpl.extractFileTail(ReaderImpl.java:555) at org.apache.orc.impl.ReaderImpl.(ReaderImpl.java:370) at org.apache.hadoop.hive.ql.io.orc.ReaderImpl.(ReaderImpl.java:61) at org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader(OrcFile.java:105) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails(OrcInputFormat.java:1647) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.callInternal(OrcInputFormat.java:1533) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.access$2700(OrcInputFormat.java:1329) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1513) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1510) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1688) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1510) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1329) at java.util.concurrent.FutureTask.run(FutureTask.java:266){code} was: When hive is run with multiple threads for data ingestion to ozone. After ingestion is done, during read we see this below error. This issue is found during hive testing. {code:java} caused by: org.apache.hadoop.ozone.common.OzoneChecksumException: Checksum mismatch at index 0 at org.apache.hadoop.ozone.common.ChecksumData.verifyChecksumDataMatches(ChecksumData.java:143) at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:239) at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:217) at org.apache.hadoop.hdds.scm.storage.BlockInputStream.readChunkFromContainer(BlockInputStream.java:227) at org.apache.hadoop.hdds.scm.storage.BlockInputStream.seek(BlockInputStream.java:259) at org.apache.hadoop.ozone.client.io.KeyInputStream$ChunkInputStreamEntry.seek(KeyInputStream.java:249) at org.apache.hadoop.ozone.client.io.KeyInputStream.seek(KeyInputStream.java:180) at org.apache.hadoop.fs.ozone.OzoneFSInputStream.seek(OzoneFSInputStream.java:62) at org.apache.hadoop.fs.FSInputStream.read(FSInputStream.java:82) at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121) at org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:111) at org.apache.orc.impl.ReaderImpl.extractFileTail(ReaderImpl.java:555) at org.apache.orc.impl.ReaderImpl.(ReaderImpl.java:370) at org.apache.hadoop.hive.ql.io.orc.ReaderImpl.(ReaderImpl.java:61) at org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader(OrcFile.java:105) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails(OrcInputFormat.java:1647) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.callInternal(OrcInputFormat.java:1533) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.access$2700(OrcInputFormat.java:1329) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1513) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1510) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at
[jira] [Updated] (HDDS-1121) Key read failure when data is written parallel in to Ozone
[ https://issues.apache.org/jira/browse/HDDS-1121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bharat Viswanadham updated HDDS-1121: - Attachment: HDDS-1121.00.patch > Key read failure when data is written parallel in to Ozone > -- > > Key: HDDS-1121 > URL: https://issues.apache.org/jira/browse/HDDS-1121 > Project: Hadoop Distributed Data Store > Issue Type: Bug >Reporter: Bharat Viswanadham >Assignee: Bharat Viswanadham >Priority: Major > Attachments: HDDS-1121.00.patch > > > When hive is run with multiple threads for data ingestion to ozone. After > ingestion is done, during read we see this below error. > This issue is found during hive testing. > {code:java} > caused by: org.apache.hadoop.ozone.common.OzoneChecksumException: Checksum > mismatch at index 0 > at > org.apache.hadoop.ozone.common.ChecksumData.verifyChecksumDataMatches(ChecksumData.java:143) > at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:239) > at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:217) > at > org.apache.hadoop.hdds.scm.storage.BlockInputStream.readChunkFromContainer(BlockInputStream.java:227) > at > org.apache.hadoop.hdds.scm.storage.BlockInputStream.seek(BlockInputStream.java:259) > at > org.apache.hadoop.ozone.client.io.KeyInputStream$ChunkInputStreamEntry.seek(KeyInputStream.java:249) > at > org.apache.hadoop.ozone.client.io.KeyInputStream.seek(KeyInputStream.java:180) > at > org.apache.hadoop.fs.ozone.OzoneFSInputStream.seek(OzoneFSInputStream.java:62) > at org.apache.hadoop.fs.FSInputStream.read(FSInputStream.java:82) > at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121) > at > org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:111) > at org.apache.orc.impl.ReaderImpl.extractFileTail(ReaderImpl.java:555) > at org.apache.orc.impl.ReaderImpl.(ReaderImpl.java:370) > at org.apache.hadoop.hive.ql.io.orc.ReaderImpl.(ReaderImpl.java:61) > at org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader(OrcFile.java:105) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails(OrcInputFormat.java:1647) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.callInternal(OrcInputFormat.java:1533) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.access$2700(OrcInputFormat.java:1329) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1513) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1510) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1688) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1510) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1329) > at java.util.concurrent.FutureTask.run(FutureTask.java:266){code} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDDS-1121) Key read failure when data is written parallel in to Ozone
[ https://issues.apache.org/jira/browse/HDDS-1121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bharat Viswanadham updated HDDS-1121: - Target Version/s: 0.4.0 > Key read failure when data is written parallel in to Ozone > -- > > Key: HDDS-1121 > URL: https://issues.apache.org/jira/browse/HDDS-1121 > Project: Hadoop Distributed Data Store > Issue Type: Bug >Reporter: Bharat Viswanadham >Assignee: Bharat Viswanadham >Priority: Major > Attachments: HDDS-1121.00.patch > > > When hive is run with multiple threads for data ingestion to ozone. After > ingestion is done, during read we see this below error. > This issue is found during hive testing. > {code:java} > caused by: org.apache.hadoop.ozone.common.OzoneChecksumException: Checksum > mismatch at index 0 > at > org.apache.hadoop.ozone.common.ChecksumData.verifyChecksumDataMatches(ChecksumData.java:143) > at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:239) > at org.apache.hadoop.ozone.common.Checksum.verifyChecksum(Checksum.java:217) > at > org.apache.hadoop.hdds.scm.storage.BlockInputStream.readChunkFromContainer(BlockInputStream.java:227) > at > org.apache.hadoop.hdds.scm.storage.BlockInputStream.seek(BlockInputStream.java:259) > at > org.apache.hadoop.ozone.client.io.KeyInputStream$ChunkInputStreamEntry.seek(KeyInputStream.java:249) > at > org.apache.hadoop.ozone.client.io.KeyInputStream.seek(KeyInputStream.java:180) > at > org.apache.hadoop.fs.ozone.OzoneFSInputStream.seek(OzoneFSInputStream.java:62) > at org.apache.hadoop.fs.FSInputStream.read(FSInputStream.java:82) > at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121) > at > org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:111) > at org.apache.orc.impl.ReaderImpl.extractFileTail(ReaderImpl.java:555) > at org.apache.orc.impl.ReaderImpl.(ReaderImpl.java:370) > at org.apache.hadoop.hive.ql.io.orc.ReaderImpl.(ReaderImpl.java:61) > at org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader(OrcFile.java:105) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails(OrcInputFormat.java:1647) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.callInternal(OrcInputFormat.java:1533) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.access$2700(OrcInputFormat.java:1329) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1513) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1510) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1688) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1510) > at > org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1329) > at java.util.concurrent.FutureTask.run(FutureTask.java:266){code} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org