[ https://issues.apache.org/jira/browse/HADOOP-15224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Raphael Azzolini resolved HADOOP-15224. --------------------------------------- Resolution: Fixed > Add option to set checksum on S3 object uploads > ----------------------------------------------- > > Key: HADOOP-15224 > URL: https://issues.apache.org/jira/browse/HADOOP-15224 > Project: Hadoop Common > Issue Type: Sub-task > Components: fs/s3 > Affects Versions: 3.0.0 > Reporter: Steve Loughran > Assignee: Raphael Azzolini > Priority: Minor > Labels: pull-request-available > Fix For: 3.5.0, 3.4.2 > > > [~rdblue] reports sometimes he sees corrupt data on S3. Given MD5 checks from > upload to S3, its likelier to have happened in VM RAM, HDD or nearby. > If the MD5 checksum for each block was built up as data was written to it, > and checked against the etag RAM/HDD storage of the saved blocks could be > removed as sources of corruption > The obvious place would be > {{org.apache.hadoop.fs.s3a.S3ADataBlocks.DataBlock}} -- This message was sent by Atlassian Jira (v8.20.10#820010) --------------------------------------------------------------------- To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-dev-h...@hadoop.apache.org