[
https://issues.apache.org/jira/browse/HADOOP-19654?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=18032833#comment-18032833
]
ASF GitHub Bot commented on HADOOP-19654:
-----------------------------------------
steveloughran commented on PR #7882:
URL: https://github.com/apache/hadoop/pull/7882#issuecomment-3444243683
OK, this is all related to checksums on multipart puts.
If you declare that checksums are always required on requests, you MUST
define a checksum algorithm to use for multipart put, otherwise upload
completions fail. I have no idea why, will file some SDK bug report to say
"this is wrong" and simply change our settings to
- checksums NOT always required
- MD5 always enabled
- checksum algorithm is CRC32C (will test with third party store)
checksums in MPUs breaks a couple of the multipart uploader tests; more
worried that about a ITestS3AOpenCost test failing with checksum verification
being enabled (slow, expensive). I need to make sure that this is not an SDK
regression.
> Upgrade AWS SDK to 2.33.x
> -------------------------
>
> Key: HADOOP-19654
> URL: https://issues.apache.org/jira/browse/HADOOP-19654
> Project: Hadoop Common
> Issue Type: Improvement
> Components: build, fs/s3
> Affects Versions: 3.5.0
> Reporter: Steve Loughran
> Assignee: Steve Loughran
> Priority: Major
> Labels: pull-request-available
>
> Upgrade to a recent version of 2.33.x or later while off the critical path of
> things.
> HADOOP-19485 froze the sdk at a version which worked with third party stores.
> Apparently the new version works; early tests show that Bulk Delete calls
> with third party stores complain about lack of md5 headers, so some tuning is
> clearly going to be needed.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]