[
https://issues.apache.org/jira/browse/HADOOP-15576?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16570980#comment-16570980
]
Steve Loughran commented on HADOOP-15576:
-----------------------------------------
I think this is good to go in, if you are happy with my changes
And there's no need for protobuf; we aren't worrying about wire compat over
time, are we?
But there's still some ambiguity about things, something which surfaces
precisely because the spec of HDFS-13713 doesn't exist, and we are left looking
at the behaviours of the two existing implementations and guessing which are
"the reference" behaviours vs "implementation artifacts"
* what policy for 0-entry commits MUST be (here: fail)
* what happens if your MPU complete call, there are uploaded parts which are
not listed?
* what happens if >1 part is listed twice in completion
* what happens if you try to upload a part after the MPU has completed
If we had the time, I'd say "pull that specification task into this one and
define things alongside the tests", that being how I like to do tests and
reverse-engineer a spec from behaviours: build the spec, come up with ways to
break the preconditions, see what happens when you try that in tests, fix
code/refine spec.
But...we are approaching the cutoff for 3.2, and ideally I'd like this in ASAP,
along with the other 3.2 features. Getting this in gives us time to finish &
Review those.
So
I'm +1 for this patch as is. It is working for me against us-west-1 + s3guard,
where it wasn't before. (I'm not reliably testing encryption BTW, as there's no
test actually verifying the object has the encryption header.)
If you are happy with the patch-as-modified, it's good to go. But we do need
that spec still, which I'd like before the actual apps using this stuff come
together
> S3A Multipart Uploader to work with S3Guard and encryption
> -----------------------------------------------------------
>
> Key: HADOOP-15576
> URL: https://issues.apache.org/jira/browse/HADOOP-15576
> Project: Hadoop Common
> Issue Type: Sub-task
> Components: fs/s3
> Affects Versions: 3.2
> Reporter: Steve Loughran
> Assignee: Ewan Higgs
> Priority: Blocker
> Attachments: HADOOP-15576-005.patch, HADOOP-15576-007.patch,
> HADOOP-15576-008.patch, HADOOP-15576.001.patch, HADOOP-15576.002.patch,
> HADOOP-15576.003.patch, HADOOP-15576.004.patch
>
>
> The new Multipart Uploader API of HDFS-13186 needs to work with S3Guard, with
> the tests to demonstrate this
> # move from low-level calls of S3A client to calls of WriteOperationHelper;
> adding any new methods are needed there.
> # Tests. the tests of HDFS-13713.
> # test execution, with -DS3Guard, -DAuth
> There isn't an S3A version of {{AbstractSystemMultipartUploaderTest}}, and
> even if there was, it might not show that S3Guard was bypassed, because
> there's no checks that listFiles/listStatus shows the newly committed files.
> Similarly, because MPU requests are initiated in S3AMultipartUploader,
> encryption settings are't picked up. Files being uploaded this way *are not
> being encrypted*
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]