[
https://issues.apache.org/jira/browse/HADOOP-14081?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15869810#comment-15869810
]
Steve Loughran commented on HADOOP-14081:
-----------------------------------------
# as usual, please state which s3 region you tested with. As this plays with
blocks, please use -Dscale for the scale tests, ideally choosing a larger file
size for more rigorousnes
# This patch will break against HADOOP-14208, whichever gets in first will
force the other to catch up. That's a bug so has priority for me, but this is a
lower cost patch which I can vote in. So: once you confirm the testing, I'll
take a look at this
> S3A: Consider avoiding array copy in S3ABlockOutputStream (ByteArrayBlock)
> --------------------------------------------------------------------------
>
> Key: HADOOP-14081
> URL: https://issues.apache.org/jira/browse/HADOOP-14081
> Project: Hadoop Common
> Issue Type: Sub-task
> Components: fs/s3
> Reporter: Rajesh Balamohan
> Priority: Minor
> Attachments: HADOOP-14081.001.patch
>
>
> In {{S3ADataBlocks::ByteArrayBlock}}, data is copied whenever {{startUpload}}
> is called. It might be possible to directly access the byte[] array from
> ByteArrayOutputStream.
> Might have to extend ByteArrayOutputStream and create a method like
> getInputStream() which can return ByteArrayInputStream. This would avoid
> expensive array copy during large upload.
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]