Steve Loughran commented on HADOOP-14081:

# as usual, please state which s3 region you tested with. As this plays with 
blocks, please use -Dscale for the scale tests, ideally choosing a larger file 
size for more rigorousnes
# This patch will break against HADOOP-14208, whichever gets in first will 
force the other to catch up. That's a bug so has priority for me, but this is a 
lower cost patch which I can vote in. So: once you confirm the testing, I'll 
take a look at this

> S3A: Consider avoiding array copy in S3ABlockOutputStream (ByteArrayBlock)
> --------------------------------------------------------------------------
>                 Key: HADOOP-14081
>                 URL: https://issues.apache.org/jira/browse/HADOOP-14081
>             Project: Hadoop Common
>          Issue Type: Sub-task
>          Components: fs/s3
>            Reporter: Rajesh Balamohan
>            Priority: Minor
>         Attachments: HADOOP-14081.001.patch
> In {{S3ADataBlocks::ByteArrayBlock}}, data is copied whenever {{startUpload}} 
> is called. It might be possible to directly access the byte[] array from 
> ByteArrayOutputStream. 
> Might have to extend ByteArrayOutputStream and create a method like 
> getInputStream() which can return ByteArrayInputStream.  This would avoid 
> expensive array copy during large upload.

This message was sent by Atlassian JIRA

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org

Reply via email to