[ https://issues.apache.org/jira/browse/HADOOP-14081?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15869810#comment-15869810 ]
Steve Loughran commented on HADOOP-14081: ----------------------------------------- # as usual, please state which s3 region you tested with. As this plays with blocks, please use -Dscale for the scale tests, ideally choosing a larger file size for more rigorousnes # This patch will break against HADOOP-14208, whichever gets in first will force the other to catch up. That's a bug so has priority for me, but this is a lower cost patch which I can vote in. So: once you confirm the testing, I'll take a look at this > S3A: Consider avoiding array copy in S3ABlockOutputStream (ByteArrayBlock) > -------------------------------------------------------------------------- > > Key: HADOOP-14081 > URL: https://issues.apache.org/jira/browse/HADOOP-14081 > Project: Hadoop Common > Issue Type: Sub-task > Components: fs/s3 > Reporter: Rajesh Balamohan > Priority: Minor > Attachments: HADOOP-14081.001.patch > > > In {{S3ADataBlocks::ByteArrayBlock}}, data is copied whenever {{startUpload}} > is called. It might be possible to directly access the byte[] array from > ByteArrayOutputStream. > Might have to extend ByteArrayOutputStream and create a method like > getInputStream() which can return ByteArrayInputStream. This would avoid > expensive array copy during large upload. -- This message was sent by Atlassian JIRA (v6.3.15#6346) --------------------------------------------------------------------- To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org