[
https://issues.apache.org/jira/browse/HADOOP-9454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13686297#comment-13686297
]
Hadoop QA commented on HADOOP-9454:
-----------------------------------
{color:red}-1 overall{color}. Here are the results of testing the latest
attachment
http://issues.apache.org/jira/secure/attachment/12588273/HADOOP-9454-10.patch
against trunk revision .
{color:green}+1 @author{color}. The patch does not contain any @author
tags.
{color:green}+1 tests included{color}. The patch appears to include 2 new
or modified test files.
{color:red}-1 javac{color}. The applied patch generated 1154 javac
compiler warnings (more than the trunk's current 1152 warnings).
{color:green}+1 javadoc{color}. The javadoc tool did not generate any
warning messages.
{color:green}+1 eclipse:eclipse{color}. The patch built with
eclipse:eclipse.
{color:green}+1 findbugs{color}. The patch does not introduce any new
Findbugs (version 1.3.9) warnings.
{color:green}+1 release audit{color}. The applied patch does not increase
the total number of release audit warnings.
{color:green}+1 core tests{color}. The patch passed unit tests in
hadoop-common-project/hadoop-common.
{color:green}+1 contrib tests{color}. The patch passed contrib unit tests.
Test results:
https://builds.apache.org/job/PreCommit-HADOOP-Build/2662//testReport/
Javac warnings:
https://builds.apache.org/job/PreCommit-HADOOP-Build/2662//artifact/trunk/patchprocess/diffJavacWarnings.txt
Console output:
https://builds.apache.org/job/PreCommit-HADOOP-Build/2662//console
This message is automatically generated.
> Support multipart uploads for s3native
> --------------------------------------
>
> Key: HADOOP-9454
> URL: https://issues.apache.org/jira/browse/HADOOP-9454
> Project: Hadoop Common
> Issue Type: Improvement
> Components: fs/s3
> Reporter: Jordan Mendelson
> Attachments: HADOOP-9454-10.patch
>
>
> The s3native filesystem is limited to 5 GB file uploads to S3, however the
> newest version of jets3t supports multipart uploads to allow storing multi-TB
> files. While the s3 filesystem lets you bypass this restriction by uploading
> blocks, it is necessary for us to output our data into Amazon's
> publicdatasets bucket which is shared with others.
> Amazon has added a similar feature to their distribution of hadoop as has
> MapR.
> Please note that while this supports large copies, it does not yet support
> parallel copies because jets3t doesn't expose an API yet that allows it
> without hadoop controlling the threads unlike with upload.
> By default, this patch does not enable multipart uploads. To enable them and
> parallel uploads:
> add the following keys to your hadoop config:
> <property>
> <name>fs.s3n.multipart.uploads.enabled</name>
> <value>true</value>
> </property>
> <property>
> <name>fs.s3n.multipart.uploads.block.size</name>
> <value>67108864</value>
> </property>
> <property>
> <name>fs.s3n.multipart.copy.block.size</name>
> <value>5368709120</value>
> </property>
> create a /etc/hadoop/conf/jets3t.properties file with or similar to:
> storage-service.internal-error-retry-max=5
> storage-service.disable-live-md5=false
> threaded-service.max-thread-count=20
> threaded-service.admin-max-thread-count=20
> s3service.max-thread-count=20
> s3service.admin-max-thread-count=20
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira