[ https://issues.apache.org/jira/browse/HADOOP-15961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16748039#comment-16748039 ]
Steve Loughran commented on HADOOP-15961: ----------------------------------------- build the patch off trunk.e.g {code} git diff trunk...HEAD > ~/hadoop-patches/work/HADOOP-16961-002.patch {code} (assuming you have that hadoop-patches/work) dir, or similar). Then attach that patch to the JIRA > S3A committers: make sure there's regular progress() calls > ---------------------------------------------------------- > > Key: HADOOP-15961 > URL: https://issues.apache.org/jira/browse/HADOOP-15961 > Project: Hadoop Common > Issue Type: Sub-task > Components: fs/s3 > Reporter: Steve Loughran > Assignee: lqjacklee > Priority: Minor > Attachments: HADOOP-15961-001.patch, HADOOP-15961-002.patch > > > MAPREDUCE-7164 highlights how inside job/task commit more context.progress() > callbacks are needed, just for HDFS. > the S3A committers should be reviewed similarly. > At a glance: > StagingCommitter.commitTaskInternal() is at risk if a task write upload > enough data to the localfs that the upload takes longer than the timeout. > it should call progress it every single file commits, or better: modify > {{uploadFileToPendingCommit}} to take a Progressable for progress callbacks > after every part upload. -- This message was sent by Atlassian JIRA (v7.6.3#76005) --------------------------------------------------------------------- To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org