[
https://issues.apache.org/jira/browse/HADOOP-657?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12605824#action_12605824
]
Hadoop QA commented on HADOOP-657:
----------------------------------
-1 overall. Here are the results of testing the latest attachment
http://issues.apache.org/jira/secure/attachment/12384165/diskspaceest_v4.patch
against trunk revision 668867.
+1 @author. The patch does not contain any @author tags.
-1 tests included. The patch doesn't appear to include any new or modified
tests.
Please justify why no tests are needed for this patch.
+1 javadoc. The javadoc tool did not generate any warning messages.
+1 javac. The applied patch does not increase the total number of javac
compiler warnings.
+1 findbugs. The patch does not introduce any new Findbugs warnings.
+1 release audit. The applied patch does not increase the total number of
release audit warnings.
+1 core tests. The patch passed core unit tests.
+1 contrib tests. The patch passed contrib unit tests.
Test results:
http://hudson.zones.apache.org/hudson/job/Hadoop-Patch/2677/testReport/
Findbugs warnings:
http://hudson.zones.apache.org/hudson/job/Hadoop-Patch/2677/artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Checkstyle results:
http://hudson.zones.apache.org/hudson/job/Hadoop-Patch/2677/artifact/trunk/build/test/checkstyle-errors.html
Console output:
http://hudson.zones.apache.org/hudson/job/Hadoop-Patch/2677/console
This message is automatically generated.
> Free temporary space should be modelled better
> ----------------------------------------------
>
> Key: HADOOP-657
> URL: https://issues.apache.org/jira/browse/HADOOP-657
> Project: Hadoop Core
> Issue Type: Improvement
> Components: mapred
> Affects Versions: 0.17.0
> Reporter: Owen O'Malley
> Assignee: Ari Rabkin
> Fix For: 0.19.0
>
> Attachments: diskspaceest.patch, diskspaceest_v2.patch,
> diskspaceest_v3.patch, diskspaceest_v4.patch
>
>
> Currently, there is a configurable size that must be free for a task tracker
> to accept a new task. However, that isn't a very good model of what the task
> is likely to take. I'd like to propose:
> Map tasks: totalInputSize * conf.getFloat("map.output.growth.factor", 1.0) /
> numMaps
> Reduce tasks: totalInputSize * 2 * conf.getFloat("map.output.growth.factor",
> 1.0) / numReduces
> where totalInputSize is the size of all the maps inputs for the given job.
> To start a new task,
> newTaskAllocation + (sum over running tasks of (1.0 - done) * allocation)
> >=
> free disk * conf.getFloat("mapred.max.scratch.allocation", 0.90);
> So in English, we will model the expected sizes of tasks and only task tasks
> that should leave us a 10% margin. With:
> map.output.growth.factor -- the relative size of the transient data relative
> to the map inputs
> mapred.max.scratch.allocation -- the maximum amount of our disk we want to
> allocate to tasks.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.