[
https://issues.apache.org/jira/browse/HADOOP-5652?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12700720#action_12700720
]
Hudson commented on HADOOP-5652:
--------------------------------
Integrated in Hadoop-trunk #811 (See
[http://hudson.zones.apache.org/hudson/job/Hadoop-trunk/811/])
> Reduce does not respect in-memory segment memory limit when number of on disk
> segments == io.sort.factor
> --------------------------------------------------------------------------------------------------------
>
> Key: HADOOP-5652
> URL: https://issues.apache.org/jira/browse/HADOOP-5652
> Project: Hadoop Core
> Issue Type: Bug
> Components: mapred
> Affects Versions: 0.20.0
> Reporter: Chris Douglas
> Assignee: Chris Douglas
> Priority: Minor
> Fix For: 0.21.0
>
> Attachments: 5652-0.patch, 5652-1.patch
>
>
> If the number of on-disk segments is exactly {{io.sort.factor}}, then map
> output segments may be left in memory for the reduce contrary to the
> specification in {{mapred.job.reduce.input.buffer.percent}}.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.