[
https://issues.apache.org/jira/browse/MAPREDUCE-3343?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13206393#comment-13206393
]
Matt Foley commented on MAPREDUCE-3343:
---------------------------------------
This patch has been tested at user sites and is believed stable. Nathan Roberts
requested that I include it in 1.0.1, as its absence is causing ops problems
with 1.0.0.
> TaskTracker Out of Memory because of distributed cache
> ------------------------------------------------------
>
> Key: MAPREDUCE-3343
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-3343
> Project: Hadoop Map/Reduce
> Issue Type: Bug
> Components: mrv1
> Affects Versions: 0.20.205.0
> Reporter: Ahmed Radwan
> Assignee: zhaoyunjiong
> Labels: mapreduce, patch
> Fix For: 1.0.1
>
> Attachments: MAPREDUCE-3343_rev2.patch,
> mapreduce-3343-release-0.20.205.0.patch
>
>
> This Out of Memory happens when you run large number of jobs (using the
> distributed cache) on a TaskTracker.
> Seems the basic issue is with the distributedCacheManager (instance of
> TrackerDistributedCacheManager in TaskTracker.java), this gets created during
> TaskTracker.initialize(), and it keeps references to
> TaskDistributedCacheManager for every submitted job via the jobArchives Map,
> also references to CacheStatus via cachedArchives map. I am not seeing these
> cleaned up between jobs, so this can out of memory problems after really
> large number of jobs are submitted. We have seen this issue in a number of
> cases.
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators:
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira