[
https://issues.apache.org/jira/browse/MAPREDUCE-430?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12747514#action_12747514
]
Arun C Murthy commented on MAPREDUCE-430:
-----------------------------------------
Aaron - I apologize for being unclear and assuming people got the context.
The distinction mainly arises from the desire to weed out node/hardware
failures as opposed to application errors. The the distinction is rooted in the
desire to treat node/hardware errors (disk, corrupt RAM/NIC etc.) differently
so as to quickly and accurately penalize the node (e.g. blacklist the
tasktracker). Currently we use _all_ task failures uniformly to penalize the
tasktracker... clearly penalizing a tasktracker for an application error which
results in an OOM (for e.g.) is injudicious. Does that make sense?
> Task stuck in cleanup with OutOfMemoryErrors
> --------------------------------------------
>
> Key: MAPREDUCE-430
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-430
> Project: Hadoop Map/Reduce
> Issue Type: Bug
> Reporter: Amareshwari Sriramadasu
> Assignee: Amar Kamat
> Fix For: 0.20.1
>
> Attachments: MAPREDUCE-430-v1.11.patch,
> MAPREDUCE-430-v1.12-branch-0.20.patch, MAPREDUCE-430-v1.12.patch,
> MAPREDUCE-430-v1.6-branch-0.20.patch, MAPREDUCE-430-v1.6.patch,
> MAPREDUCE-430-v1.7.patch, MAPREDUCE-430-v1.8.patch
>
>
> Obesrved a task with OutOfMemory error, stuck in cleanup.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.