[
https://issues.apache.org/jira/browse/HADOOP-1324?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12494004
]
Hadoop QA commented on HADOOP-1324:
-----------------------------------
+1
http://issues.apache.org/jira/secure/attachment/12356848/HADOOP-1324_20070507_1.patch
applied and successfully tested against trunk revision r534975.
Test results:
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/120/testReport/
Console output:
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/120/console
> FSError encountered by one running task should not be fatal to other tasks on
> that node
> ---------------------------------------------------------------------------------------
>
> Key: HADOOP-1324
> URL: https://issues.apache.org/jira/browse/HADOOP-1324
> Project: Hadoop
> Issue Type: Improvement
> Components: mapred
> Affects Versions: 0.12.3
> Reporter: Devaraj Das
> Assigned To: Arun C Murthy
> Attachments: HADOOP-1324_20070507_1.patch
>
>
> Currently, if one task encounters a FSError, it reports that to the
> TaskTracker and the TaskTracker reinitializes itself and effectively loses
> state of all the other running tasks too. This can probably be improved
> especially after the fix for HADOOP-1252. The TaskTracker should probably
> avoid reinitializing itself and instead get blacklisted for that job. Other
> tasks should be allowed to continue as long as they can (complete
> successfully, or, fail either due to disk problems or otherwise).
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.