[ 
https://issues.apache.org/jira/browse/HADOOP-4638?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12677247#action_12677247
 ] 

Amar Kamat commented on HADOOP-4638:
------------------------------------

Amareshwari,
The job gets ignored before it gets added to jobtracker (i.e if the filename is 
not recoverable or restoration of master file fails). Once the filename is 
recovered, the recovery manager recovers whatever it can and continues. There 
is no killing done after that. Hence there is no need to do a job.kill(). 

> Exception thrown in/from RecoveryManager.recover() should be caught and 
> handled
> -------------------------------------------------------------------------------
>
>                 Key: HADOOP-4638
>                 URL: https://issues.apache.org/jira/browse/HADOOP-4638
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: mapred
>    Affects Versions: 0.20.0
>            Reporter: Amar Kamat
>            Assignee: Amar Kamat
>            Priority: Blocker
>             Fix For: 0.20.0
>
>         Attachments: HADOOP-4638-v1.1.patch, HADOOP-4638-v1.3.patch, 
> HADOOP-4638-v1.6.patch, HADOOP-4638-v1.8.patch
>
>
> {{RecoveryManager.recover()}} can throw an exception while recovering a job. 
> Since the {{JobTracker}} calls {{RecoveryManager.recover()}} from 
> {{offerService()}}, any failure in recovery will cause {{JobTracker}} to 
> crash. Ideally the {{RecoveryManager}} should log the failure encountered 
> while recovering the job and continue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to