Each reduce task (Nutch indexing job) gets as far as 66%, and then fails with 
the following error:

"Task failed to report status for 600 seconds. Killing."

In the end, no reduce task completes successfully. Besides solves this issue, I was wondering if I could update code and configuration and start the reduce phase again without the need to redo all map tasks (that saves me 2 hours). Assuming of course that the output of the map tasks has not changed.

Mathijs



Arun C Murthy wrote:
Hi Mathijs,

Mathijs Homminga wrote:

We have some troubles with the reduce phase of our job.
Is it possible to re-execute the reduce tasks without the need to do all map tasks again?


That the MR-framework already does... you don't have to re-execute the maps for the *failed* reduces. Are you noticing something else?

What are the 'troubles' you allude to? Also with once we get HADOOP-1127 in, you should try turing on 'speculative execution' - that helps when some tasks are very slow w.r.t other similar tasks.

Arun

Thanks!
Mathijs Homminga


Reply via email to