[ 
http://issues.apache.org/jira/browse/HADOOP-580?page=comments#action_12440557 ] 
            
Benjamin Reed commented on HADOOP-580:
--------------------------------------

No. I'm very against running code in the Trackers (as my mail indicates :). The 
idea would be that you would spawn off a child process at the beginning of a 
job and kill it at the end. (Or some variation on that theme.)

> Job setup and take down on Nodes
> --------------------------------
>
>                 Key: HADOOP-580
>                 URL: http://issues.apache.org/jira/browse/HADOOP-580
>             Project: Hadoop
>          Issue Type: New Feature
>          Components: mapred
>            Reporter: Benjamin Reed
>
> It would be nice if there was a hook for doing job provisioning and cleanup 
> on compute nodes. The TaskTracker implicitly knows when a job starts (a task 
> for the job is received) and pollForTaskWithClosedJob() will explicitly say 
> that a job is finished if a Map task has been run (If only Reduce tasks have 
> run and are finished I don't think pollForTaskWithClosedJob() will return 
> anything will it?), but child Tasks do not get this information.
> It would be nice if there was a hook so that programmers could do some 
> provisioning when a job starts and cleanup when a job ends. Caching addresses 
> some of the provisioning, but in some cases a helper daemon may need to be 
> started or the results of queries need to be retrieved and having startJob(), 
> finishJob() callbacks that happen exactly once for each node that runs part 
> of the job would be wonderful.

-- 
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators: 
http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to