Submitting job information via DFS in Map/Reduce causing consistency and 
performance issues
-------------------------------------------------------------------------------------------

                 Key: HADOOP-5350
                 URL: https://issues.apache.org/jira/browse/HADOOP-5350
             Project: Hadoop Core
          Issue Type: Improvement
          Components: mapred
            Reporter: Hemanth Yamijala


Job submission involves two steps: submitting jobs to the System directory on 
DFS (done by the client), then submit the job via the JobSubmissionProtocol to 
JobTracker. This two step process is seen to have some issues:

- Since the files need to be read from DFS, slowness in the DFS can cause job 
initialization to become costly. We faced this as described in HADOOP-5286 and 
HADOOP-4664.
- The two step process could lead to inconsistent information being left around 
- like in HADOOP-5327 and HADOOP-5335.

This JIRA is to explore options to remove the two step process in submitting a 
job.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to