[
https://issues.apache.org/jira/browse/HADOOP-5977?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12716507#action_12716507
]
Vinod K V commented on HADOOP-5977:
-----------------------------------
+1 given that we are doing 2-3 roundtrips to DFS in the JobInProgress
constructor itself.
Further, the current code has a bug. JT downloads job files from the DFS in the
constructor onto its local file system. When a job is rejected by access/queue
checks, even though the job files on the DFS are deleted, the files on JT's
local FS are never deleted. This happens because JobInProgress.garbageCollect()
is never called.
> Avoid creating JobInProgress objects before Access checks and Queues checks
> are done in JobTracker submitJob
> -------------------------------------------------------------------------------------------------------------
>
> Key: HADOOP-5977
> URL: https://issues.apache.org/jira/browse/HADOOP-5977
> Project: Hadoop Core
> Issue Type: Improvement
> Components: mapred
> Reporter: rahul k singh
>
> In JobTracker submitJob , JobInProgress instance gets created . after this
> checks are done for access and queue state. In event of checks failed . There
> isn't any use for these JIP objects , hence in event of failure only reason
> these objects were created was to get conf data and be deleted.
> We need to fetch the information required to only do the checks instead of
> creating a JobInProgress object
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.