[ 
https://issues.apache.org/jira/browse/MAPREDUCE-1919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12887670#action_12887670
 ] 

Balaji Rajagopalan commented on MAPREDUCE-1919:
-----------------------------------------------

+    JobInfo jInfo = wovenClient.getJobInfo(jobId);
+    LOG.info("Waiting till the job is completed...");
+    while (!jInfo.getStatus().isJobComplete()) {
+      UtilsForTests.waitFor(100);
+      jInfo = wovenClient.getJobInfo(jobId);
+    }

Vinay please use the building block JTClient::isJobStopped. There are couple of 
places this needs to be fixed. 

+  private int runTool(Configuration job, Tool tool, 
+    String[] jobArgs) throws Exception {
+    int returnStatus = ToolRunner.run(job, tool, jobArgs);
+    return returnStatus;
+  }

The above can be made as a helper class, since it will be useful for many test 
cases. 



> [Herriot] Test for verification of per cache file ref  count.
> -------------------------------------------------------------
>
>                 Key: MAPREDUCE-1919
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-1919
>             Project: Hadoop Map/Reduce
>          Issue Type: Task
>          Components: test
>            Reporter: Vinay Kumar Thota
>            Assignee: Vinay Kumar Thota
>         Attachments: 1919-ydist-security.patch, MAPREDUCE-1919.patch
>
>
> It covers the following scenarios.
> 1. Run the job with two distributed cache files and verify whether job is 
> succeeded or not.
> 2.  Run the job with distributed cache files and remove one cache file from 
> the DFS when it is localized.verify whether the job is failed or not.
> 3.  Run the job with two distribute cache files and the size of  one file 
> should be larger than local.cache.size.Verify  whether job is succeeded or 
> not.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to