[ 
https://issues.apache.org/jira/browse/MAPREDUCE-1819?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ramkumar Vadali updated MAPREDUCE-1819:
---------------------------------------

    Attachment: MAPREDUCE-1819.5.patch

Modifying patch after svn up.

Test Results:


test-junit:
    [junit] WARNING: multiple versions of ant detected in path for junit 
    [junit]          
jar:file:/home/rvadali/local/external/ant/lib/ant.jar!/org/apache/tools/ant/Project.class
    [junit]      and 
jar:file:/home/rvadali/.ivy2/cache/ant/ant/jars/ant-1.6.5.jar!/org/apache/tools/ant/Project.class
    [junit] Running org.apache.hadoop.hdfs.TestRaidDfs
    [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 28.33 sec
    [junit] Running org.apache.hadoop.raid.TestDirectoryTraversal
    [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 8.603 sec
    [junit] Running org.apache.hadoop.raid.TestRaidHar
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 69.846 sec
    [junit] Running org.apache.hadoop.raid.TestRaidNode
    [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 408.472 sec
    [junit] Running org.apache.hadoop.raid.TestRaidPurge
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 44.013 sec

test:

BUILD SUCCESSFUL
Total time: 9 minutes 34 seconds


> RaidNode should be smarter in submitting Raid jobs
> --------------------------------------------------
>
>                 Key: MAPREDUCE-1819
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-1819
>             Project: Hadoop Map/Reduce
>          Issue Type: Task
>          Components: contrib/raid
>    Affects Versions: 0.20.1
>            Reporter: Ramkumar Vadali
>            Assignee: Ramkumar Vadali
>         Attachments: MAPREDUCE-1819.4.patch, MAPREDUCE-1819.5.patch, 
> MAPREDUCE-1819.patch, MAPREDUCE-1819.patch.2, MAPREDUCE-1819.patch.3
>
>
> The RaidNode currently computes parity files as follows:
> 1. Using RaidNode.selectFiles() to figure out what files to raid for a policy
> 2. Using #1 repeatedly for each configured policy to accumulate a list of 
> files. 
> 3. Submitting a mapreduce job with the list of files from #2 using 
> DistRaid.doDistRaid()
> This task addresses the fact that #2 and #3 happen sequentially. The proposal 
> is to submit a separate mapreduce job for the list of files for each policy 
> and use another thread to track the progress of the submitted jobs. This will 
> help reduce the time taken for files to be raided.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to