[ 
https://issues.apache.org/jira/browse/HDFS-1171?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12921054#action_12921054
 ] 

Patrick Kling commented on HDFS-1171:
-------------------------------------

In order to be able to handle a larger number of simultaneous block losses, we 
should have the option of dispatching map reduce jobs to fix the affected files 
in a distributed fashion. For greater flexibility, we propose a configuration 
option that specifies whether to use the local or the distributed 
implementation of the block fixer.

> RaidNode should fix missing blocks directly on Data Node
> --------------------------------------------------------
>
>                 Key: HDFS-1171
>                 URL: https://issues.apache.org/jira/browse/HDFS-1171
>             Project: Hadoop HDFS
>          Issue Type: Task
>          Components: contrib/raid
>    Affects Versions: 0.20.1
>            Reporter: Ramkumar Vadali
>
> RaidNode currently does not fix missing blocks. The missing blocks have to be 
> fixed manually.
> This task proposes that recovery be more automated:
> 1. RaidNode periodically fetches a list of corrupt files from the NameNode
> 2. If the corrupt files has a RAID parity file, RaidNode identifies missing 
> block(s) in the file and recomputes the block(s) using the parity file and 
> other good blocks
> 3. RaidNode sends the generated block contents to a DataNode
>    a. RaidNode chooses a DataNode with the most available space to send the 
> block. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to