[ https://issues.apache.org/jira/browse/HDFS-1171?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12921054#action_12921054 ]
Patrick Kling commented on HDFS-1171: ------------------------------------- In order to be able to handle a larger number of simultaneous block losses, we should have the option of dispatching map reduce jobs to fix the affected files in a distributed fashion. For greater flexibility, we propose a configuration option that specifies whether to use the local or the distributed implementation of the block fixer. > RaidNode should fix missing blocks directly on Data Node > -------------------------------------------------------- > > Key: HDFS-1171 > URL: https://issues.apache.org/jira/browse/HDFS-1171 > Project: Hadoop HDFS > Issue Type: Task > Components: contrib/raid > Affects Versions: 0.20.1 > Reporter: Ramkumar Vadali > > RaidNode currently does not fix missing blocks. The missing blocks have to be > fixed manually. > This task proposes that recovery be more automated: > 1. RaidNode periodically fetches a list of corrupt files from the NameNode > 2. If the corrupt files has a RAID parity file, RaidNode identifies missing > block(s) in the file and recomputes the block(s) using the parity file and > other good blocks > 3. RaidNode sends the generated block contents to a DataNode > a. RaidNode chooses a DataNode with the most available space to send the > block. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.