[ https://issues.apache.org/jira/browse/HADOOP-1300?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12502015 ]
Koji Noguchi commented on HADOOP-1300: -------------------------------------- > > is there any reasons to adopt different policies for allocation and > > deletion? > For allocation, where to place a replica has an effect on performance. But > for deletion, the cost of deleting any replica is the same. I meant when replication is set to 4 or higher, are we "maximizing the number of unique racks" for the rest of the blocks? Talking with Hairong, now I understand that we don't do this in allocation/replication to simplify the code. Thanks! > deletion of excess replicas does not take into account 'rack-locality' > ---------------------------------------------------------------------- > > Key: HADOOP-1300 > URL: https://issues.apache.org/jira/browse/HADOOP-1300 > Project: Hadoop > Issue Type: Bug > Components: dfs > Reporter: Koji Noguchi > Assignee: Hairong Kuang > Attachments: excessDel.patch > > > One rack went down today, resulting in one missing block/file. > Looking at the log, this block was originally over-replicated. > 3 replicas on one rack and 1 replica on another. > Namenode decided to delete the latter, leaving 3 replicas on the same rack. > It'll be nice if the deletion is also rack-aware. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.