[
https://issues.apache.org/jira/browse/HDFS-13088?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16572361#comment-16572361
]
Íñigo Goiri commented on HDFS-13088:
------------------------------------
I think this is the least intrusive way to provide over-replication for
provided blocks.
Ideally, this should be done at a file level but as you mentioned this would
require many more changes including interface ones.
I would vote to start with this; moving forward we should have a story for what
happens when this setting and the per-file configuration are available.
> Allow HDFS files/blocks to be over-replicated.
> ----------------------------------------------
>
> Key: HDFS-13088
> URL: https://issues.apache.org/jira/browse/HDFS-13088
> Project: Hadoop HDFS
> Issue Type: Sub-task
> Reporter: Virajith Jalaparti
> Assignee: Virajith Jalaparti
> Priority: Major
> Attachments: HDFS-13088.001.patch, HDFS-13088.002.patch
>
>
> This JIRA is to add a per-file "over-replication" factor to HDFS. As
> mentioned in HDFS-13069, the over-replication factor will be the excess
> replicas that will be allowed to exist for a file or block. This is
> beneficial if the application deems additional replicas for a file are
> needed. In the case of HDFS-13069, it would allow copies of data in PROVIDED
> storage to be cached locally in HDFS in a read-through manner.
> The Namenode will not proactively meet the over-replication i.e., it does not
> schedule replications if the number of replicas for a block is less than
> (replication factor + over-replication factor) as long as they are more than
> the replication factor of the file.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]