[ https://issues.apache.org/jira/browse/HDFS-13088?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16572361#comment-16572361 ]
Íñigo Goiri commented on HDFS-13088: ------------------------------------ I think this is the least intrusive way to provide over-replication for provided blocks. Ideally, this should be done at a file level but as you mentioned this would require many more changes including interface ones. I would vote to start with this; moving forward we should have a story for what happens when this setting and the per-file configuration are available. > Allow HDFS files/blocks to be over-replicated. > ---------------------------------------------- > > Key: HDFS-13088 > URL: https://issues.apache.org/jira/browse/HDFS-13088 > Project: Hadoop HDFS > Issue Type: Sub-task > Reporter: Virajith Jalaparti > Assignee: Virajith Jalaparti > Priority: Major > Attachments: HDFS-13088.001.patch, HDFS-13088.002.patch > > > This JIRA is to add a per-file "over-replication" factor to HDFS. As > mentioned in HDFS-13069, the over-replication factor will be the excess > replicas that will be allowed to exist for a file or block. This is > beneficial if the application deems additional replicas for a file are > needed. In the case of HDFS-13069, it would allow copies of data in PROVIDED > storage to be cached locally in HDFS in a read-through manner. > The Namenode will not proactively meet the over-replication i.e., it does not > schedule replications if the number of replicas for a block is less than > (replication factor + over-replication factor) as long as they are more than > the replication factor of the file. -- This message was sent by Atlassian JIRA (v7.6.3#76005) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org