[ 
https://issues.apache.org/jira/browse/HDFS-5662?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-5662:
-----------------------------

    Attachment: HDFS-5662.branch2.3.patch

The patch to branch2.3 is slightly different in the unit test. Attached it here.

> Can't decommission a DataNode due to file's replication factor larger than 
> the rest of the cluster size
> -------------------------------------------------------------------------------------------------------
>
>                 Key: HDFS-5662
>                 URL: https://issues.apache.org/jira/browse/HDFS-5662
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: namenode
>            Reporter: Brandon Li
>            Assignee: Brandon Li
>             Fix For: 2.3.0
>
>         Attachments: HDFS-5662.001.patch, HDFS-5662.002.patch, 
> HDFS-5662.branch2.3.patch
>
>
> A datanode can't be decommissioned if it has replica that belongs to a file 
> with a replication factor larger than the rest of the cluster size.
> One way to fix this is to have some kind of minimum replication factor 
> setting and thus any datanode can be decommissioned regardless of the largest 
> replication factor it's related to. 



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)

Reply via email to