[
https://issues.apache.org/jira/browse/HADOOP-1300?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12502015
]
Koji Noguchi commented on HADOOP-1300:
--------------------------------------
> > is there any reasons to adopt different policies for allocation and
> > deletion?
> For allocation, where to place a replica has an effect on performance. But
> for deletion, the cost of deleting any replica is the same.
I meant when replication is set to 4 or higher, are we "maximizing the number
of unique racks" for the rest of the blocks?
Talking with Hairong, now I understand that we don't do this in
allocation/replication to simplify the code.
Thanks!
> deletion of excess replicas does not take into account 'rack-locality'
> ----------------------------------------------------------------------
>
> Key: HADOOP-1300
> URL: https://issues.apache.org/jira/browse/HADOOP-1300
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Reporter: Koji Noguchi
> Assignee: Hairong Kuang
> Attachments: excessDel.patch
>
>
> One rack went down today, resulting in one missing block/file.
> Looking at the log, this block was originally over-replicated.
> 3 replicas on one rack and 1 replica on another.
> Namenode decided to delete the latter, leaving 3 replicas on the same rack.
> It'll be nice if the deletion is also rack-aware.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.