Hi Bing,

I would recommend that you add your 50 new nodes to the cluster and the 
decommission the 50 nodes you want to get rid of.  You can do the decommission 
in one operation (albeit a lengthy operation) by adding the nodes you want to 
decommission to your HDFS exclude file and running `hdfs dfsadmin 
-refreshNodes`.  The decommission process will ensure that the data from your 
old nodes are distributed across your cluster.

Cheers,

Rich

On Jul 1, 2015, at 3:30 AM, Bing Jiang 
<jiangbinglo...@gmail.com<mailto:jiangbinglo...@gmail.com>> wrote:

hi, guys.

I want to move all the blocks from 50 datanodes to another 50 new datanodes. 
There is a very easy idea that we can add the new 50 nodes to hadoop cluster 
firstly, then decommission the other 50 nodes one by one. But, I believe it is 
not an efficient way to reach the goal.
So I plan to get the help of  the idea of Hdfs balancer, which limits the 
movements  happen to the 100 nodes. But I need to disable write_op for those 
nodes to be decommissioned.

Is there a way to make DataNode in safe mode (read only) mode?

Of course, regarding to blocks movement between nodes,  Any thoughts will be 
appreciated.



--
Bing Jiang


Rich Haase| Sr. Software Engineer | Pandora
m (303) 887-1146 | rha...@pandora.com<mailto:rha...@pandora.com>




Reply via email to