Hey Scott,
Hadoop tends to get confused by nodes with multiple hostnames or multiple IP
addresses. Is this your case?
I can't remember precisely what our admin does, but I think he puts in the IP
address which Hadoop listens on in the exclude-hosts file.
Look in the output of
hadoop
Hi Scott,
You might be hitting two different issues.
1) Decommission not finishing.
https://issues.apache.org/jira/browse/HDFS-694 explains decommission
never finishing due to open files in 0.20
2) Nodes showing up both in live and dead nodes.
I remember Suresh taking a look at this.
Dfsadmin -report reports the hostname for that machine and not the ip. That
machine happens to be the master node which is why I am trying to
decommission the data node there since I only want the data node running on
the slave nodes. Dfs admin -report reports all the ips for the slave nodes.
One
Hey Scott,
If the node shows up in the dead nodes and the live nodes as you say, it's
definitely not even attempting to be decommissioned. If HDFS was attempting
decommissioning and you restart the namenode, then it would only show up in the
dead nodes list.
Another option is to just turn
I followed the steps mentioned here:
http://developer.yahoo.com/hadoop/tutorial/module2.html#decommission to
decommission a data node. What I see from the namenode is the hostname of
the machine that I decommissioned shows up in both the list of dead nodes
but also live nodes where its admin