hey sid ,

thanks but i tried for the same also..

1: first  stop cluster

*"please make note that this file must be included at the namenode startup"*

is this wat you mean is to add file at *dfs.hosts.exclude* in hdfs-site and

after that start-dfs.sh ,start-mapred.sh

and after starting server , i need to add the ip:port at exlude file ...

then cmd "*-refreshNodes*"

regards
Vikas


On Mon, Sep 19, 2011 at 6:40 PM, Siddharth Tiwari <siddharth.tiw...@live.com
> wrote:

>  Hi vikas,
>
> Please include your exclude file in *hdfs-site.xml* under the 
> property*dfs.hosts.exclude
> *.
> please make note that this file must be included at the namenode startup
> and then u hv to edit it for the datanodes you want to exclude. Secondly,
> after editin it for the ip:port of the retiring datanode, use the command
> *bin/hadoop dfsadmin -refreshNodes*.
>
> also, Decommission does not happen momentarily since it requires
> replication of potentially a large number of blocks and we do not want the
> cluster to be overwhelmed with just this one job. The decommission progress
> can be monitored on the name-node Web UI. Until all blocks are replicated
> the node will be in "Decommission In Progress" state. When decommission is
> done the state will change to "Decommissioned". The nodes can be removed
> whenever decommission is finished.
>
> The decommission process can be terminated at any time by editing the
> configuration or the exclude files and repeating the *-refreshNodes*command.
>
> hope it helps.
>
> **------------------------**
> *Cheers !!!*
> *Siddharth Tiwari*
> *TCS world wide Data warehouse and Analytic Team - Americas*
> Have a refreshing day !!!
>
>
> ------------------------------
> Date: Mon, 19 Sep 2011 12:50:13 +0530
> Subject: Decommission of datanode(Urgent)
> From: vikas.srivast...@one97.net
> To: user@hive.apache.org; ayonsi...@yahoo.com; viral.baja...@gmail.com
> CC: nitin2.ku...@one97.net; abhinav.mis...@one97.net
>
>
> Hey folks !!
>
> i tried to decommission datanode from hadoop cluster... steps i followed
>
> 1: add  this in core site
>
> * <property>
>     <name>dfs.hosts.exclude</name>
>     <value>/home/hadoop/excludes</value>
>     <final>true</final>
>   </property>
> *
> 1: add  this in mapred-site
> *  <property>
>     <name>mapred.hosts.exclude</name>
>
>     <value>/home/hadoop/excludes</value>
>     <final>true</final>
>   </property>*
>
>
> 3:create a excludes file and add *ip:port* in that
>
> exp: *10.0.3.31:50010*
>
> 4: run cmd
>
> *hadoop dfsadmin -refreshNodes*
>
>
> *5: After that my live nodes became 0 and all nodes became dead.. i
> checked namenode logs where i found these error msgs*
>
> 2011-09-19 12:33:47,695 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 24 on 9000, call sendHeartbeat(DatanodeRegistration(
> 10.0.3.16:50010, storageID=DS-1703098060-10.0.3.16-50010-1298269611944,
> infoPort=50075, ipcPort=50020), 2012206694400, 1650194042865, 271003275264,
> 0, 1) from 10.0.3.16:38587: error:
> org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode
> denied communication with namenode: 10.0.3.16:50010
> org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException:
> Datanode denied communication with namenode: 10.0.3.16:50010
>         at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.handleHeartbeat(FSNamesystem.java:2235)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.sendHeartbeat(NameNode.java:704)
>         at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>         at java.lang.reflect.Method.invoke(Method.java:597)
>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:396)
>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)
> 2011-09-19 12:33:47,701 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 7 on 9000, call sendHeartbeat(DatanodeRegistration(10.0.5.36:50010,
> storageID=DS-809855347-10.0.5.36-50010-1316252293924, infoPort=50075,
> ipcPort=50020), 1938687860736, 1390486994944, 457712619520, 0, 1) from
> 10.0.5.36:58924: error:
> org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode
> denied communication with namenode: 10.0.5.36:50010
> org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException:
> Datanode denied communication with namenode: 10.0.5.36:50010
>         at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.handleHeartbeat(FSNamesystem.java:2235)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.sendHeartbeat(NameNode.java:704)
>         at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>         at java.lang.reflect.Method.invoke(Method.java:597)
>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:396)
>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)
>
>
>
>
> please suggest any help would be appreciated!!!!!!!!!!
>
>
> --
> With Regards
> Vikas Srivastava
>
> DWH & Analytics Team
> Mob:+91 9560885900
> One97 | Let's get talking !
>
>


-- 
With Regards
Vikas Srivastava

DWH & Analytics Team
Mob:+91 9560885900
One97 | Let's get talking !

Reply via email to