I can SSH both ways .i.e. From master to slave and slave to master. the datanode is getting intialized at master but the log at slave looks like this
************************************************************/ 2009-06-04 15:20:06,066 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting DataNode STARTUP_MSG: host = ************ STARTUP_MSG: args = [] STARTUP_MSG: version = 0.18.3 STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18 -r 736250; compiled by 'ndaley' on Thu Jan 22 23:12:08 UTC 2009 ************************************************************/ 2009-06-04 15:20:08,826 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: master/198.55.35.229:54310. Already tried 0 time(s). 2009-06-04 15:20:09,829 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: master/198.55.35.229:54310. Already tried 1 time(s). 2009-06-04 15:20:10,831 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: master/198.55.35.229:54310. Already tried 2 time(s). 2009-06-04 15:20:11,832 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: master/198.55.35.229:54310. Already tried 3 time(s). 2009-06-04 15:20:12,834 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: master/198.55.35.229:54310. Already tried 4 time(s). 2009-06-04 15:20:13,837 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: master/198.55.35.229:54310. Already tried 5 time(s). 2009-06-04 15:20:14,840 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: master/198.55.35.229:54310. Already tried 6 time(s). 2009-06-04 15:20:15,841 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: master/198.55.35.229:54310. Already tried 7 time(s). 2009-06-04 15:20:16,844 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: master/198.55.35.229:54310. Already tried 8 time(s). 2009-06-04 15:20:17,847 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: master/198.55.35.229:54310. Already tried 9 time(s). 2009-06-04 15:20:17,873 ERROR org.apache.hadoop.dfs.DataNode: java.io.IOException: Call to master/198.55.35.229:54310 failed on local exception: java.net.NoRouteToHostException: No route to host at org.apache.hadoop.ipc.Client.wrapException(Client.java:751) at org.apache.hadoop.ipc.Client.call(Client.java:719) at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216) at org.apache.hadoop.dfs.$Proxy4.getProtocolVersion(Unknown Source) at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:348) at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:335) at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:372) at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:309) at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:286) at org.apache.hadoop.dfs.DataNode.startDataNode(DataNode.java:277) at org.apache.hadoop.dfs.DataNode.<init>(DataNode.java:223) at org.apache.hadoop.dfs.DataNode.makeInstance(DataNode.java:3071) at org.apache.hadoop.dfs.DataNode.instantiateDataNode(DataNode.java:3026) at org.apache.hadoop.dfs.DataNode.createDataNode(DataNode.java:3034) at org.apache.hadoop.dfs.DataNode.main(DataNode.java:3156) Caused by: java.net.NoRouteToHostException: No route to host at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:574) at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:100) at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:301) at org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:178) at org.apache.hadoop.ipc.Client.getConnection(Client.java:820) at org.apache.hadoop.ipc.Client.call(Client.java:705) ... 13 more 2009-06-04 15:20:17,874 INFO org.apache.hadoop.dfs.DataNode: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down DataNode at ******* **********************************************************************************88 Please suggest. Asif. On Thu, Jun 4, 2009 at 4:15 PM, asif md <asif.d...@gmail.com> wrote: > @Ravi > > thanx ravi .. i'm now using my a definded tmp dir so the second issue is > resolved. > > But i have ssh keys tht have passwords. But i am able to ssh to the slave > and master from the master. > > should i be able to do tht from the slave as well. > > @ALL > > Any suggestions. > > Thanx > > Asif. > > > On Thu, Jun 4, 2009 at 3:17 PM, Ravi Phulari <rphul...@yahoo-inc.com>wrote: > >> From logs looks like your Hadoop cluster is facing two different issues >> . >> >> At Slave >> >> 1. exception: java.net.NoRouteToHostException: No route to host in >> your logs >> >> Diagnosis - One of your nodes cannot be reached correctly. Make sure you >> can ssh to your master and slave and passwordless ssh keys are set . >> >> At master >> 2. java.io.IOException: Incompatible namespaceIDs in >> Diagnosis - Your Hadoop namespaceID became corrupted. Unfortunately the >> easiest thing to do reformat the HDFS >> >> As you have not configured hadoop.tmp.dir or data.dir by default Hadoop >> will you /tmp as directory for temporary files , log files and for data dir >> which is not good practice . >> I would suggest you using some tmpForHadoop dir somewhere else than /tmp >> . >> >> - >> Ravi >> >> On 6/4/09 12:39 PM, "asif md" <asif.d...@gmail.com> wrote: >> >> Hello all, >> >> I'm trying to setup a two node cluster < remote > using the following >> tutorials >> { NOTE : i'm ignoring the tmp directory property in hadoop-site.xml >> suggested by Michael } >> >> Running Hadoop On Ubuntu Linux (Single-Node Cluster) - Michael G. >> Noll< >> http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29 >> > >> Running Hadoop On Ubuntu Linux (Multi-Node Cluster) - Michael G. >> Noll< >> http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_%28Multi-Node_Cluster%29 >> > >> >> >> I get the following logs when try to run $HADOOP_HOME/bin/start-dfs.sh at >> master. >> >> >> *************************************************************************************************** >> AT MASTER : >> ----------------- >> 2009-06-04 12:16:30,864 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG: >> /************************************************************ >> STARTUP_MSG: Starting DataNode >> STARTUP_MSG: host = ******* >> STARTUP_MSG: args = [] >> STARTUP_MSG: version = 0.18.3 >> STARTUP_MSG: build = >> https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18 -r >> 736250; >> compiled by 'ndaley' on Thu Jan 22 23:12:08 UTC 2009 >> ************************************************************/ >> 2009-06-04 12:16:31,071 ERROR org.apache.hadoop.dfs.DataNode: >> java.io.IOException: Incompatible namespaceIDs in /tmp/*****/dfs/data: >> namenode namespaceID = 34351921; datanode namespaceID = 539590337 >> at >> org.apache.hadoop.dfs.DataStorage.doTransition(DataStorage.java:226) >> at >> >> org.apache.hadoop.dfs.DataStorage.recoverTransitionRead(DataStorage.java:141) >> at org.apache.hadoop.dfs.DataNode.startDataNode(DataNode.java:306) >> at org.apache.hadoop.dfs.DataNode.<init>(DataNode.java:223) >> at org.apache.hadoop.dfs.DataNode.makeInstance(DataNode.java:3071) >> at >> org.apache.hadoop.dfs.DataNode.instantiateDataNode(DataNode.java:3026) >> at >> org.apache.hadoop.dfs.DataNode.createDataNode(DataNode.java:3034) >> at org.apache.hadoop.dfs.DataNode.main(DataNode.java:3156) >> >> 2009-06-04 12:16:31,071 INFO org.apache.hadoop.dfs.DataNode: SHUTDOWN_MSG: >> /************************************************************ >> >> ******************************************************************************************************* >> AT SLAVE : >> ---------------- >> 2009-06-04 12:16:28,203 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG: >> /************************************************************ >> STARTUP_MSG: Starting DataNode >> STARTUP_MSG: host = **************** >> STARTUP_MSG: args = [] >> STARTUP_MSG: version = 0.18.3 >> STARTUP_MSG: build = >> https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18 -r >> 736250; >> compiled by 'ndaley' on Thu Jan 22 23:12:08 UTC 2009 >> ************************************************************/ >> 2009-06-04 12:16:32,175 INFO org.apache.hadoop.ipc.Client: Retrying >> connect >> to server: master/198.55.35.229:54310. Already tried 0 time(s). >> 2009-06-04 12:16:33,178 INFO org.apache.hadoop.ipc.Client: Retrying >> connect >> to server: master/198.55.35.229:54310. Already tried 1 time(s). >> 2009-06-04 12:16:34,181 INFO org.apache.hadoop.ipc.Client: Retrying >> connect >> to server: master/198.55.35.229:54310. Already tried 2 time(s). >> 2009-06-04 12:16:35,184 INFO org.apache.hadoop.ipc.Client: Retrying >> connect >> to server: master/198.55.35.229:54310. Already tried 3 time(s). >> 2009-06-04 12:16:36,187 INFO org.apache.hadoop.ipc.Client: Retrying >> connect >> to server: master/198.55.35.229:54310. Already tried 4 time(s). >> 2009-06-04 12:16:37,190 INFO org.apache.hadoop.ipc.Client: Retrying >> connect >> to server: master/198.55.35.229:54310. Already tried 5 time(s). >> 2009-06-04 12:16:38,193 INFO org.apache.hadoop.ipc.Client: Retrying >> connect >> to server: master/198.55.35.229:54310. Already tried 6 time(s). >> 2009-06-04 12:16:39,196 INFO org.apache.hadoop.ipc.Client: Retrying >> connect >> to server: master/198.55.35.229:54310. Already tried 7 time(s). >> 2009-06-04 12:16:40,198 INFO org.apache.hadoop.ipc.Client: Retrying >> connect >> to server: master/198.55.35.229:54310. Already tried 8 time(s). >> 2009-06-04 12:16:41,200 INFO org.apache.hadoop.ipc.Client: Retrying >> connect >> to server: master/198.55.35.229:54310. Already tried 9 time(s). >> 2009-06-04 12:16:41,222 ERROR org.apache.hadoop.dfs.DataNode: >> java.io.IOException: Call to master/198.55.35.229:54310 failed on local >> exception: java.net.NoRouteToHostException: No route to host >> at org.apache.hadoop.ipc.Client.wrapException(Client.java:751) >> at org.apache.hadoop.ipc.Client.call(Client.java:719) >> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216) >> at org.apache.hadoop.dfs.$Proxy4.getProtocolVersion(Unknown Source) >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:348) >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:335) >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:372) >> at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:309) >> at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:286) >> at org.apache.hadoop.dfs.DataNode.startDataNode(DataNode.java:277) >> at org.apache.hadoop.dfs.DataNode.<init>(DataNode.java:223) >> at org.apache.hadoop.dfs.DataNode.makeInstance(DataNode.java:3071) >> at >> org.apache.hadoop.dfs.DataNode.instantiateDataNode(DataNode.java:3026) >> at org.apache.hadoop.dfs.DataNode.createDataNode(DataNode.java:3034) >> at org.apache.hadoop.dfs.DataNode.main(DataNode.java:3156) >> Caused by: java.net.NoRouteToHostException: No route to host >> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) >> at >> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:574) >> at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:100) >> at >> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:301) >> at >> org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:178) >> at org.apache.hadoop.ipc.Client.getConnection(Client.java:820) >> at org.apache.hadoop.ipc.Client.call(Client.java:705) >> ... 13 more >> >> 2009-06-04 12:16:41,222 INFO org.apache.hadoop.dfs.DataNode: SHUTDOWN_MSG: >> /************************************************************ >> SHUTDOWN_MSG: Shutting down DataNode at >> opencirrus-1262.hpl.hp.com/198.55.36.243 >> ************************************************************/ >> >> >> >> PLEASE COMMENT. >> >> Thanks. >> >> Asif. >> >> >> Ravi >> -- >> >> >