Does 'ping lca2-s3-pc01' resolve from lca2-s3-pc04 and vise-versa? Are your 'slaves' and 'master' configuration files configured correctly? You can also try stopping everything, deleting all of your Hadoop data on each machine (by default in /tmp), reformating the namenode, and starting all again.
Alex On Tue, Nov 4, 2008 at 11:11 AM, <[EMAIL PROTECTED]> wrote: > Hi, > > I am trying to use hadoop 0.18.1. After I start the hadoop, I am able to > see namenode running on the master. But, datanode on the client machine is > unable to connect to the namenode. I use 2 machines with hostnames > lca2-s3-pc01 and lca2-s3-pc04 respectively. It shows the following message > in the client log file. > > 2008-11-04 17:19:25,253 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG: > /************************************************************ > STARTUP_MSG: Starting DataNode > STARTUP_MSG: host = lca2-s3-pc04/127.0.1.1 > STARTUP_MSG: args = [] > STARTUP_MSG: version = 0.18.1 > STARTUP_MSG: build = > http://svn.apache.org/repos/asf/hadoop/core/branches/bran > ch-0.18 -r 694836; compiled by 'hadoopqa' on Fri Sep 12 23:29:35 UTC 2008 > ************************************************************/ > 2008-11-04 17:19:26,464 INFO org.apache.hadoop.ipc.Client: Retrying connect > to s > erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 0 time(s). > 2008-11-04 17:19:27,468 INFO org.apache.hadoop.ipc.Client: Retrying connect > to s > erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 1 time(s). > 2008-11-04 17:19:28,472 INFO org.apache.hadoop.ipc.Client: Retrying connect > to s > erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 2 time(s). > 2008-11-04 17:19:29,476 INFO org.apache.hadoop.ipc.Client: Retrying connect > to s > erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 3 time(s). > 2008-11-04 17:19:30,479 INFO org.apache.hadoop.ipc.Client: Retrying connect > to s > erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 4 time(s). > 2008-11-04 17:19:31,483 INFO org.apache.hadoop.ipc.Client: Retrying connect > to s > erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 5 time(s). > 2008-11-04 17:19:32,487 INFO org.apache.hadoop.ipc.Client: Retrying connect > to s > erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 6 time(s). > 2008-11-04 17:19:33,491 INFO org.apache.hadoop.ipc.Client: Retrying connect > to s > erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 7 time(s). > 2008-11-04 17:19:34,495 INFO org.apache.hadoop.ipc.Client: Retrying connect > to s > erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 8 time(s). > 2008-11-04 17:19:35,499 INFO org.apache.hadoop.ipc.Client: Retrying connect > to s > erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 9 time(s). > 2008-11-04 17:19:35,502 ERROR org.apache.hadoop.dfs.DataNode: > java.io.IOExceptio > n: Call failed on local exception > at org.apache.hadoop.ipc.Client.call(Client.java:718) > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216) > 2008-11-at org.apache.hadoop.dfs.$Proxy4.getProtocolVersion(Unknown Source) > to s > at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:319) > at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:306) > at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:343) > at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:288) > at org.apache.hadoop.dfs.DataNode.startDataNode(DataNode.java:244) > at org.apache.hadoop.dfs.DataNode.<init>(DataNode.java:190) > at org.apache.hadoop.dfs.DataNode.makeInstance(DataNode.java:2987) > at > org.apache.hadoop.dfs.DataNode.instantiateDataNode(DataNode.java:2942 > ) > at org.apache.hadoop.dfs.DataNode.createDataNode(DataNode.java:2950) > at org.apache.hadoop.dfs.DataNode.main(DataNode.java:3072) > Caused by: java.net.ConnectException: Connection refused > at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > at > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:574 > ) > at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:100) > at > org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:30 > 0) > at > org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:177) > at org.apache.hadoop.ipc.Client.getConnection(Client.java:789) > at org.apache.hadoop.ipc.Client.call(Client.java:704) > ... 12 more > erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 5 time(s). > 2008-11-04 17:19:35,502 INFO org.apache.hadoop.dfs.DataNode: SHUTDOWN_MSG:t > to s > /************************************************************e(s). > SHUTDOWN_MSG: Shutting down DataNode at lca2-s3-pc04/127.0.1.1 > ************************************************************/haracters > > Here is the hadoop-site configuration file data that I use on both the > master and the client. > > <configuration> > > <property> > <name>hadoop.tmp.dir</name> > <value>/opt/okkam/datastore/hadoop</value> > </property> > > <property> > <name>fs.default.name</name> > <value>hdfs://lca2-s3-pc01:9000</value> > </property> > > <property> > <name>dfs.replication</name> > <value>2</value> > </property> > > </configuration> > > Could you please tell the mistake I am committing. > > Thanks a lot in advance, > Srikanth. >