Is it possible there is a firewall blocking port 9000 on one or more of the machines.

We had that happen to us with some machines that were kickstarted by our IT, the firewall was configured to only allow ssh.

[EMAIL PROTECTED] wrote:
Hi,

I am trying to use hadoop 0.18.1. After I start the hadoop, I am able to see namenode running on the master. But, datanode on the client machine is unable to connect to the namenode. I use 2 machines with hostnames lca2-s3-pc01 and lca2-s3-pc04 respectively. It shows the following message in the client log file.

2008-11-04 17:19:25,253 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting DataNode
STARTUP_MSG:   host = lca2-s3-pc04/127.0.1.1
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.18.1
STARTUP_MSG: build = http://svn.apache.org/repos/asf/hadoop/core/branches/bran
ch-0.18 -r 694836; compiled by 'hadoopqa' on Fri Sep 12 23:29:35 UTC 2008
************************************************************/
2008-11-04 17:19:26,464 INFO org.apache.hadoop.ipc.Client: Retrying connect to s
erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 0 time(s).
2008-11-04 17:19:27,468 INFO org.apache.hadoop.ipc.Client: Retrying connect to s
erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 1 time(s).
2008-11-04 17:19:28,472 INFO org.apache.hadoop.ipc.Client: Retrying connect to s
erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 2 time(s).
2008-11-04 17:19:29,476 INFO org.apache.hadoop.ipc.Client: Retrying connect to s
erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 3 time(s).
2008-11-04 17:19:30,479 INFO org.apache.hadoop.ipc.Client: Retrying connect to s
erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 4 time(s).
2008-11-04 17:19:31,483 INFO org.apache.hadoop.ipc.Client: Retrying connect to s
erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 5 time(s).
2008-11-04 17:19:32,487 INFO org.apache.hadoop.ipc.Client: Retrying connect to s
erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 6 time(s).
2008-11-04 17:19:33,491 INFO org.apache.hadoop.ipc.Client: Retrying connect to s
erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 7 time(s).
2008-11-04 17:19:34,495 INFO org.apache.hadoop.ipc.Client: Retrying connect to s
erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 8 time(s).
2008-11-04 17:19:35,499 INFO org.apache.hadoop.ipc.Client: Retrying connect to s
erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 9 time(s).
2008-11-04 17:19:35,502 ERROR org.apache.hadoop.dfs.DataNode: java.io.IOExceptio
n: Call failed on local exception
        at org.apache.hadoop.ipc.Client.call(Client.java:718)
        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)
2008-11-at org.apache.hadoop.dfs.$Proxy4.getProtocolVersion(Unknown Source) to s
        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:319)
        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:306)
        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:343)
        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:288)
at org.apache.hadoop.dfs.DataNode.startDataNode(DataNode.java:244)
        at org.apache.hadoop.dfs.DataNode.<init>(DataNode.java:190)
at org.apache.hadoop.dfs.DataNode.makeInstance(DataNode.java:2987) at org.apache.hadoop.dfs.DataNode.instantiateDataNode(DataNode.java:2942
)
at org.apache.hadoop.dfs.DataNode.createDataNode(DataNode.java:2950)
        at org.apache.hadoop.dfs.DataNode.main(DataNode.java:3072)
Caused by: java.net.ConnectException: Connection refused
        at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:574
)
        at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:100)
at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:30
0)
at org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:177)
        at org.apache.hadoop.ipc.Client.getConnection(Client.java:789)
        at org.apache.hadoop.ipc.Client.call(Client.java:704)
        ... 12 more
erver: lca2-s3-pc01/128.178.156.221:9000. Already tried 5 time(s).
2008-11-04 17:19:35,502 INFO org.apache.hadoop.dfs.DataNode: SHUTDOWN_MSG:t to s
/************************************************************e(s).
SHUTDOWN_MSG: Shutting down DataNode at lca2-s3-pc04/127.0.1.1
************************************************************/haracters

Here is the hadoop-site configuration file data that I use on both the master and the client.

<configuration>

<property>
<name>hadoop.tmp.dir</name>
<value>/opt/okkam/datastore/hadoop</value>
</property>

<property>
<name>fs.default.name</name>
<value>hdfs://lca2-s3-pc01:9000</value>
</property>

<property>
<name>dfs.replication</name>
<value>2</value>
</property>

</configuration>

Could you please tell the mistake I am committing.

Thanks a lot in advance,
Srikanth.

Reply via email to