If you have a static address for the machine, make sure that your
hosts file is pointing to the static address for the namenode host
name as opposed to the 127.0.0.1 address. It should look something
like this with the values replaced with your values.

127.0.0.1               localhost.localdomain localhost
192.x.x.x               yourhost.yourdomain.com yourhost

- Edward

On Wed, Jul 23, 2008 at 6:03 AM, Jose Vidal <[EMAIL PROTECTED]> wrote:
> I'm trying to install hadoop on our linux machine but after
> start-all.sh none of the slaves can connect:
>
> 2008-07-22 16:35:27,534 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting DataNode
> STARTUP_MSG:   host = thetis/127.0.0.1
> STARTUP_MSG:   args = []
> STARTUP_MSG:   version = 0.16.4
> STARTUP_MSG:   build = 
> http://svn.apache.org/repos/asf/hadoop/core/branches/bran
> ch-0.16 -r 652614; compiled by 'hadoopqa' on Fri May  2 00:18:12 UTC 2008
> ************************************************************/
> 2008-07-22 16:35:27,643 WARN org.apache.hadoop.dfs.DataNode: Invalid 
> directory i
> n dfs.data.dir: directory is not writable: /work
> 2008-07-22 16:35:27,699 INFO org.apache.hadoop.ipc.Client: Retrying connect 
> to s
> erver: hermes.cse.sc.edu/129.252.130.148:9000. Already tried 1 time(s).
> 2008-07-22 16:35:28,700 INFO org.apache.hadoop.ipc.Client: Retrying connect 
> to s
> erver: hermes.cse.sc.edu/129.252.130.148:9000. Already tried 2 time(s).
> 2008-07-22 16:35:29,700 INFO org.apache.hadoop.ipc.Client: Retrying connect 
> to s
> erver: hermes.cse.sc.edu/129.252.130.148:9000. Already tried 3 time(s).
> 2008-07-22 16:35:30,701 INFO org.apache.hadoop.ipc.Client: Retrying connect 
> to s
> erver: hermes.cse.sc.edu/129.252.130.148:9000. Already tried 4 time(s).
> 2008-07-22 16:35:31,702 INFO org.apache.hadoop.ipc.Client: Retrying connect 
> to s
> erver: hermes.cse.sc.edu/129.252.130.148:9000. Already tried 5 time(s).
> 2008-07-22 16:35:32,702 INFO org.apache.hadoop.ipc.Client: Retrying connect 
> to s
> erver: hermes.cse.sc.edu/129.252.130.148:9000. Already tried 6 time(s).
>
> same for the tasktrackers (port 9001).
>
> I think the problem has something to do with name resolution. Check these out:
>
> [EMAIL PROTECTED]:~/hadoop-0.16.4> telnet hermes.cse.sc.edu 9000
> Trying 127.0.0.1...
> Connected to hermes.cse.sc.edu (127.0.0.1).
> Escape character is '^]'.
> bye
> Connection closed by foreign host.
>
> [EMAIL PROTECTED]:~/hadoop-0.16.4> host hermes.cse.sc.edu
> hermes.cse.sc.edu has address 129.252.130.148
>
> [EMAIL PROTECTED]:~/hadoop-0.16.4> telnet 129.252.130.148 9000
> Trying 129.252.130.148...
> telnet: connect to address 129.252.130.148: Connection refused
> telnet: Unable to connect to remote host: Connection refused
>
> So, the first one connects but not the second one, but they both go to
> the same machine:port. My guess is that the hadoop server is closing
> the connection, but why?
>
> Thanks,
> Jose
>
> --
> Jose M. Vidal <[EMAIL PROTECTED]> http://jmvidal.cse.sc.edu
> University of South Carolina http://www.multiagent.com
>



-- 
Best regards,
Edward J. Yoon,
http://blog.udanax.org

Reply via email to