i am following the instructions to setup a multi-node cluster at
http://hadoop.apache.org/docs/r2.4.1/hadoop-project-dist/hadoop-common/ClusterSetup.html
.

my problem is that when i run the script to start up the slave datanodes,
no slave datanode is started (more on this later).

i have two nodes so far that i am experimenting with
1. node1 (this is the namenode)
2. node2 (this is the datanode)

on node1 (namenode), i start the namenode daemon as follows. there is no
problem here.

$HADOOP_PREFIX/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR --script
hdfs start namenode

on node1 (namenode), i start the datanode daemon (on node2) as follows.

$HADOOP_PREFIX/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR --script
hdfs start datanode

it is at this point that the problems begin:
a) the datanode daemon is started on node1 (instead of node2)
b) the datanode daemon is not started on node2 (the only slave defined)
c) the only job of node1 is to be a namenode, not also a datanode

my $HADOOP_CONF_DIR/slaves has one single entry (line):
node2

my /etc/hosts file looks like the following
127.0.0.1 localhost 192.168.0.10 node1
192.168.0.11 node2

why does the script to start the datanode slave daemons not start them and
only start a local slave daemon?

on node2, when i run the script to start the datanode daemon, it is able to
find node1:8020 and become active as part of the cluster. but this is
strange to me, because this script should be called/executed from the
namenode to start all datanodes on the slaves (not the other way around).

any idea what i am doing wrong?

Reply via email to