That helps a lot actually. I will try setting up my hosts file tomorrow and make the other changes you suggested.
Thanks! Mithila Nagendra wrote: > > Hi, > The replication factor has to be set to 1. Also for you dfs and job > tracker > configuration you should insert the name of the node rather than the i.p > address. > > For instance: > <value>192.168.1.10:54310</value> > > can be: > > <value>master:54310</value> > > The nodes can be renamed by renaming them in the hosts files in /etc > folder. > It should look like the following: > > # Do not remove the following line, or various programs > # that require network functionality will fail. > 127.0.0.1 localhost.localdomain localhost node01 > 192.168.0.1 node01 > 192.168.0.2 node02 > 192.168.0.3 node03 > > Hope this helps > Mithila > > On Wed, Apr 15, 2009 at 9:40 PM, jpe30 <jpotte...@gmail.com> wrote: > >> >> I'm setting up a Hadoop cluster and I have the name node and job tracker >> up >> and running. However, I cannot get any of my datanodes or tasktrackers >> to >> start. Here is my hadoop-site.xml file... >> >> >> >> <?xml version="1.0"?> >> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >> >> <!-- Put site-specific property overrides in this file. --> >> >> <configuration> >> >> <property> >> <name>hadoop.tmp.dir</name> >> <value>/home/hadoop/h_temp</value> >> <description>A base for other temporary directories.</description> >> </property> >> >> <property> >> <name>dfs.data.dir</name> >> <value>/home/hadoop/data</value> >> </property> >> >> <property> >> <name>fs.default.name</name> >> <value>192.168.1.10:54310</value> >> <description>The name of the default file system. A URI whose >> scheme and authority determine the FileSystem implementation. The >> uri's scheme determines the config property (fs.SCHEME.impl) naming >> the FileSystem implementation class. The uri's authority is used to >> determine the host, port, etc. for a filesystem.</description> >> <final>true</final> >> </property> >> >> <property> >> <name>mapred.job.tracker</name> >> <value>192.168.1.10:54311</value> >> <description>The host and port that the MapReduce job tracker runs >> at. If "local", then jobs are run in-process as a single map >> and reduce task. >> </description> >> </property> >> >> <property> >> <name>dfs.replication</name> >> <value>0</value> >> <description>Default block replication. >> The actual number of replications can be specified when the file is >> created. >> The default is used if replication is not specified in create time. >> </description> >> </property> >> >> </configuration> >> >> >> and here is the error I'm getting... >> >> >> >> >> 2009-04-15 14:00:48,208 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG: >> /************************************************************ >> STARTUP_MSG: Starting DataNode >> STARTUP_MSG: host = java.net.UnknownHostException: myhost: myhost >> STARTUP_MSG: args = [] >> STARTUP_MSG: version = 0.18.3 >> STARTUP_MSG: build = >> https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18 -r >> 736250; >> compiled by 'ndaley' on Thu Jan 22 23:12:0$ >> ************************************************************/ >> 2009-04-15 14:00:48,355 ERROR org.apache.hadoop.dfs.DataNode: >> java.net.UnknownHostException: myhost: myhost >> at java.net.InetAddress.getLocalHost(InetAddress.java:1353) >> at org.apache.hadoop.net.DNS.getDefaultHost(DNS.java:185) >> at org.apache.hadoop.dfs.DataNode.startDataNode(DataNode.java:249) >> at org.apache.hadoop.dfs.DataNode.<init>(DataNode.java:223) >> at >> org.apache.hadoop.dfs.DataNode.makeInstance(DataNode.java:3071) >> at >> org.apache.hadoop.dfs.DataNode.instantiateDataNode(DataNode.java:3026) >> at >> org.apache.hadoop.dfs.DataNode.createDataNode(DataNode.java:3034) >> at org.apache.hadoop.dfs.DataNode.main(DataNode.java:3156) >> >> 2009-04-15 14:00:48,356 INFO org.apache.hadoop.dfs.DataNode: >> SHUTDOWN_MSG: >> /************************************************************ >> SHUTDOWN_MSG: Shutting down DataNode at java.net.UnknownHostException: >> myhost: myhost >> ************************************************************/ >> >> -- >> View this message in context: >> http://www.nabble.com/Datanode-Setup-tp23064660p23064660.html >> Sent from the Hadoop core-user mailing list archive at Nabble.com. >> >> > > -- View this message in context: http://www.nabble.com/Datanode-Setup-tp23064660p23065220.html Sent from the Hadoop core-user mailing list archive at Nabble.com.