You may also need to set mapred.job.tracker in mapred-site.xml
-seunghwa On Thu, 2009-06-18 at 10:42 -0700, llpind wrote: > Hey All, > > I'm able to start my master server, but none of the slave nodes come up > (unless I list the master as the slave). After searching a bit, seems > people have this problem when they forget to set df.default.name, but i've > got it set in core-site.xml (listed below). They all have the error below > on start up: > > STARTUP_MSG: Starting DataNode > STARTUP_MSG: host = slave1/192.168.0.234 > STARTUP_MSG: args = [] > STARTUP_MSG: version = 0.20.0 > STARTUP_MSG: build = > https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.20 -r 763504; > compiled by 'ndaley' on Thu Apr 9 05:18:40 UTC 2009 > ************************************************************/ > 2009-06-18 09:06:49,369 ERROR > org.apache.hadoop.hdfs.server.datanode.DataNode: > java.lang.NullPointerException > at > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:134) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:156) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:160) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:246) > > at > org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:216) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1283) > > at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1238) > > at > org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1246) > > at > org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1368) > > 2009-06-18 09:06:49,370 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG: > /************************************************************ > SHUTDOWN_MSG: Shutting down DataNode at slave1/192.168.0.234 > ************************************************************/ > > ============================================== > core-site.xml > ============================================== > > <property> > <name>fs.default.name</name> > <value>hdfs://master:54310</value> > <description>The name of the default file system. A URI whose > scheme and authority determine the FileSystem implementation. The > uri's scheme determines the config property (fs.SCHEME.impl) naming > the FileSystem implementation class. The uri's authority is used to > determine the host, port, etc. for a filesystem.</description> > </property> > <property> > <name>hadoop.tmp.dir</name> > <value>/data/hadoop-0.20.0-${user.name}</value> > <description>A base for other temporary directories.</description> > </property>