Make sure your hosts file on your namenode is setup correctly: 127.0.0.1 localhost.localdomain localhost 10.x.x.x myhostname.mydomain.com myhostname
As opposed to: 127.0.0.1 localhost.localdomain localhost myhostname.mydomain.com myhostname The problem may be that the machine is listening on only the local interface. If you do a ping myhostname from the local box you should receive the real IP and not the loopback address. Let me know if this was the problem or if you need more help. Dennis Kubes cybercouf wrote: > I'm trying to setup hadoop using these guides: > http://wiki.apache.org/nutch/NutchHadoopTutorial and > http://www.nabble.com/Nutch-Step-by-Step-Maybe-someone-will-find-this-useful---tf3526281.html > > But i'm stuck at the early step: having a single machine running. > using nutch 0.8.1 and so the provided hadoop "hadoop-0.4.0-patched.jar" > JVM sun 1.5.0_11 > > When I start the namenode (using ./bin/start-all.sh) I have this in the > namenode-log: > > 2007-05-02 12:39:51,335 INFO util.Credential - Checking Resource aliases > 2007-05-02 12:39:51,349 INFO http.HttpServer - Version Jetty/5.1.4 > 2007-05-02 12:39:51,350 WARN servlet.WebApplicationContext - Web > application not found > /home/nutch/search/file:/home/nutch/search/lib/hadoop-0.4.0-patched.jar!/webapps/dfs > 2007-05-02 12:39:51,351 WARN servlet.WebApplicationContext - Configuration > error on > /home/nutch/search/file:/home/nutch/search/lib/hadoop-0.4.0-patched.jar!/webapps/dfs > java.io.FileNotFoundException: > /home/nutch/search/file:/home/nutch/search/lib/hadoop-0.4.0-patched.jar!/webapps/dfs > at > org.mortbay.jetty.servlet.WebApplicationContext.resolveWebApp(WebApplicationContext.java:266) > at > org.mortbay.jetty.servlet.WebApplicationContext.doStart(WebApplicationContext.java:449) > at org.mortbay.util.Container.start(Container.java:72) > at org.mortbay.http.HttpServer.doStart(HttpServer.java:753) > at org.mortbay.util.Container.start(Container.java:72) > at > org.apache.hadoop.mapred.StatusHttpServer.start(StatusHttpServer.java:138) > at org.apache.hadoop.dfs.FSNamesystem.<init>(FSNamesystem.java:173) > at org.apache.hadoop.dfs.NameNode.<init>(NameNode.java:91) > at org.apache.hadoop.dfs.NameNode.<init>(NameNode.java:82) > at org.apache.hadoop.dfs.NameNode.main(NameNode.java:491) > 2007-05-02 12:39:51,353 INFO util.Container - Started > HttpContext[/logs,/logs] > 2007-05-02 12:39:51,353 INFO util.Container - Started > HttpContext[/static,/static] > 2007-05-02 12:39:51,357 INFO http.SocketListener - Started SocketListener > on 0.0.0.0:50070 > > and after I can't access it: > $ ./bin/hadoop dfs -ls > ls: Connection refused > > hadoop.log: > 2007-05-02 12:41:40,030 WARN fs.DFSClient - Problem renewing lease for > DFSClient_2015604182: java.net.ConnectException: Connection refused > at java.net.PlainSocketImpl.socketConnect(Native Method) > [...] > > > > 1. I can't understant why there is this FileNotFound execption, I didn't > change anything in the hadoop nutch jar file. > > 2. It looks like the namenode is running (when I stop it I have the message > "stopping namenode"), but why I can't access it ? (is this ip from the log > correct? 0.0.0.0:50070) > all is on the same machine, and my conf file looks ok: > fs.default.name myhostname:9000 > mapred.job.tracker myhostname:9001 > mapred.map.tasks 2 > mapred.reduce.tasks 2 > dfs.name.dir /home/nutch/filesystem/name > dfs.data.dir /home/nutch/filesystem/data > mapred.system.dir /home/nutch/filesystem/mapreduce/system > mapred.local.dir /home/nutch/filesystem/mapreduce/local > dfs.replication 1 ------------------------------------------------------------------------- This SF.net email is sponsored by DB2 Express Download DB2 Express C - the FREE version of DB2 express and take control of your XML. No limits. Just data. Click to get it now. http://sourceforge.net/powerbar/db2/ _______________________________________________ Nutch-general mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/nutch-general
