Yes it is. Here more details:
$ cat /etc/hosts 127.0.0.1 localhost 84.x.x.x myhostname.mydomain.com myhostname # ping myhostname PING myhostname.mydomain.com (84.x.x.x) 56(84) bytes of data. 64 bytes from myhostname.mydomain.com (84.x.x.x): icmp_seq=1 ttl=64 time=0.017 ms and when i do start-all.sh, the namenode seems running: # netstat -tupl Active Internet connections (only servers) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name tcp6 0 0 *:50070 *:* LISTEN 18241/java tcp6 0 0 *:ssh *:* LISTEN 3350/sshd tcp6 0 0 *:50010 *:* LISTEN 18279/java and also I noticed that my nutch user (from who i launch all the scripts) is not allowed to ping nutch:~/search$ ping myhostname ping: icmp open socket: Operation not permitted but that shouldn't be linked to the fact to failed to open a java socket? (java.net.ConnectException: Connection refused) thanks for your help! Dennis Kubes wrote: > > Make sure your hosts file on your namenode is setup correctly: > > 127.0.0.1 localhost.localdomain localhost > 10.x.x.x myhostname.mydomain.com myhostname > > As opposed to: > > 127.0.0.1 localhost.localdomain localhost > myhostname.mydomain.com myhostname > > The problem may be that the machine is listening on only the local > interface. If you do a ping myhostname from the local box you should > receive the real IP and not the loopback address. > > Let me know if this was the problem or if you need more help. > > Dennis Kubes > > cybercouf wrote: >> I'm trying to setup hadoop using these guides: >> http://wiki.apache.org/nutch/NutchHadoopTutorial and >> http://www.nabble.com/Nutch-Step-by-Step-Maybe-someone-will-find-this-useful---tf3526281.html >> >> But i'm stuck at the early step: having a single machine running. >> using nutch 0.8.1 and so the provided hadoop "hadoop-0.4.0-patched.jar" >> JVM sun 1.5.0_11 >> >> When I start the namenode (using ./bin/start-all.sh) I have this in the >> namenode-log: >> >> 2007-05-02 12:39:51,335 INFO util.Credential - Checking Resource aliases >> 2007-05-02 12:39:51,349 INFO http.HttpServer - Version Jetty/5.1.4 >> 2007-05-02 12:39:51,350 WARN servlet.WebApplicationContext - Web >> application not found >> /home/nutch/search/file:/home/nutch/search/lib/hadoop-0.4.0-patched.jar!/webapps/dfs >> 2007-05-02 12:39:51,351 WARN servlet.WebApplicationContext - >> Configuration >> error on >> /home/nutch/search/file:/home/nutch/search/lib/hadoop-0.4.0-patched.jar!/webapps/dfs >> java.io.FileNotFoundException: >> /home/nutch/search/file:/home/nutch/search/lib/hadoop-0.4.0-patched.jar!/webapps/dfs >> at >> org.mortbay.jetty.servlet.WebApplicationContext.resolveWebApp(WebApplicationContext.java:266) >> at >> org.mortbay.jetty.servlet.WebApplicationContext.doStart(WebApplicationContext.java:449) >> at org.mortbay.util.Container.start(Container.java:72) >> at org.mortbay.http.HttpServer.doStart(HttpServer.java:753) >> at org.mortbay.util.Container.start(Container.java:72) >> at >> org.apache.hadoop.mapred.StatusHttpServer.start(StatusHttpServer.java:138) >> at org.apache.hadoop.dfs.FSNamesystem.<init>(FSNamesystem.java:173) >> at org.apache.hadoop.dfs.NameNode.<init>(NameNode.java:91) >> at org.apache.hadoop.dfs.NameNode.<init>(NameNode.java:82) >> at org.apache.hadoop.dfs.NameNode.main(NameNode.java:491) >> 2007-05-02 12:39:51,353 INFO util.Container - Started >> HttpContext[/logs,/logs] >> 2007-05-02 12:39:51,353 INFO util.Container - Started >> HttpContext[/static,/static] >> 2007-05-02 12:39:51,357 INFO http.SocketListener - Started >> SocketListener >> on 0.0.0.0:50070 >> >> and after I can't access it: >> $ ./bin/hadoop dfs -ls >> ls: Connection refused >> >> hadoop.log: >> 2007-05-02 12:41:40,030 WARN fs.DFSClient - Problem renewing lease for >> DFSClient_2015604182: java.net.ConnectException: Connection refused >> at java.net.PlainSocketImpl.socketConnect(Native Method) >> [...] >> >> >> >> 1. I can't understant why there is this FileNotFound execption, I didn't >> change anything in the hadoop nutch jar file. >> >> 2. It looks like the namenode is running (when I stop it I have the >> message >> "stopping namenode"), but why I can't access it ? (is this ip from the >> log >> correct? 0.0.0.0:50070) >> all is on the same machine, and my conf file looks ok: >> fs.default.name myhostname:9000 >> mapred.job.tracker myhostname:9001 >> mapred.map.tasks 2 >> mapred.reduce.tasks 2 >> dfs.name.dir /home/nutch/filesystem/name >> dfs.data.dir /home/nutch/filesystem/data >> mapred.system.dir /home/nutch/filesystem/mapreduce/system >> mapred.local.dir /home/nutch/filesystem/mapreduce/local >> dfs.replication 1 > > -- View this message in context: http://www.nabble.com/nutch-and-hadoop%3A-can%27t-launch-properly-the-name-node-tf3680311.html#a10286866 Sent from the Nutch - User mailing list archive at Nabble.com. ------------------------------------------------------------------------- This SF.net email is sponsored by DB2 Express Download DB2 Express C - the FREE version of DB2 express and take control of your XML. No limits. Just data. Click to get it now. http://sourceforge.net/powerbar/db2/ _______________________________________________ Nutch-general mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/nutch-general
