> > I finally managed to setup and run Hama in fully distributed mode (thanks a > lot to Thomas Jungblut!) >
No problem, that's my "job" ;)). That is great. Have fun! 2011/9/19 Luis Eduardo Pineda Morales <[email protected]> > Hi all! > > I finally managed to setup and run Hama in fully distributed mode (thanks a > lot to Thomas Jungblut!) > > I'm using Hama 0.3.0 and Hadoop 0.20.2 with IPv4 as in > http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/ > > Same settings didn't work with Hadoop 0.20.203 (said to be the most recent > stable version). > Hope these settings are useful for you. > > Luis > > > On 15 Sep 2011, at 19:25, Thomas Jungblut wrote: > > Hey, I'm sorry, the IPv6 was misleading. > On your screenshot I see that you are using an Append version of Hadoop. > Did you try it with 0.20.2? > > 2011/9/15 Luis Eduardo Pineda Morales <[email protected]> > >> Hi Thomas, apparently IPv6 wasn't the problem, since now hadoop is running >> in IPv4 and i still get the same exceptions in hama. >> >> pineda@server00:~/hadoop$ jps >> 10592 NameNode >> 10922 Jps >> 10695 DataNode >> 10844 SecondaryNameNode >> >> pineda@server00:~/hadoop$ lsof -i >> COMMAND PID USER FD TYPE DEVICE SIZE NODE NAME >> java 10592 pineda 46u IPv4 2559447 TCP *:50272 (LISTEN) >> java 10592 pineda 56u IPv4 2559684 TCP server00:54310 (LISTEN) >> java 10592 pineda 67u IPv4 2559694 TCP *:50070 (LISTEN) >> java 10592 pineda 71u IPv4 2559771 TCP >> server00:54310->server00:51666 (ESTABLISHED) >> java 10592 pineda 72u IPv4 2559810 TCP >> server00:51668->server00:54310 (ESTABLISHED) >> java 10592 pineda 73u IPv4 2559811 TCP >> server00:54310->server00:51668 (ESTABLISHED) >> java 10592 pineda 77u IPv4 2560218 TCP >> server00:54310->server00:51671 (ESTABLISHED) >> java 10695 pineda 46u IPv4 2559682 TCP *:44935 (LISTEN) >> java 10695 pineda 52u IPv4 2559764 TCP >> server00:51666->server00:54310 (ESTABLISHED) >> java 10695 pineda 60u IPv4 2559892 TCP *:50010 (LISTEN) >> java 10695 pineda 61u IPv4 2559899 TCP *:50075 (LISTEN) >> java 10695 pineda 66u IPv4 2560208 TCP *:50020 (LISTEN) >> java 10844 pineda 46u IPv4 2560204 TCP *:41188 (LISTEN) >> java 10844 pineda 52u IPv4 2560217 TCP >> server00:51671->server00:54310 (ESTABLISHED) >> java 10844 pineda 59u IPv4 2560225 TCP *:50090 (LISTEN) >> >> >> also the web interface doesn't show any errors: and I'm able to run >> hadoop shell commands. Any other idea? :-/ >> >> Luis >> >> >> >> >> On 15 Sep 2011, at 18:17, Thomas Jungblut wrote: >> >> > Hi Luis, >> > >> > it doesn't mean that it is working, just because there is no exception. >> > Thanks that you appended your lsof output, because Hadoop does not >> support >> > IPv6. >> > >> > Please setup Hadoop correctly [1] and then use Hama. >> > For example here is my lsof -i output: >> > >> > hadoop@raynor:/home/thomasjungblut$ lsof -i >> >> COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME >> >> java 1144 hadoop 33u IPv4 8819 0t0 TCP *:49737 (LISTEN) >> >> java 1144 hadoop 37u IPv4 9001 0t0 TCP raynor:9001 >> (LISTEN) >> >> java 1144 hadoop 47u IPv4 9222 0t0 TCP *:50070 (LISTEN) >> >> java 1144 hadoop 52u IPv4 9429 0t0 TCP >> >> raynor:9001->findlay:35283 (ESTABLISHED) >> >> java 1144 hadoop 53u IPv4 9431 0t0 TCP >> >> raynor:9001->karrigan:57345 (ESTABLISHED) >> >> java 1249 hadoop 33u IPv4 8954 0t0 TCP *:54235 (LISTEN) >> >> java 1249 hadoop 44u IPv4 9422 0t0 TCP *:50010 (LISTEN) >> >> java 1249 hadoop 45u IPv4 9426 0t0 TCP *:50075 (LISTEN) >> >> >> > >> > There are two ways to determine if Hadoop is setup correctly: >> > >> > 1. Look at the Webinterface of the Namenode [2] and see that there is >> no >> > Safemode message or datanode missing. >> > 2. Or run a sample MapReduce Job, for example WordCount [3]. >> > >> > If Hama is not working afterwards, you ask your next question again. >> > >> > Thanks and good luck :) >> > >> > [1] >> > >> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/ >> > [2] >> > >> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/#hdfs-name-node-web-interface >> > [3] >> > >> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/#run-the-mapreduce-job >> > >> > >> > 2011/9/15 Luis Eduardo Pineda Morales <[email protected]> >> > >> >> Hi all, >> >> >> >> I am attempting to run the distributed mode. I have HDFS running in a >> >> single machine (pseudo-distributed mode): >> >> >> >> pineda@server00:~/hadoop$ jps >> >> 472 SecondaryNameNode >> >> 1429 Jps >> >> 32733 NameNode >> >> 364 DataNode >> >> >> >> pineda@server00:~/hadoop$ lsof -i >> >> COMMAND PID USER FD TYPE DEVICE SIZE NODE NAME >> >> java 364 pineda 46u IPv6 2532945 TCP *:41462 (LISTEN) >> >> java 364 pineda 52u IPv6 2533275 TCP >> >> server00:42445->server00:54310 (ESTABLISHED) >> >> java 364 pineda 60u IPv6 2533307 TCP *:50010 (LISTEN) >> >> java 364 pineda 61u IPv6 2533511 TCP *:50075 (LISTEN) >> >> java 364 pineda 66u IPv6 2533518 TCP *:50020 (LISTEN) >> >> java 472 pineda 46u IPv6 2533286 TCP *:43098 (LISTEN) >> >> java 472 pineda 59u IPv6 2533536 TCP *:50090 (LISTEN) >> >> java 32733 pineda 46u IPv6 2532751 TCP *:54763 (LISTEN) >> >> java 32733 pineda 56u IPv6 2533062 TCP server00:54310 >> (LISTEN) >> >> java 32733 pineda 67u IPv6 2533081 TCP *:50070 (LISTEN) >> >> java 32733 pineda 76u IPv6 2533276 TCP >> >> server00:54310->server00:42445 (ESTABLISHED) >> >> >> >> i.e. fs.defaul.name = hdfs://server00:54310/ >> >> >> >> then I run hama in server04 (groom in server03, zookeeper in server05): >> >> >> >> pineda@server04:~/hama$ bin/start-bspd.sh >> >> server05: starting zookeeper, logging to >> >> /logs/hama-pineda-zookeeper-server05.out >> >> starting bspmaster, logging to /logs/hama-pineda-bspmaster-server04.out >> >> 2011-09-15 17:08:43.349:INFO::Logging to STDERR via >> >> org.mortbay.log.StdErrLog >> >> 2011-09-15 17:08:43.409:INFO::jetty-0.3.0-incubating >> >> server03: starting groom, logging to >> /logs/hama-pineda-groom-server03.out >> >> >> >> this is my hama-site.xml file: >> >> >> >> <configuration> >> >> <property> >> >> <name>bsp.master.address</name> >> >> <value>server04</value> >> >> </property> >> >> >> >> <property> >> >> <name>fs.default.name</name> >> >> <value>hdfs://server00:54310</value> >> >> </property> >> >> >> >> <property> >> >> <name>hama.zookeeper.quorum</name> >> >> <value>server05</value> >> >> </property> >> >> </configuration> >> >> >> >> >> >> In theory I can connect to the HDFS, because I don't get any >> >> ConnectException, but Hama doesn't run, and I get this Exception trace >> in my >> >> bspmaster.log after the Jetty is bound: >> >> >> >> >> >> 2011-09-15 17:08:43,409 INFO org.apache.hama.http.HttpServer: Jetty >> bound >> >> to port 40013 >> >> 2011-09-15 17:08:44,070 INFO org.apache.hama.bsp.BSPMaster: problem >> >> cleaning system directory: null >> >> java.io.IOException: Call to server00/192.168.122.10:54310 failed on >> local >> >> exception: java.io.EOFException >> >> at org.apache.hadoop.ipc.Client.wrapException(Client.java:775) >> >> at org.apache.hadoop.ipc.Client.call(Client.java:743) >> >> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) >> >> at $Proxy4.getProtocolVersion(Unknown Source) >> >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359) >> >> at >> >> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:106) >> >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:207) >> >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:170) >> >> at >> >> >> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82) >> >> at >> >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378) >> >> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) >> >> at >> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390) >> >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196) >> >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95) >> >> at org.apache.hama.bsp.BSPMaster.<init>(BSPMaster.java:263) >> >> at org.apache.hama.bsp.BSPMaster.startMaster(BSPMaster.java:421) >> >> at org.apache.hama.bsp.BSPMaster.startMaster(BSPMaster.java:415) >> >> at org.apache.hama.BSPMasterRunner.run(BSPMasterRunner.java:46) >> >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) >> >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) >> >> at org.apache.hama.BSPMasterRunner.main(BSPMasterRunner.java:56) >> >> Caused by: java.io.EOFException >> >> at java.io.DataInputStream.readInt(DataInputStream.java:375) >> >> at >> >> >> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:501) >> >> at org.apache.hadoop.ipc.Client$Connection.run(Client.java:446) >> >> >> >> >> >> Do you know how to fix this? Do you know what is the directory that it >> is >> >> trying to clean? >> >> >> >> Any idea is welcomed! >> >> >> >> Thanks, >> >> Luis. >> > >> > >> > >> > >> > -- >> > Thomas Jungblut >> > Berlin >> > >> > mobile: 0170-3081070 >> > >> > business: [email protected] >> > private: [email protected] >> >> >> > > > -- > Thomas Jungblut > Berlin > > mobile: 0170-3081070 > > business: [email protected] > private: [email protected] > > > -- Thomas Jungblut Berlin mobile: 0170-3081070 business: [email protected] private: [email protected]
