Hi Thomas, Could you provide a concrete instance to illustrate the advantage of HAMA, when HAMA vs. MapReduce?
For example,SSSP on HAMA vs. SSSP on MapReduce. So ,I can catch the idea of HAMA quickly. Thank you very much! Changguanghui -----邮件原件----- 发件人: Thomas Jungblut [mailto:[email protected]] 发送时间: 2011年9月19日 23:17 收件人: Luis Eduardo Pineda Morales 抄送: [email protected] 主题: Re: Hama help (how the distributed mode is working) > > I finally managed to setup and run Hama in fully distributed mode (thanks a > lot to Thomas Jungblut!) > No problem, that's my "job" ;)). That is great. Have fun! 2011/9/19 Luis Eduardo Pineda Morales <[email protected]> > Hi all! > > I finally managed to setup and run Hama in fully distributed mode (thanks a > lot to Thomas Jungblut!) > > I'm using Hama 0.3.0 and Hadoop 0.20.2 with IPv4 as in > http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/ > > Same settings didn't work with Hadoop 0.20.203 (said to be the most recent > stable version). > Hope these settings are useful for you. > > Luis > > > On 15 Sep 2011, at 19:25, Thomas Jungblut wrote: > > Hey, I'm sorry, the IPv6 was misleading. > On your screenshot I see that you are using an Append version of Hadoop. > Did you try it with 0.20.2? > > 2011/9/15 Luis Eduardo Pineda Morales <[email protected]> > >> Hi Thomas, apparently IPv6 wasn't the problem, since now hadoop is running >> in IPv4 and i still get the same exceptions in hama. >> >> pineda@server00:~/hadoop$ jps >> 10592 NameNode >> 10922 Jps >> 10695 DataNode >> 10844 SecondaryNameNode >> >> pineda@server00:~/hadoop$ lsof -i >> COMMAND PID USER FD TYPE DEVICE SIZE NODE NAME >> java 10592 pineda 46u IPv4 2559447 TCP *:50272 (LISTEN) >> java 10592 pineda 56u IPv4 2559684 TCP server00:54310 (LISTEN) >> java 10592 pineda 67u IPv4 2559694 TCP *:50070 (LISTEN) >> java 10592 pineda 71u IPv4 2559771 TCP >> server00:54310->server00:51666 (ESTABLISHED) >> java 10592 pineda 72u IPv4 2559810 TCP >> server00:51668->server00:54310 (ESTABLISHED) >> java 10592 pineda 73u IPv4 2559811 TCP >> server00:54310->server00:51668 (ESTABLISHED) >> java 10592 pineda 77u IPv4 2560218 TCP >> server00:54310->server00:51671 (ESTABLISHED) >> java 10695 pineda 46u IPv4 2559682 TCP *:44935 (LISTEN) >> java 10695 pineda 52u IPv4 2559764 TCP >> server00:51666->server00:54310 (ESTABLISHED) >> java 10695 pineda 60u IPv4 2559892 TCP *:50010 (LISTEN) >> java 10695 pineda 61u IPv4 2559899 TCP *:50075 (LISTEN) >> java 10695 pineda 66u IPv4 2560208 TCP *:50020 (LISTEN) >> java 10844 pineda 46u IPv4 2560204 TCP *:41188 (LISTEN) >> java 10844 pineda 52u IPv4 2560217 TCP >> server00:51671->server00:54310 (ESTABLISHED) >> java 10844 pineda 59u IPv4 2560225 TCP *:50090 (LISTEN) >> >> >> also the web interface doesn't show any errors: and I'm able to run >> hadoop shell commands. Any other idea? :-/ >> >> Luis >> >> >> >> >> On 15 Sep 2011, at 18:17, Thomas Jungblut wrote: >> >> > Hi Luis, >> > >> > it doesn't mean that it is working, just because there is no exception. >> > Thanks that you appended your lsof output, because Hadoop does not >> support >> > IPv6. >> > >> > Please setup Hadoop correctly [1] and then use Hama. >> > For example here is my lsof -i output: >> > >> > hadoop@raynor:/home/thomasjungblut$ lsof -i >> >> COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME >> >> java 1144 hadoop 33u IPv4 8819 0t0 TCP *:49737 (LISTEN) >> >> java 1144 hadoop 37u IPv4 9001 0t0 TCP raynor:9001 >> (LISTEN) >> >> java 1144 hadoop 47u IPv4 9222 0t0 TCP *:50070 (LISTEN) >> >> java 1144 hadoop 52u IPv4 9429 0t0 TCP >> >> raynor:9001->findlay:35283 (ESTABLISHED) >> >> java 1144 hadoop 53u IPv4 9431 0t0 TCP >> >> raynor:9001->karrigan:57345 (ESTABLISHED) >> >> java 1249 hadoop 33u IPv4 8954 0t0 TCP *:54235 (LISTEN) >> >> java 1249 hadoop 44u IPv4 9422 0t0 TCP *:50010 (LISTEN) >> >> java 1249 hadoop 45u IPv4 9426 0t0 TCP *:50075 (LISTEN) >> >> >> > >> > There are two ways to determine if Hadoop is setup correctly: >> > >> > 1. Look at the Webinterface of the Namenode [2] and see that there is >> no >> > Safemode message or datanode missing. >> > 2. Or run a sample MapReduce Job, for example WordCount [3]. >> > >> > If Hama is not working afterwards, you ask your next question again. >> > >> > Thanks and good luck :) >> > >> > [1] >> > >> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/ >> > [2] >> > >> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/#hdfs-name-node-web-interface >> > [3] >> > >> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/#run-the-mapreduce-job >> > >> > >> > 2011/9/15 Luis Eduardo Pineda Morales <[email protected]> >> > >> >> Hi all, >> >> >> >> I am attempting to run the distributed mode. I have HDFS running in a >> >> single machine (pseudo-distributed mode): >> >> >> >> pineda@server00:~/hadoop$ jps >> >> 472 SecondaryNameNode >> >> 1429 Jps >> >> 32733 NameNode >> >> 364 DataNode >> >> >> >> pineda@server00:~/hadoop$ lsof -i >> >> COMMAND PID USER FD TYPE DEVICE SIZE NODE NAME >> >> java 364 pineda 46u IPv6 2532945 TCP *:41462 (LISTEN) >> >> java 364 pineda 52u IPv6 2533275 TCP >> >> server00:42445->server00:54310 (ESTABLISHED) >> >> java 364 pineda 60u IPv6 2533307 TCP *:50010 (LISTEN) >> >> java 364 pineda 61u IPv6 2533511 TCP *:50075 (LISTEN) >> >> java 364 pineda 66u IPv6 2533518 TCP *:50020 (LISTEN) >> >> java 472 pineda 46u IPv6 2533286 TCP *:43098 (LISTEN) >> >> java 472 pineda 59u IPv6 2533536 TCP *:50090 (LISTEN) >> >> java 32733 pineda 46u IPv6 2532751 TCP *:54763 (LISTEN) >> >> java 32733 pineda 56u IPv6 2533062 TCP server00:54310 >> (LISTEN) >> >> java 32733 pineda 67u IPv6 2533081 TCP *:50070 (LISTEN) >> >> java 32733 pineda 76u IPv6 2533276 TCP >> >> server00:54310->server00:42445 (ESTABLISHED) >> >> >> >> i.e. fs.defaul.name = hdfs://server00:54310/ >> >> >> >> then I run hama in server04 (groom in server03, zookeeper in server05): >> >> >> >> pineda@server04:~/hama$ bin/start-bspd.sh >> >> server05: starting zookeeper, logging to >> >> /logs/hama-pineda-zookeeper-server05.out >> >> starting bspmaster, logging to /logs/hama-pineda-bspmaster-server04.out >> >> 2011-09-15 17:08:43.349:INFO::Logging to STDERR via >> >> org.mortbay.log.StdErrLog >> >> 2011-09-15 17:08:43.409:INFO::jetty-0.3.0-incubating >> >> server03: starting groom, logging to >> /logs/hama-pineda-groom-server03.out >> >> >> >> this is my hama-site.xml file: >> >> >> >> <configuration> >> >> <property> >> >> <name>bsp.master.address</name> >> >> <value>server04</value> >> >> </property> >> >> >> >> <property> >> >> <name>fs.default.name</name> >> >> <value>hdfs://server00:54310</value> >> >> </property> >> >> >> >> <property> >> >> <name>hama.zookeeper.quorum</name> >> >> <value>server05</value> >> >> </property> >> >> </configuration> >> >> >> >> >> >> In theory I can connect to the HDFS, because I don't get any >> >> ConnectException, but Hama doesn't run, and I get this Exception trace >> in my >> >> bspmaster.log after the Jetty is bound: >> >> >> >> >> >> 2011-09-15 17:08:43,409 INFO org.apache.hama.http.HttpServer: Jetty >> bound >> >> to port 40013 >> >> 2011-09-15 17:08:44,070 INFO org.apache.hama.bsp.BSPMaster: problem >> >> cleaning system directory: null >> >> java.io.IOException: Call to server00/192.168.122.10:54310 failed on >> local >> >> exception: java.io.EOFException >> >> at org.apache.hadoop.ipc.Client.wrapException(Client.java:775) >> >> at org.apache.hadoop.ipc.Client.call(Client.java:743) >> >> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) >> >> at $Proxy4.getProtocolVersion(Unknown Source) >> >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359) >> >> at >> >> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:106) >> >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:207) >> >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:170) >> >> at >> >> >> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82) >> >> at >> >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378) >> >> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) >> >> at >> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390) >> >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196) >> >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95) >> >> at org.apache.hama.bsp.BSPMaster.<init>(BSPMaster.java:263) >> >> at org.apache.hama.bsp.BSPMaster.startMaster(BSPMaster.java:421) >> >> at org.apache.hama.bsp.BSPMaster.startMaster(BSPMaster.java:415) >> >> at org.apache.hama.BSPMasterRunner.run(BSPMasterRunner.java:46) >> >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) >> >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) >> >> at org.apache.hama.BSPMasterRunner.main(BSPMasterRunner.java:56) >> >> Caused by: java.io.EOFException >> >> at java.io.DataInputStream.readInt(DataInputStream.java:375) >> >> at >> >> >> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:501) >> >> at org.apache.hadoop.ipc.Client$Connection.run(Client.java:446) >> >> >> >> >> >> Do you know how to fix this? Do you know what is the directory that it >> is >> >> trying to clean? >> >> >> >> Any idea is welcomed! >> >> >> >> Thanks, >> >> Luis. >> > >> > >> > >> > >> > -- >> > Thomas Jungblut >> > Berlin >> > >> > mobile: 0170-3081070 >> > >> > business: [email protected] >> > private: [email protected] >> >> >> > > > -- > Thomas Jungblut > Berlin > > mobile: 0170-3081070 > > business: [email protected] > private: [email protected] > > > -- Thomas Jungblut Berlin mobile: 0170-3081070 business: [email protected] private: [email protected]
