I have created an issue for that: HAMA-642<https://issues.apache.org/jira/browse/HAMA-642>
2012/9/14 Thomas Jungblut <[email protected]> > Basically I think that the graph should fit into memory of your task. > So the messages could cause the overflow. > > You can try out the DiskQueue, this can be configured with setting the > property "hama.messenger.queue.class" to > "org.apache.hama.bsp.message.DiskQueue". > > This will immediately flush the messages to disk. However this is > experimental currently, so if you try it out please tell us if it helped. > > Thanks. > > To further scale this, we should write vertices that don't fit in memory > on the disk. I will add another jira for that soon. > > 2012/9/14 庄克琛 <[email protected]> > >> oh, the HDFS block size is 128Mb, not 64Mb, so the 73Mb graph will not >> be split-ed on the HDFS. >> >> 2012/9/14 庄克琛 <[email protected]> >> >> > em... I have try your configure advise and restart the hama. >> > I use the Google web graph( >> > http://wiki.apache.org/hama/WriteHamaGraphFile ), >> > Nodes: 875713 Edges: 5105039, which is about 73Mb, upload to a small >> HDFS >> > cluster(block size is 64Mb), test the PageRank in ( >> > http://wiki.apache.org/hama/WriteHamaGraphFile ), got the result as: >> > ################ >> > function@624-PC:~/hadoop-1.0.3/hama-0.6.0$ hama jar hama-6-P* >> > input-google ouput-google >> > 12/09/14 14:27:50 INFO bsp.FileInputFormat: Total input paths to >> process : >> > 1 >> > 12/09/14 14:27:50 INFO bsp.FileInputFormat: Total # of splits: 3 >> > 12/09/14 14:27:50 INFO bsp.BSPJobClient: Running job: >> job_201008141420_0004 >> > 12/09/14 14:27:53 INFO bsp.BSPJobClient: Current supersteps number: 0 >> > Java HotSpot(TM) Server VM warning: Attempt to allocate stack guard >> pages >> > failed. >> > ################### >> > >> > Last time the supersteps could be 1 or 2, then the same result. >> > the task attempt****.err files are empty. >> > Is the graph too large? >> > I test on a small graph, get the right Rank results >> > >> > >> > 2012/9/14 Edward J. Yoon <[email protected]> >> > >> > I've added multi-step partitioning method to save memory[1]. >> >> >> >> Please try to configure below property to hama-site.xml. >> >> >> >> <property> >> >> <name>hama.graph.multi.step.partitioning.interval</name> >> >> <value>10000000</value> >> >> </property> >> >> >> >> 1. https://issues.apache.org/jira/browse/HAMA-599 >> >> >> >> On Fri, Sep 14, 2012 at 3:13 PM, 庄克琛 <[email protected]> wrote: >> >> > HI, Actually I use this ( >> >> > >> >> >> https://builds.apache.org/job/Hama-Nightly/672/artifact/.repository/org/apache/hama/hama-dist/0.6.0-SNAPSHOT/ >> >> > ) >> >> > to test again, I mean use this 0.6.0SNAPSHOT version replace >> everything, >> >> > got the same out of memory results. I just don't know what cause the >> >> out of >> >> > memory fails, only some small graph computing can be finished. Is >> this >> >> > version finished the " >> >> > [HAMA-596<https://issues.apache.org/jira/browse/HAMA-596>]:Optimize >> >> > memory usage of graph job" ? >> >> > Thanks >> >> > >> >> > 2012/9/14 Thomas Jungblut <[email protected]> >> >> > >> >> >> Hey, what jar did you exactly replace? >> >> >> Am 14.09.2012 07:49 schrieb "庄克琛" <[email protected]>: >> >> >> >> >> >> > hi, every one: >> >> >> > I use the hama-0.5.0 with the hadoop-1.0.3, try to do some large >> >> graphs >> >> >> > analysis. >> >> >> > When I test the PageRank examples, as the ( >> >> >> > http://wiki.apache.org/hama/WriteHamaGraphFile) shows, I download >> >> the >> >> >> > graph >> >> >> > data, and run the PageRank job on a small distributed cluser, I >> can >> >> only >> >> >> > get the out of memory failed, with Superstep 0,1,2 works well, >> then >> >> get >> >> >> the >> >> >> > memory out fail.(Each computer have 2G memory) But when I test >> some >> >> small >> >> >> > graph, everything went well. >> >> >> > Also I try the trunk version( >> >> >> > https://builds.apache.org/job/Hama-Nightly/672/changes#detail3), >> >> replace >> >> >> > my >> >> >> > hama-0.5.0 with the hama-0.6.0-snapshot, only get the same >> results. >> >> >> > Anyone got better ideas? >> >> >> > >> >> >> > Thanks! >> >> >> > >> >> >> > -- >> >> >> > >> >> >> > *Zhuang Kechen >> >> >> > * >> >> >> > >> >> >> >> >> > >> >> > >> >> > >> >> > -- >> >> > >> >> > *Zhuang Kechen* >> >> > >> >> > School of Computer Science & Technology >> >> > >> >> > ** >> >> > Nanjing University of Science & Technology >> >> > >> >> > Lab.623, School of Computer Sci. & Tech. >> >> > >> >> > No.200, Xiaolingwei Street >> >> > >> >> > Nanjing, Jiangsu, 210094 >> >> > >> >> > P.R. China >> >> > >> >> > Tel: 025-84315982** >> >> > >> >> > Email: [email protected] >> >> >> >> >> >> >> >> -- >> >> Best Regards, Edward J. Yoon >> >> @eddieyoon >> >> >> > >> > >> > >> > -- >> > >> > *Zhuang Kechen >> > * >> > >> > >> > >> >> >> -- >> >> *Zhuang Kechen* >> >> School of Computer Science & Technology >> >> ** >> Nanjing University of Science & Technology >> >> Lab.623, School of Computer Sci. & Tech. >> >> No.200, Xiaolingwei Street >> >> Nanjing, Jiangsu, 210094 >> >> P.R. China >> >> Tel: 025-84315982** >> >> Email: [email protected] >> > >
