Hello:
I have a graph of size 2B edges, 200M vertices. I am trying to compute
pagerank of this graph. I have 10 node cluster, 8G memory.
I am unable to run giraph job successfully. My question is more generic:
1. How should one choose right amount of heap space? .. is there a way of
telling
Can you paste your cluster information ? I am also struggling to make it
work on 75M vertices and 100s of million edges.
On Fri, Jul 26, 2013 at 8:02 AM, jerome richard jeromerichard...@msn.comwrote:
Hi,
I encountered a critical scaling problem using Giraph. I made a very
simple algorithm
Hello:
I am struggling to make PageRank run on 75M nodes with each node having
1-75000 edges.
I am constantly getting zookeeper timeouts irrespective of my configuration.
- I have 21 node hadoop cluster, each node having 4 cores, 4GB memory.
- Data is stored in hbase as adjacency matrix
- I am
(ZooKeeperExt.java:152)
at org.apache.giraph.bsp.BspService.getJobState(BspService.java:667)
On Wed, Jul 24, 2013 at 6:30 PM, Puneet Jain puneetdabu...@gmail.comwrote:
Hello:
I am struggling to make PageRank run on 75M nodes with each node having
1-75000 edges.
I am constantly getting zookeeper
Hello:
I have a graph with over million nodes and each node may be connected to
thousands of edges. My graph is stored in hbase as :
source, colon_sep_list_of_connected_nodes
I have thousands of such rows in my HBase table. I am facing issue in
running standard algorithms such as PageRank,
Hello:
I am unable to get rid of this exception, although my 3 port is always
available, it get this binding exception. NettyServer eventually finds
available port since it has 20 retries.
2013-06-25 09:38:04,203 WARN org.apache.giraph.comm.netty.NettyServer:
start: Likely failed to bind on
My run method:
args = new String[] {
/user/root/giraph-input/gooftech.tsv, /user/root/giraph-output,
'Shanae Dailey', X:2181};
if (args.length 4) {
System.err.println(printUsage());
System.exit(1);
}
if (args.length 4) {
System.err
.println(too many arguments.
+ Did you forget to quote