Giraph 1.0 | Computation stuck at map 100% - reduce 0% for my algorithm only, at multi-node cluster

2014-09-29 Thread Panagiotis Eustratiadis
Good morning, I have been having a problem the past few days which sadly I can't solve. First of all I set up a Hadoop 0.20.203.0 cluster of two nodes a master and a slave. I followed this tutorial for the settings: http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-c

RE: Using a custom graph partitioning stratergy with giraph

2014-09-29 Thread Pavan Kumar A
we have two inputs - vertex & edgesif we partition edges vertices based on a map, then when we want to send messages we should be able to know which partition a vertex is on. typically we send messages to targetIds of outgoing edges, edge transation helps encode mapping information into targetId

Re: Using a custom graph partitioning stratergy with giraph

2014-09-29 Thread Charith Wickramarachchi
Hi Pavan, Thanks for the details. I went through the code specially the extension points you mentioned. I am not clear about the function of the edge Translation (org.apache.giraph .mapping.translate.TranslateEdge) class. Could you please explain the idea of this translation process. In my case I

RE: Graph re-partitioning

2014-09-29 Thread Pavan Kumar A
If you are using hashpartitioning, then as long as number of workers is same, partitions will remain unchanged, though they might run on a different worker. However, yes graph is always partitioned. Date: Mon, 29 Sep 2014 15:01:37 -0400 Subject: Graph re-partitioning From: xuhongne...@gmail.com T

Graph re-partitioning

2014-09-29 Thread xuhong zhang
Hello, Will Giraph re-partition the graph each time running a job on this graph? Is there anyway to directly load the partitioned graph from last job? Thanks -- Xuhong Zhang

How do I control which tasks run on which hosts?

2014-09-29 Thread Matthew Cornell
Hi Folks, I have a small CDH4 cluster of five hosts (four compute nodes and a head node - call them 0-3 and 'w') where hosts 0-3 have 4 cores and 16GB RAM each, and 'w' has 32 cores and 64GB RAM. All five hosts are running mapreduce tasktracker services, and 'w' is also running the jobtracker. Res