Now that I have the loading and computation completing
successfully, I am having issues when saving the edges back to
disk. During the saving step, the machines will get to ~1-2
partitions before the cluster freezes up entirely (as in, I
can't even SSH into the machine or view the Hadoop web
consol
Thanks for your help! I'll try it out.
--
Andrew
On Tue, Sep 9, 2014, at 10:25 AM, Pavan Kumar A wrote:
yes, you should implement your own edge store.
please take a look at ByteArrayEdges for example and modify it
to use BigDataOutput & BigDataInput instead of
ExtendedByteArrayOutput/Inpu
Great, thanks for pointing me in the right direction. All of
the edge values are strings (in a Text object) and point to and
from vertices with Text IDs, but none of the values should be
greater than 60 bytes or so during the loading step. The size
will increase during computation because I am modi
Hey,
I am currently running Giraph on a semi-large dataset of 600
million edges (the edges are directed, so I've used the
ReverseEdgeDuplicator for an expected total of 1.2b edges). I
am running into an issue during superstep -1 when the edges are
being loaded-- I receive a
"java.lang.NegativeAr