Lockup During Edge Saving

2014-09-11 Thread Andrew Munsell
Now that I have the loading and computation completing successfully, I am having issues when saving the edges back to disk. During the saving step, the machines will get to ~1-2 partitions before the cluster freezes up entirely (as in, I can't even SSH into the machine or view the Hadoop web consol

Re: NegativeArraySizeException with large dataset

2014-09-09 Thread Andrew Munsell
Thanks for your help! I'll try it out. -- Andrew On Tue, Sep 9, 2014, at 10:25 AM, Pavan Kumar A wrote: yes, you should implement your own edge store. please take a look at ByteArrayEdges for example and modify it to use BigDataOutput & BigDataInput instead of ExtendedByteArrayOutput/Inpu

Re: NegativeArraySizeException with large dataset

2014-09-09 Thread Andrew Munsell
Great, thanks for pointing me in the right direction. All of the edge values are strings (in a Text object) and point to and from vertices with Text IDs, but none of the values should be greater than 60 bytes or so during the loading step. The size will increase during computation because I am modi

NegativeArraySizeException with large dataset

2014-09-08 Thread Andrew Munsell
Hey, I am currently running Giraph on a semi-large dataset of 600 million edges (the edges are directed, so I've used the ReverseEdgeDuplicator for an expected total of 1.2b edges). I am running into an issue during superstep -1 when the edges are being loaded-- I receive a "java.lang.NegativeAr