Actually, I take that back. It seems it does succeeded in creating
partitions - it just struggles with it sometimes. Should I be worried about
these errors if partition directories seem to be filling up?
On Sep 11, 2013 6:38 PM, "Claudio Martella" <claudio.marte...@gmail.com>
wrote:

> Giraph does not offload partitions or messages to HDFS in the out-of-core
> module. It uses local disk on the computing nodes. By defualt, it uses the
> tasktracker local directory where for example the distributed cache is
> stored.
>
> Could you provide the stacktrace Giraph is spitting when failing?
>
>
> On Thu, Sep 12, 2013 at 12:54 AM, Alexander Asplund <alexaspl...@gmail.com
> > wrote:
>
>> Hi,
>>
>> I'm still trying to get Giraph to work on a graph that requires more
>> memory that is available. The problem is that when the Workers try to
>> offload partitions, the offloading fails. The DiskBackedPartitionStore
>> fails to create the directory
>> _bsp/_partitions/job-xxxx/part-vertices-xxx (roughly from recall).
>>
>> The input or computation will then continue for a while, which I
>> believe is because it is still managing to hold everything in memory -
>> but at some point it reaches the limit where there simply is no more
>> heap space, and it crashes with OOM.
>>
>> Has anybody had this problem with giraph failing to make HDFS directories?
>>
>
>
>
> --
>    Claudio Martella
>    claudio.marte...@gmail.com
>

Reply via email to