Re: Issues running a large MapReduce job over a complete HBase table

2010-12-07 Thread Gabriel Reid
Hi St.Ack, > You might be leaking scanners but that should have no effect on number > of open store files.  On deploy of a region, we open its store files > and hold them open and do not open others -- not unless > compacting/splitting. > > Hope this helps, Yes, huge help, thank you very much for

Re: Issues running a large MapReduce job over a complete HBase table

2010-12-06 Thread Gabriel Reid
message?  You might tune the thread stack size and that > might give you headroom you need.  How many nodes in your cluster and > how much RAM they have? > > Thanks, > St.Ack > P.S. Yes, bigger files could help but OOME in DN is a little unusual. > > On Mon, Dec 6, 2010 at 4:3

Re: Issues running a large MapReduce job over a complete HBase table

2010-12-06 Thread Gabriel Reid
> you see OOMEs, I would like to know what it has consumed. You are > saying the Hadoop DataNodes actually crash with the OOME? > > Lars > > On Mon, Dec 6, 2010 at 9:02 AM, Gabriel Reid wrote: >> Hi, >> >> We're currently running into issues with running

Issues running a large MapReduce job over a complete HBase table

2010-12-06 Thread Gabriel Reid
Hi, We're currently running into issues with running a MapReduce job over a complete HBase table - we can't seem to find a balance between having dfs.datanode.max.xcievers set too low (and getting "xceiverCount X exceeds the limit of concurrent xcievers") and getting OutOfMemoryErrors on datanodes