Thanks again.  One of the things we struggle with currently on the
RDBMS, is the organisation of 250million records to complex
taxonomies, and also point in polygon intersections.  Having such
memory available the MR jobs allows us to consider loading taxonomies
/ polygons / RTree indexes into memory to do those calculations in
parallel with MR.  I was playing with that a couple of years ago when
I first ventured into Hadoop
(http://biodivertido.blogspot.com/2008/11/reproducing-spatial-joins-using-hadoop.html)
but might get back into it...

Tim

On Thu, Oct 14, 2010 at 8:07 PM, Jean-Daniel Cryans <jdcry...@apache.org> wrote:
>> I had it in my mind that HBase liked big memory, hence assuming the
>> region servers should stay on the 24G machines with plenty of memory
>> at their  disposal.  We'll come up with a test platform and then try
>> some benchmarking and do a blog on it all and share.
>>
>
> They do, but because of JVM limitations the recommended setting is
> around 4-8GB. Giving more would cause bigger heap fragmentation
> issues, leading to full GC pauses, which could cause session timeouts.
>
> J-D
>

Reply via email to