Cache uses only 2G of heap no matter how much allocated?
--------------------------------------------------------

                 Key: HBASE-2106
                 URL: https://issues.apache.org/jira/browse/HBASE-2106
             Project: Hadoop HBase
          Issue Type: Bug
            Reporter: stack


Here is a question from Adam Silberstein who is having trouble writing this 
list:

"I'm running on a particular cluster of machines and seeing worse than expected 
performance.  I've run on a different, but identical, cluster of machines and 
gotten about 2-4x better throughput and latency.  I've checked all of the 
configuration settings and they are all the same.  The only thing I observe 
that looks odd is the memory usage by each region server.  I've allocated 6 GB 
heap space to each.  There is a total of 20GB of records on each.  When I use 
"top" to observe memory usage, I see that the region server process does have 
6GB of virtual memory.  When I start running a workload, I can see the amount 
of resident memory climbing.  However, it always stops at 2GB.  Since I have 20 
GB on each region server, I'd expect the process to use all the memory 
allocated it

"Is there a reasonable explanation for why it stops at 2GB, and/or has anyone 
seen this happen"

I took a look at the cache code and I see that it has an int for maximum size 
in the constructor argument though internally its using a long for maximum 
size.  I wonder if this is the issue?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to