Cache uses only 2G of heap no matter how much allocated?
--------------------------------------------------------
Key: HBASE-2106
URL: https://issues.apache.org/jira/browse/HBASE-2106
Project: Hadoop HBase
Issue Type: Bug
Reporter: stack
Here is a question from Adam Silberstein who is having trouble writing this
list:
"I'm running on a particular cluster of machines and seeing worse than expected
performance. I've run on a different, but identical, cluster of machines and
gotten about 2-4x better throughput and latency. I've checked all of the
configuration settings and they are all the same. The only thing I observe
that looks odd is the memory usage by each region server. I've allocated 6 GB
heap space to each. There is a total of 20GB of records on each. When I use
"top" to observe memory usage, I see that the region server process does have
6GB of virtual memory. When I start running a workload, I can see the amount
of resident memory climbing. However, it always stops at 2GB. Since I have 20
GB on each region server, I'd expect the process to use all the memory
allocated it
"Is there a reasonable explanation for why it stops at 2GB, and/or has anyone
seen this happen"
I took a look at the cache code and I see that it has an int for maximum size
in the constructor argument though internally its using a long for maximum
size. I wonder if this is the issue?
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.