Hi, I am trying to configure my hbase (version 1.0) phoenix (version - 4.6) cluster to utilize as much memory as possible on the server hardware. We have an OLAP workload that allows users to perform interactive analysis over huge sets of data. While reading about hbase configuration I came across two configs
1. Hbase bucket cache <http://blog.asquareb.com/blog/2014/11/24/how-to-leverage-large-physical-memory-to-improve-hbase-read-performance> (off heap) which looks like a good option to bypass garbage collection. 2. Hadoop pinned hdfs blocks <http://blog.cloudera.com/blog/2014/08/new-in-cdh-5-1-hdfs-read-caching/> (max locked memory) concept that loads the hdfs blocks in memory, but given that hbase is configured with short circuit reads I assume this config may not be of much help. Instead it would be right to increase hbase region server heap memory. Is my understanding right? We use HBase with Phoenix. Kindly let me know your thoughts or suggestions on any more options that I should explore Thanks, Amit.
