http://git-wip-us.apache.org/repos/asf/hbase/blob/92aa9dc8/src/main/asciidoc/hbase-default.xml
----------------------------------------------------------------------
diff --git a/src/main/asciidoc/hbase-default.xml 
b/src/main/asciidoc/hbase-default.xml
new file mode 100644
index 0000000..630bc8e
--- /dev/null
+++ b/src/main/asciidoc/hbase-default.xml
@@ -0,0 +1,1393 @@
+
+
+:doctype: book
+:numbered:
+:toc: left
+:icons: font
+:experimental:
+
+[[_hbase_default_configurations]]
+== HBase Default Configuration
+
+The documentation below is generated using the default hbase configuration 
file, _hbase-default.xml_, as source.
+
+
+
+[[_hbase.tmp.dir]]
+.hbase.tmp.dir
+Description:: Temporary directory on the local filesystem.
+    Change this setting to point to a location more permanent
+    than '/tmp', the usual resolve for java.io.tmpdir, as the
+    '/tmp' directory is cleared on machine restart.
+Default:: ${java.io.tmpdir}/hbase-${user.name} +
+  
+[[_hbase.rootdir]]
+.hbase.rootdir
+Description:: The directory shared by region servers and into
+    which HBase persists.  The URL should be 'fully-qualified'
+    to include the filesystem scheme.  For example, to specify the
+    HDFS directory '/hbase' where the HDFS instance's namenode is
+    running at namenode.example.org on port 9000, set this value to:
+    hdfs://namenode.example.org:9000/hbase.  By default, we write
+    to whatever ${hbase.tmp.dir} is set too -- usually /tmp --
+    so change this configuration or else all data will be lost on
+    machine restart.
+Default:: ${hbase.tmp.dir}/hbase +
+  
+[[_hbase.cluster.distributed]]
+.hbase.cluster.distributed
+Description:: The mode the cluster will be in. Possible values are
+      false for standalone mode and true for distributed mode.  If
+      false, startup will run all HBase and ZooKeeper daemons together
+      in the one JVM.
+Default:: false +
+  
+[[_hbase.zookeeper.quorum]]
+.hbase.zookeeper.quorum
+Description:: Comma separated list of servers in the ZooKeeper ensemble
+    (This config. should have been named hbase.zookeeper.ensemble).
+    For example, "host1.mydomain.com,host2.mydomain.com,host3.mydomain.com".
+    By default this is set to localhost for local and pseudo-distributed modes
+    of operation. For a fully-distributed setup, this should be set to a full
+    list of ZooKeeper ensemble servers. If HBASE_MANAGES_ZK is set in 
hbase-env.sh
+    this is the list of servers which hbase will start/stop ZooKeeper on as
+    part of cluster start/stop.  Client-side, we will take this list of
+    ensemble members and put it together with the hbase.zookeeper.clientPort
+    config. and pass it into zookeeper constructor as the connectString
+    parameter.
+Default:: localhost +
+  
+[[_hbase.local.dir]]
+.hbase.local.dir
+Description:: Directory on the local filesystem to be used
+    as a local storage.
+Default:: ${hbase.tmp.dir}/local/ +
+  
+[[_hbase.master.info.port]]
+.hbase.master.info.port
+Description:: The port for the HBase Master web UI.
+    Set to -1 if you do not want a UI instance run.
+Default:: 16010 +
+  
+[[_hbase.master.info.bindAddress]]
+.hbase.master.info.bindAddress
+Description:: The bind address for the HBase Master web UI
+    
+Default:: 0.0.0.0 +
+  
+[[_hbase.master.logcleaner.plugins]]
+.hbase.master.logcleaner.plugins
+Description:: A comma-separated list of BaseLogCleanerDelegate invoked by
+    the LogsCleaner service. These WAL cleaners are called in order,
+    so put the cleaner that prunes the most files in front. To
+    implement your own BaseLogCleanerDelegate, just put it in HBase's classpath
+    and add the fully qualified class name here. Always add the above
+    default log cleaners in the list.
+Default:: org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner +
+  
+[[_hbase.master.logcleaner.ttl]]
+.hbase.master.logcleaner.ttl
+Description:: Maximum time a WAL can stay in the .oldlogdir directory,
+    after which it will be cleaned by a Master thread.
+Default:: 600000 +
+  
+[[_hbase.master.hfilecleaner.plugins]]
+.hbase.master.hfilecleaner.plugins
+Description:: A comma-separated list of BaseHFileCleanerDelegate invoked by
+    the HFileCleaner service. These HFiles cleaners are called in order,
+    so put the cleaner that prunes the most files in front. To
+    implement your own BaseHFileCleanerDelegate, just put it in HBase's 
classpath
+    and add the fully qualified class name here. Always add the above
+    default log cleaners in the list as they will be overwritten in
+    hbase-site.xml.
+Default:: org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner +
+  
+[[_hbase.master.catalog.timeout]]
+.hbase.master.catalog.timeout
+Description:: Timeout value for the Catalog Janitor from the master to
+    META.
+Default:: 600000 +
+  
+[[_hbase.master.infoserver.redirect]]
+.hbase.master.infoserver.redirect
+Description:: Whether or not the Master listens to the Master web
+      UI port (hbase.master.info.port) and redirects requests to the web
+      UI server shared by the Master and RegionServer.
+Default:: true +
+  
+[[_hbase.regionserver.port]]
+.hbase.regionserver.port
+Description:: The port the HBase RegionServer binds to.
+Default:: 16020 +
+  
+[[_hbase.regionserver.info.port]]
+.hbase.regionserver.info.port
+Description:: The port for the HBase RegionServer web UI
+    Set to -1 if you do not want the RegionServer UI to run.
+Default:: 16030 +
+  
+[[_hbase.regionserver.info.bindAddress]]
+.hbase.regionserver.info.bindAddress
+Description:: The address for the HBase RegionServer web UI
+Default:: 0.0.0.0 +
+  
+[[_hbase.regionserver.info.port.auto]]
+.hbase.regionserver.info.port.auto
+Description:: Whether or not the Master or RegionServer
+    UI should search for a port to bind to. Enables automatic port
+    search if hbase.regionserver.info.port is already in use.
+    Useful for testing, turned off by default.
+Default:: false +
+  
+[[_hbase.regionserver.handler.count]]
+.hbase.regionserver.handler.count
+Description:: Count of RPC Listener instances spun up on RegionServers.
+    Same property is used by the Master for count of master handlers.
+Default:: 30 +
+  
+[[_hbase.ipc.server.callqueue.handler.factor]]
+.hbase.ipc.server.callqueue.handler.factor
+Description:: Factor to determine the number of call queues.
+      A value of 0 means a single queue shared between all the handlers.
+      A value of 1 means that each handler has its own queue.
+Default:: 0.1 +
+  
+[[_hbase.ipc.server.callqueue.read.ratio]]
+.hbase.ipc.server.callqueue.read.ratio
+Description:: Split the call queues into read and write queues.
+      The specified interval (which should be between 0.0 and 1.0)
+      will be multiplied by the number of call queues.
+      A value of 0 indicate to not split the call queues, meaning that both 
read and write
+      requests will be pushed to the same set of queues.
+      A value lower than 0.5 means that there will be less read queues than 
write queues.
+      A value of 0.5 means there will be the same number of read and write 
queues.
+      A value greater than 0.5 means that there will be more read queues than 
write queues.
+      A value of 1.0 means that all the queues except one are used to dispatch 
read requests.
+
+      Example: Given the total number of call queues being 10
+      a read.ratio of 0 means that: the 10 queues will contain both read/write 
requests.
+      a read.ratio of 0.3 means that: 3 queues will contain only read requests
+      and 7 queues will contain only write requests.
+      a read.ratio of 0.5 means that: 5 queues will contain only read requests
+      and 5 queues will contain only write requests.
+      a read.ratio of 0.8 means that: 8 queues will contain only read requests
+      and 2 queues will contain only write requests.
+      a read.ratio of 1 means that: 9 queues will contain only read requests
+      and 1 queues will contain only write requests.
+    
+Default:: 0 +
+  
+[[_hbase.ipc.server.callqueue.scan.ratio]]
+.hbase.ipc.server.callqueue.scan.ratio
+Description:: Given the number of read call queues, calculated from the total 
number
+      of call queues multiplied by the callqueue.read.ratio, the scan.ratio 
property
+      will split the read call queues into small-read and long-read queues.
+      A value lower than 0.5 means that there will be less long-read queues 
than short-read queues.
+      A value of 0.5 means that there will be the same number of short-read 
and long-read queues.
+      A value greater than 0.5 means that there will be more long-read queues 
than short-read queues
+      A value of 0 or 1 indicate to use the same set of queues for gets and 
scans.
+
+      Example: Given the total number of read call queues being 8
+      a scan.ratio of 0 or 1 means that: 8 queues will contain both long and 
short read requests.
+      a scan.ratio of 0.3 means that: 2 queues will contain only long-read 
requests
+      and 6 queues will contain only short-read requests.
+      a scan.ratio of 0.5 means that: 4 queues will contain only long-read 
requests
+      and 4 queues will contain only short-read requests.
+      a scan.ratio of 0.8 means that: 6 queues will contain only long-read 
requests
+      and 2 queues will contain only short-read requests.
+    
+Default:: 0 +
+  
+[[_hbase.regionserver.msginterval]]
+.hbase.regionserver.msginterval
+Description:: Interval between messages from the RegionServer to Master
+    in milliseconds.
+Default:: 3000 +
+  
+[[_hbase.regionserver.regionSplitLimit]]
+.hbase.regionserver.regionSplitLimit
+Description:: Limit for the number of regions after which no more region
+    splitting should take place. This is not a hard limit for the number of
+    regions but acts as a guideline for the regionserver to stop splitting 
after
+    a certain limit. Default is MAX_INT; i.e. do not block splitting.
+Default:: 2147483647 +
+  
+[[_hbase.regionserver.logroll.period]]
+.hbase.regionserver.logroll.period
+Description:: Period at which we will roll the commit log regardless
+    of how many edits it has.
+Default:: 3600000 +
+  
+[[_hbase.regionserver.logroll.errors.tolerated]]
+.hbase.regionserver.logroll.errors.tolerated
+Description:: The number of consecutive WAL close errors we will allow
+    before triggering a server abort.  A setting of 0 will cause the
+    region server to abort if closing the current WAL writer fails during
+    log rolling.  Even a small value (2 or 3) will allow a region server
+    to ride over transient HDFS errors.
+Default:: 2 +
+  
+[[_hbase.regionserver.hlog.reader.impl]]
+.hbase.regionserver.hlog.reader.impl
+Description:: The WAL file reader implementation.
+Default:: org.apache.hadoop.hbase.regionserver.wal.ProtobufLogReader +
+  
+[[_hbase.regionserver.hlog.writer.impl]]
+.hbase.regionserver.hlog.writer.impl
+Description:: The WAL file writer implementation.
+Default:: org.apache.hadoop.hbase.regionserver.wal.ProtobufLogWriter +
+  
+[[_hbase.master.distributed.log.replay]]
+.hbase.master.distributed.log.replay
+Description:: Enable 'distributed log replay' as default engine splitting
+    WAL files on server crash.  This default is new in hbase 1.0.  To fall
+    back to the old mode 'distributed log splitter', set the value to
+    'false'.  'Disributed log replay' improves MTTR because it does not
+    write intermediate files.  'DLR' required that 'hfile.format.version'
+    be set to version 3 or higher. 
+    
+Default:: true +
+  
+[[_hbase.regionserver.global.memstore.size]]
+.hbase.regionserver.global.memstore.size
+Description:: Maximum size of all memstores in a region server before new
+      updates are blocked and flushes are forced. Defaults to 40% of heap.
+      Updates are blocked and flushes are forced until size of all memstores
+      in a region server hits 
hbase.regionserver.global.memstore.size.lower.limit.
+Default:: 0.4 +
+  
+[[_hbase.regionserver.global.memstore.size.lower.limit]]
+.hbase.regionserver.global.memstore.size.lower.limit
+Description:: Maximum size of all memstores in a region server before flushes 
are forced.
+      Defaults to 95% of hbase.regionserver.global.memstore.size.
+      A 100% value for this value causes the minimum possible flushing to 
occur when updates are 
+      blocked due to memstore limiting.
+Default:: 0.95 +
+  
+[[_hbase.regionserver.optionalcacheflushinterval]]
+.hbase.regionserver.optionalcacheflushinterval
+Description:: 
+    Maximum amount of time an edit lives in memory before being automatically 
flushed.
+    Default 1 hour. Set it to 0 to disable automatic flushing.
+Default:: 3600000 +
+  
+[[_hbase.regionserver.catalog.timeout]]
+.hbase.regionserver.catalog.timeout
+Description:: Timeout value for the Catalog Janitor from the regionserver to 
META.
+Default:: 600000 +
+  
+[[_hbase.regionserver.dns.interface]]
+.hbase.regionserver.dns.interface
+Description:: The name of the Network Interface from which a region server
+      should report its IP address.
+Default:: default +
+  
+[[_hbase.regionserver.dns.nameserver]]
+.hbase.regionserver.dns.nameserver
+Description:: The host name or IP address of the name server (DNS)
+      which a region server should use to determine the host name used by the
+      master for communication and display purposes.
+Default:: default +
+  
+[[_hbase.regionserver.region.split.policy]]
+.hbase.regionserver.region.split.policy
+Description:: 
+      A split policy determines when a region should be split. The various 
other split policies that
+      are available currently are ConstantSizeRegionSplitPolicy, 
DisabledRegionSplitPolicy,
+      DelimitedKeyPrefixRegionSplitPolicy, KeyPrefixRegionSplitPolicy etc.
+    
+Default:: 
org.apache.hadoop.hbase.regionserver.IncreasingToUpperBoundRegionSplitPolicy +
+  
+[[_zookeeper.session.timeout]]
+.zookeeper.session.timeout
+Description:: ZooKeeper session timeout in milliseconds. It is used in two 
different ways.
+      First, this value is used in the ZK client that HBase uses to connect to 
the ensemble.
+      It is also used by HBase when it starts a ZK server and it is passed as 
the 'maxSessionTimeout'. See
+      
http://hadoop.apache.org/zookeeper/docs/current/zookeeperProgrammers.html#ch_zkSessions.
+      For example, if a HBase region server connects to a ZK ensemble that's 
also managed by HBase, then the
+      session timeout will be the one specified by this configuration. But, a 
region server that connects
+      to an ensemble managed with a different configuration will be subjected 
that ensemble's maxSessionTimeout. So,
+      even though HBase might propose using 90 seconds, the ensemble can have 
a max timeout lower than this and
+      it will take precedence. The current default that ZK ships with is 40 
seconds, which is lower than HBase's.
+    
+Default:: 90000 +
+  
+[[_zookeeper.znode.parent]]
+.zookeeper.znode.parent
+Description:: Root ZNode for HBase in ZooKeeper. All of HBase's ZooKeeper
+      files that are configured with a relative path will go under this node.
+      By default, all of HBase's ZooKeeper file path are configured with a
+      relative path, so they will all go under this directory unless changed.
+Default:: /hbase +
+  
+[[_zookeeper.znode.rootserver]]
+.zookeeper.znode.rootserver
+Description:: Path to ZNode holding root region location. This is written by
+      the master and read by clients and region servers. If a relative path is
+      given, the parent folder will be ${zookeeper.znode.parent}. By default,
+      this means the root location is stored at /hbase/root-region-server.
+Default:: root-region-server +
+  
+[[_zookeeper.znode.acl.parent]]
+.zookeeper.znode.acl.parent
+Description:: Root ZNode for access control lists.
+Default:: acl +
+  
+[[_hbase.zookeeper.dns.interface]]
+.hbase.zookeeper.dns.interface
+Description:: The name of the Network Interface from which a ZooKeeper server
+      should report its IP address.
+Default:: default +
+  
+[[_hbase.zookeeper.dns.nameserver]]
+.hbase.zookeeper.dns.nameserver
+Description:: The host name or IP address of the name server (DNS)
+      which a ZooKeeper server should use to determine the host name used by 
the
+      master for communication and display purposes.
+Default:: default +
+  
+[[_hbase.zookeeper.peerport]]
+.hbase.zookeeper.peerport
+Description:: Port used by ZooKeeper peers to talk to each other.
+    See 
http://hadoop.apache.org/zookeeper/docs/r3.1.1/zookeeperStarted.html#sc_RunningReplicatedZooKeeper
+    for more information.
+Default:: 2888 +
+  
+[[_hbase.zookeeper.leaderport]]
+.hbase.zookeeper.leaderport
+Description:: Port used by ZooKeeper for leader election.
+    See 
http://hadoop.apache.org/zookeeper/docs/r3.1.1/zookeeperStarted.html#sc_RunningReplicatedZooKeeper
+    for more information.
+Default:: 3888 +
+  
+[[_hbase.zookeeper.useMulti]]
+.hbase.zookeeper.useMulti
+Description:: Instructs HBase to make use of ZooKeeper's multi-update 
functionality.
+    This allows certain ZooKeeper operations to complete more quickly and 
prevents some issues
+    with rare Replication failure scenarios (see the release note of 
HBASE-2611 for an example).
+    IMPORTANT: only set this to true if all ZooKeeper servers in the cluster 
are on version 3.4+
+    and will not be downgraded.  ZooKeeper versions before 3.4 do not support 
multi-update and
+    will not fail gracefully if multi-update is invoked (see ZOOKEEPER-1495).
+Default:: true +
+  
+[[_hbase.config.read.zookeeper.config]]
+.hbase.config.read.zookeeper.config
+Description:: 
+        Set to true to allow HBaseConfiguration to read the
+        zoo.cfg file for ZooKeeper properties. Switching this to true
+        is not recommended, since the functionality of reading ZK
+        properties from a zoo.cfg file has been deprecated.
+Default:: false +
+  
+[[_hbase.zookeeper.property.initLimit]]
+.hbase.zookeeper.property.initLimit
+Description:: Property from ZooKeeper's config zoo.cfg.
+    The number of ticks that the initial synchronization phase can take.
+Default:: 10 +
+  
+[[_hbase.zookeeper.property.syncLimit]]
+.hbase.zookeeper.property.syncLimit
+Description:: Property from ZooKeeper's config zoo.cfg.
+    The number of ticks that can pass between sending a request and getting an
+    acknowledgment.
+Default:: 5 +
+  
+[[_hbase.zookeeper.property.dataDir]]
+.hbase.zookeeper.property.dataDir
+Description:: Property from ZooKeeper's config zoo.cfg.
+    The directory where the snapshot is stored.
+Default:: ${hbase.tmp.dir}/zookeeper +
+  
+[[_hbase.zookeeper.property.clientPort]]
+.hbase.zookeeper.property.clientPort
+Description:: Property from ZooKeeper's config zoo.cfg.
+    The port at which the clients will connect.
+Default:: 2181 +
+  
+[[_hbase.zookeeper.property.maxClientCnxns]]
+.hbase.zookeeper.property.maxClientCnxns
+Description:: Property from ZooKeeper's config zoo.cfg.
+    Limit on number of concurrent connections (at the socket level) that a
+    single client, identified by IP address, may make to a single member of
+    the ZooKeeper ensemble. Set high to avoid zk connection issues running
+    standalone and pseudo-distributed.
+Default:: 300 +
+  
+[[_hbase.client.write.buffer]]
+.hbase.client.write.buffer
+Description:: Default size of the HTable client write buffer in bytes.
+    A bigger buffer takes more memory -- on both the client and server
+    side since server instantiates the passed write buffer to process
+    it -- but a larger buffer size reduces the number of RPCs made.
+    For an estimate of server-side memory-used, evaluate
+    hbase.client.write.buffer * hbase.regionserver.handler.count
+Default:: 2097152 +
+  
+[[_hbase.client.pause]]
+.hbase.client.pause
+Description:: General client pause value.  Used mostly as value to wait
+    before running a retry of a failed get, region lookup, etc.
+    See hbase.client.retries.number for description of how we backoff from
+    this initial pause amount and how this pause works w/ retries.
+Default:: 100 +
+  
+[[_hbase.client.retries.number]]
+.hbase.client.retries.number
+Description:: Maximum retries.  Used as maximum for all retryable
+    operations such as the getting of a cell's value, starting a row update,
+    etc.  Retry interval is a rough function based on hbase.client.pause.  At
+    first we retry at this interval but then with backoff, we pretty quickly 
reach
+    retrying every ten seconds.  See HConstants#RETRY_BACKOFF for how the 
backup
+    ramps up.  Change this setting and hbase.client.pause to suit your 
workload.
+Default:: 35 +
+  
+[[_hbase.client.max.total.tasks]]
+.hbase.client.max.total.tasks
+Description:: The maximum number of concurrent tasks a single HTable instance 
will
+    send to the cluster.
+Default:: 100 +
+  
+[[_hbase.client.max.perserver.tasks]]
+.hbase.client.max.perserver.tasks
+Description:: The maximum number of concurrent tasks a single HTable instance 
will
+    send to a single region server.
+Default:: 5 +
+  
+[[_hbase.client.max.perregion.tasks]]
+.hbase.client.max.perregion.tasks
+Description:: The maximum number of concurrent connections the client will
+    maintain to a single Region. That is, if there is already
+    hbase.client.max.perregion.tasks writes in progress for this region, new 
puts
+    won't be sent to this region until some writes finishes.
+Default:: 1 +
+  
+[[_hbase.client.scanner.caching]]
+.hbase.client.scanner.caching
+Description:: Number of rows that will be fetched when calling next
+    on a scanner if it is not served from (local, client) memory. Higher
+    caching values will enable faster scanners but will eat up more memory
+    and some calls of next may take longer and longer times when the cache is 
empty.
+    Do not set this value such that the time between invocations is greater
+    than the scanner timeout; i.e. hbase.client.scanner.timeout.period
+Default:: 100 +
+  
+[[_hbase.client.keyvalue.maxsize]]
+.hbase.client.keyvalue.maxsize
+Description:: Specifies the combined maximum allowed size of a KeyValue
+    instance. This is to set an upper boundary for a single entry saved in a
+    storage file. Since they cannot be split it helps avoiding that a region
+    cannot be split any further because the data is too large. It seems wise
+    to set this to a fraction of the maximum region size. Setting it to zero
+    or less disables the check.
+Default:: 10485760 +
+  
+[[_hbase.client.scanner.timeout.period]]
+.hbase.client.scanner.timeout.period
+Description:: Client scanner lease period in milliseconds.
+Default:: 60000 +
+  
+[[_hbase.client.localityCheck.threadPoolSize]]
+.hbase.client.localityCheck.threadPoolSize
+Description:: 
+Default:: 2 +
+  
+[[_hbase.bulkload.retries.number]]
+.hbase.bulkload.retries.number
+Description:: Maximum retries.  This is maximum number of iterations
+    to atomic bulk loads are attempted in the face of splitting operations
+    0 means never give up.
+Default:: 10 +
+  
+[[_hbase.balancer.period
+    ]]
+.hbase.balancer.period
+    
+Description:: Period at which the region balancer runs in the Master.
+Default:: 300000 +
+  
+[[_hbase.regions.slop]]
+.hbase.regions.slop
+Description:: Rebalance if any regionserver has average + (average * slop) 
regions.
+Default:: 0.2 +
+  
+[[_hbase.server.thread.wakefrequency]]
+.hbase.server.thread.wakefrequency
+Description:: Time to sleep in between searches for work (in milliseconds).
+    Used as sleep interval by service threads such as log roller.
+Default:: 10000 +
+  
+[[_hbase.server.versionfile.writeattempts]]
+.hbase.server.versionfile.writeattempts
+Description:: 
+    How many time to retry attempting to write a version file
+    before just aborting. Each attempt is seperated by the
+    hbase.server.thread.wakefrequency milliseconds.
+Default:: 3 +
+  
+[[_hbase.hregion.memstore.flush.size]]
+.hbase.hregion.memstore.flush.size
+Description:: 
+    Memstore will be flushed to disk if size of the memstore
+    exceeds this number of bytes.  Value is checked by a thread that runs
+    every hbase.server.thread.wakefrequency.
+Default:: 134217728 +
+  
+[[_hbase.hregion.percolumnfamilyflush.size.lower.bound]]
+.hbase.hregion.percolumnfamilyflush.size.lower.bound
+Description:: 
+    If FlushLargeStoresPolicy is used, then every time that we hit the
+    total memstore limit, we find out all the column families whose memstores
+    exceed this value, and only flush them, while retaining the others whose
+    memstores are lower than this limit. If none of the families have their
+    memstore size more than this, all the memstores will be flushed
+    (just as usual). This value should be less than half of the total memstore
+    threshold (hbase.hregion.memstore.flush.size).
+    
+Default:: 16777216 +
+  
+[[_hbase.hregion.preclose.flush.size]]
+.hbase.hregion.preclose.flush.size
+Description:: 
+      If the memstores in a region are this size or larger when we go
+      to close, run a "pre-flush" to clear out memstores before we put up
+      the region closed flag and take the region offline.  On close,
+      a flush is run under the close flag to empty memory.  During
+      this time the region is offline and we are not taking on any writes.
+      If the memstore content is large, this flush could take a long time to
+      complete.  The preflush is meant to clean out the bulk of the memstore
+      before putting up the close flag and taking the region offline so the
+      flush that runs under the close flag has little to do.
+Default:: 5242880 +
+  
+[[_hbase.hregion.memstore.block.multiplier]]
+.hbase.hregion.memstore.block.multiplier
+Description:: 
+    Block updates if memstore has hbase.hregion.memstore.block.multiplier
+    times hbase.hregion.memstore.flush.size bytes.  Useful preventing
+    runaway memstore during spikes in update traffic.  Without an
+    upper-bound, memstore fills such that when it flushes the
+    resultant flush files take a long time to compact or split, or
+    worse, we OOME.
+Default:: 4 +
+  
+[[_hbase.hregion.memstore.mslab.enabled]]
+.hbase.hregion.memstore.mslab.enabled
+Description:: 
+      Enables the MemStore-Local Allocation Buffer,
+      a feature which works to prevent heap fragmentation under
+      heavy write loads. This can reduce the frequency of stop-the-world
+      GC pauses on large heaps.
+Default:: true +
+  
+[[_hbase.hregion.max.filesize]]
+.hbase.hregion.max.filesize
+Description:: 
+    Maximum HFile size. If the sum of the sizes of a region's HFiles has grown 
to exceed this 
+    value, the region is split in two.
+Default:: 10737418240 +
+  
+[[_hbase.hregion.majorcompaction]]
+.hbase.hregion.majorcompaction
+Description:: Time between major compactions, expressed in milliseconds. Set 
to 0 to disable
+      time-based automatic major compactions. User-requested and size-based 
major compactions will
+      still run. This value is multiplied by 
hbase.hregion.majorcompaction.jitter to cause
+      compaction to start at a somewhat-random time during a given window of 
time. The default value
+      is 7 days, expressed in milliseconds. If major compactions are causing 
disruption in your
+      environment, you can configure them to run at off-peak times for your 
deployment, or disable
+      time-based major compactions by setting this parameter to 0, and run 
major compactions in a
+      cron job or by another external mechanism.
+Default:: 604800000 +
+  
+[[_hbase.hregion.majorcompaction.jitter]]
+.hbase.hregion.majorcompaction.jitter
+Description:: A multiplier applied to hbase.hregion.majorcompaction to cause 
compaction to occur
+      a given amount of time either side of hbase.hregion.majorcompaction. The 
smaller the number,
+      the closer the compactions will happen to the 
hbase.hregion.majorcompaction
+      interval.
+Default:: 0.50 +
+  
+[[_hbase.hstore.compactionThreshold]]
+.hbase.hstore.compactionThreshold
+Description::  If more than this number of StoreFiles exist in any one Store 
+      (one StoreFile is written per flush of MemStore), a compaction is run to 
rewrite all 
+      StoreFiles into a single StoreFile. Larger values delay compaction, but 
when compaction does
+      occur, it takes longer to complete.
+Default:: 3 +
+  
+[[_hbase.hstore.flusher.count]]
+.hbase.hstore.flusher.count
+Description::  The number of flush threads. With fewer threads, the MemStore 
flushes will be
+      queued. With more threads, the flushes will be executed in parallel, 
increasing the load on
+      HDFS, and potentially causing more compactions. 
+Default:: 2 +
+  
+[[_hbase.hstore.blockingStoreFiles]]
+.hbase.hstore.blockingStoreFiles
+Description::  If more than this number of StoreFiles exist in any one Store 
(one StoreFile
+     is written per flush of MemStore), updates are blocked for this region 
until a compaction is
+      completed, or until hbase.hstore.blockingWaitTime has been exceeded.
+Default:: 10 +
+  
+[[_hbase.hstore.blockingWaitTime]]
+.hbase.hstore.blockingWaitTime
+Description::  The time for which a region will block updates after reaching 
the StoreFile limit
+    defined by hbase.hstore.blockingStoreFiles. After this time has elapsed, 
the region will stop 
+    blocking updates even if a compaction has not been completed.
+Default:: 90000 +
+  
+[[_hbase.hstore.compaction.min]]
+.hbase.hstore.compaction.min
+Description:: The minimum number of StoreFiles which must be eligible for 
compaction before 
+      compaction can run. The goal of tuning hbase.hstore.compaction.min is to 
avoid ending up with 
+      too many tiny StoreFiles to compact. Setting this value to 2 would cause 
a minor compaction 
+      each time you have two StoreFiles in a Store, and this is probably not 
appropriate. If you
+      set this value too high, all the other values will need to be adjusted 
accordingly. For most 
+      cases, the default value is appropriate. In previous versions of HBase, 
the parameter
+      hbase.hstore.compaction.min was named hbase.hstore.compactionThreshold.
+Default:: 3 +
+  
+[[_hbase.hstore.compaction.max]]
+.hbase.hstore.compaction.max
+Description:: The maximum number of StoreFiles which will be selected for a 
single minor 
+      compaction, regardless of the number of eligible StoreFiles. 
Effectively, the value of
+      hbase.hstore.compaction.max controls the length of time it takes a 
single compaction to
+      complete. Setting it larger means that more StoreFiles are included in a 
compaction. For most
+      cases, the default value is appropriate.
+Default:: 10 +
+  
+[[_hbase.hstore.compaction.min.size]]
+.hbase.hstore.compaction.min.size
+Description:: A StoreFile smaller than this size will always be eligible for 
minor compaction. 
+      HFiles this size or larger are evaluated by 
hbase.hstore.compaction.ratio to determine if 
+      they are eligible. Because this limit represents the "automatic 
include"limit for all 
+      StoreFiles smaller than this value, this value may need to be reduced in 
write-heavy 
+      environments where many StoreFiles in the 1-2 MB range are being 
flushed, because every 
+      StoreFile will be targeted for compaction and the resulting StoreFiles 
may still be under the
+      minimum size and require further compaction. If this parameter is 
lowered, the ratio check is
+      triggered more quickly. This addressed some issues seen in earlier 
versions of HBase but 
+      changing this parameter is no longer necessary in most situations. 
Default: 128 MB expressed 
+      in bytes.
+Default:: 134217728 +
+  
+[[_hbase.hstore.compaction.max.size]]
+.hbase.hstore.compaction.max.size
+Description:: A StoreFile larger than this size will be excluded from 
compaction. The effect of 
+      raising hbase.hstore.compaction.max.size is fewer, larger StoreFiles 
that do not get 
+      compacted often. If you feel that compaction is happening too often 
without much benefit, you
+      can try raising this value. Default: the value of LONG.MAX_VALUE, 
expressed in bytes.
+Default:: 9223372036854775807 +
+  
+[[_hbase.hstore.compaction.ratio]]
+.hbase.hstore.compaction.ratio
+Description:: For minor compaction, this ratio is used to determine whether a 
given StoreFile 
+      which is larger than hbase.hstore.compaction.min.size is eligible for 
compaction. Its
+      effect is to limit compaction of large StoreFiles. The value of 
hbase.hstore.compaction.ratio
+      is expressed as a floating-point decimal. A large ratio, such as 10, 
will produce a single 
+      giant StoreFile. Conversely, a low value, such as .25, will produce 
behavior similar to the 
+      BigTable compaction algorithm, producing four StoreFiles. A moderate 
value of between 1.0 and
+      1.4 is recommended. When tuning this value, you are balancing write 
costs with read costs. 
+      Raising the value (to something like 1.4) will have more write costs, 
because you will 
+      compact larger StoreFiles. However, during reads, HBase will need to 
seek through fewer 
+      StoreFiles to accomplish the read. Consider this approach if you cannot 
take advantage of 
+      Bloom filters. Otherwise, you can lower this value to something like 1.0 
to reduce the 
+      background cost of writes, and use Bloom filters to control the number 
of StoreFiles touched 
+      during reads. For most cases, the default value is appropriate.
+Default:: 1.2F +
+  
+[[_hbase.hstore.compaction.ratio.offpeak]]
+.hbase.hstore.compaction.ratio.offpeak
+Description:: Allows you to set a different (by default, more aggressive) 
ratio for determining
+      whether larger StoreFiles are included in compactions during off-peak 
hours. Works in the 
+      same way as hbase.hstore.compaction.ratio. Only applies if 
hbase.offpeak.start.hour and 
+      hbase.offpeak.end.hour are also enabled.
+Default:: 5.0F +
+  
+[[_hbase.hstore.time.to.purge.deletes]]
+.hbase.hstore.time.to.purge.deletes
+Description:: The amount of time to delay purging of delete markers with 
future timestamps. If 
+      unset, or set to 0, all delete markers, including those with future 
timestamps, are purged 
+      during the next major compaction. Otherwise, a delete marker is kept 
until the major compaction 
+      which occurs after the marker's timestamp plus the value of this 
setting, in milliseconds.
+    
+Default:: 0 +
+  
+[[_hbase.offpeak.start.hour]]
+.hbase.offpeak.start.hour
+Description:: The start of off-peak hours, expressed as an integer between 0 
and 23, inclusive.
+      Set to -1 to disable off-peak.
+Default:: -1 +
+  
+[[_hbase.offpeak.end.hour]]
+.hbase.offpeak.end.hour
+Description:: The end of off-peak hours, expressed as an integer between 0 and 
23, inclusive. Set
+      to -1 to disable off-peak.
+Default:: -1 +
+  
+[[_hbase.regionserver.thread.compaction.throttle]]
+.hbase.regionserver.thread.compaction.throttle
+Description:: There are two different thread pools for compactions, one for 
large compactions and
+      the other for small compactions. This helps to keep compaction of lean 
tables (such as
+        hbase:meta) fast. If a compaction is larger than this threshold, it
+      goes into the large compaction pool. In most cases, the default value is 
appropriate. Default:
+      2 x hbase.hstore.compaction.max x hbase.hregion.memstore.flush.size 
(which defaults to 128MB).
+      The value field assumes that the value of 
hbase.hregion.memstore.flush.size is unchanged from
+      the default.
+Default:: 2684354560 +
+  
+[[_hbase.hstore.compaction.kv.max]]
+.hbase.hstore.compaction.kv.max
+Description:: The maximum number of KeyValues to read and then write in a 
batch when flushing or
+      compacting. Set this lower if you have big KeyValues and problems with 
Out Of Memory
+      Exceptions Set this higher if you have wide, small rows. 
+Default:: 10 +
+  
+[[_hbase.storescanner.parallel.seek.enable]]
+.hbase.storescanner.parallel.seek.enable
+Description:: 
+      Enables StoreFileScanner parallel-seeking in StoreScanner,
+      a feature which can reduce response latency under special conditions.
+Default:: false +
+  
+[[_hbase.storescanner.parallel.seek.threads]]
+.hbase.storescanner.parallel.seek.threads
+Description:: 
+      The default thread pool size if parallel-seeking feature enabled.
+Default:: 10 +
+  
+[[_hfile.block.cache.size]]
+.hfile.block.cache.size
+Description:: Percentage of maximum heap (-Xmx setting) to allocate to block 
cache
+        used by a StoreFile. Default of 0.4 means allocate 40%.
+        Set to 0 to disable but it's not recommended; you need at least
+        enough cache to hold the storefile indices.
+Default:: 0.4 +
+  
+[[_hfile.block.index.cacheonwrite]]
+.hfile.block.index.cacheonwrite
+Description:: This allows to put non-root multi-level index blocks into the 
block
+          cache at the time the index is being written.
+Default:: false +
+  
+[[_hfile.index.block.max.size]]
+.hfile.index.block.max.size
+Description:: When the size of a leaf-level, intermediate-level, or root-level
+          index block in a multi-level block index grows to this size, the
+          block is written out and a new block is started.
+Default:: 131072 +
+  
+[[_hbase.bucketcache.ioengine]]
+.hbase.bucketcache.ioengine
+Description:: Where to store the contents of the bucketcache. One of: onheap, 
+      offheap, or file. If a file, set it to file:PATH_TO_FILE. See 
https://hbase.apache.org/apidocs/org/apache/hadoop/hbase/io/hfile/CacheConfig.html
 for more information.
+    
+Default::  +
+  
+[[_hbase.bucketcache.combinedcache.enabled]]
+.hbase.bucketcache.combinedcache.enabled
+Description:: Whether or not the bucketcache is used in league with the LRU 
+      on-heap block cache. In this mode, indices and blooms are kept in the 
LRU 
+      blockcache and the data blocks are kept in the bucketcache.
+Default:: true +
+  
+[[_hbase.bucketcache.size]]
+.hbase.bucketcache.size
+Description:: The size of the buckets for the bucketcache if you only use a 
single size. 
+      Defaults to the default blocksize, which is 64 * 1024.
+Default:: 65536 +
+  
+[[_hbase.bucketcache.sizes]]
+.hbase.bucketcache.sizes
+Description:: A comma-separated list of sizes for buckets for the bucketcache 
+      if you use multiple sizes. Should be a list of block sizes in order from 
smallest 
+      to largest. The sizes you use will depend on your data access patterns.
+Default::  +
+  
+[[_hfile.format.version]]
+.hfile.format.version
+Description:: The HFile format version to use for new files.
+      Version 3 adds support for tags in hfiles (See 
http://hbase.apache.org/book.html#hbase.tags).
+      Distributed Log Replay requires that tags are enabled. Also see the 
configuration
+      'hbase.replication.rpc.codec'. 
+      
+Default:: 3 +
+  
+[[_hfile.block.bloom.cacheonwrite]]
+.hfile.block.bloom.cacheonwrite
+Description:: Enables cache-on-write for inline blocks of a compound Bloom 
filter.
+Default:: false +
+  
+[[_io.storefile.bloom.block.size]]
+.io.storefile.bloom.block.size
+Description:: The size in bytes of a single block ("chunk") of a compound Bloom
+          filter. This size is approximate, because Bloom blocks can only be
+          inserted at data block boundaries, and the number of keys per data
+          block varies.
+Default:: 131072 +
+  
+[[_hbase.rs.cacheblocksonwrite]]
+.hbase.rs.cacheblocksonwrite
+Description:: Whether an HFile block should be added to the block cache when 
the
+          block is finished.
+Default:: false +
+  
+[[_hbase.rpc.timeout]]
+.hbase.rpc.timeout
+Description:: This is for the RPC layer to define how long HBase client 
applications
+        take for a remote call to time out. It uses pings to check connections
+        but will eventually throw a TimeoutException.
+Default:: 60000 +
+  
+[[_hbase.rpc.shortoperation.timeout]]
+.hbase.rpc.shortoperation.timeout
+Description:: This is another version of "hbase.rpc.timeout". For those RPC 
operation
+        within cluster, we rely on this configuration to set a short timeout 
limitation
+        for short operation. For example, short rpc timeout for region 
server's trying
+        to report to active master can benefit quicker master failover process.
+Default:: 10000 +
+  
+[[_hbase.ipc.client.tcpnodelay]]
+.hbase.ipc.client.tcpnodelay
+Description:: Set no delay on rpc socket connections.  See
+    
http://docs.oracle.com/javase/1.5.0/docs/api/java/net/Socket.html#getTcpNoDelay()
+Default:: true +
+  
+[[_hbase.master.keytab.file]]
+.hbase.master.keytab.file
+Description:: Full path to the kerberos keytab file to use for logging in
+    the configured HMaster server principal.
+Default::  +
+  
+[[_hbase.master.kerberos.principal]]
+.hbase.master.kerberos.principal
+Description:: Ex. "hbase/[email protected]".  The kerberos principal name
+    that should be used to run the HMaster process.  The principal name should
+    be in the form: user/hostname@DOMAIN.  If "_HOST" is used as the hostname
+    portion, it will be replaced with the actual hostname of the running
+    instance.
+Default::  +
+  
+[[_hbase.regionserver.keytab.file]]
+.hbase.regionserver.keytab.file
+Description:: Full path to the kerberos keytab file to use for logging in
+    the configured HRegionServer server principal.
+Default::  +
+  
+[[_hbase.regionserver.kerberos.principal]]
+.hbase.regionserver.kerberos.principal
+Description:: Ex. "hbase/[email protected]".  The kerberos principal name
+    that should be used to run the HRegionServer process.  The principal name
+    should be in the form: user/hostname@DOMAIN.  If "_HOST" is used as the
+    hostname portion, it will be replaced with the actual hostname of the
+    running instance.  An entry for this principal must exist in the file
+    specified in hbase.regionserver.keytab.file
+Default::  +
+  
+[[_hadoop.policy.file]]
+.hadoop.policy.file
+Description:: The policy configuration file used by RPC servers to make
+      authorization decisions on client requests.  Only used when HBase
+      security is enabled.
+Default:: hbase-policy.xml +
+  
+[[_hbase.superuser]]
+.hbase.superuser
+Description:: List of users or groups (comma-separated), who are allowed
+    full privileges, regardless of stored ACLs, across the cluster.
+    Only used when HBase security is enabled.
+Default::  +
+  
+[[_hbase.auth.key.update.interval]]
+.hbase.auth.key.update.interval
+Description:: The update interval for master key for authentication tokens
+    in servers in milliseconds.  Only used when HBase security is enabled.
+Default:: 86400000 +
+  
+[[_hbase.auth.token.max.lifetime]]
+.hbase.auth.token.max.lifetime
+Description:: The maximum lifetime in milliseconds after which an
+    authentication token expires.  Only used when HBase security is enabled.
+Default:: 604800000 +
+  
+[[_hbase.ipc.client.fallback-to-simple-auth-allowed]]
+.hbase.ipc.client.fallback-to-simple-auth-allowed
+Description:: When a client is configured to attempt a secure connection, but 
attempts to
+      connect to an insecure server, that server may instruct the client to
+      switch to SASL SIMPLE (unsecure) authentication. This setting controls
+      whether or not the client will accept this instruction from the server.
+      When false (the default), the client will not allow the fallback to 
SIMPLE
+      authentication, and will abort the connection.
+Default:: false +
+  
+[[_hbase.display.keys]]
+.hbase.display.keys
+Description:: When this is set to true the webUI and such will display all 
start/end keys
+                 as part of the table details, region names, etc. When this is 
set to false,
+                 the keys are hidden.
+Default:: true +
+  
+[[_hbase.coprocessor.region.classes]]
+.hbase.coprocessor.region.classes
+Description:: A comma-separated list of Coprocessors that are loaded by
+    default on all tables. For any override coprocessor method, these classes
+    will be called in order. After implementing your own Coprocessor, just put
+    it in HBase's classpath and add the fully qualified class name here.
+    A coprocessor can also be loaded on demand by setting HTableDescriptor.
+Default::  +
+  
+[[_hbase.rest.port]]
+.hbase.rest.port
+Description:: The port for the HBase REST server.
+Default:: 8080 +
+  
+[[_hbase.rest.readonly]]
+.hbase.rest.readonly
+Description:: Defines the mode the REST server will be started in. Possible 
values are:
+    false: All HTTP methods are permitted - GET/PUT/POST/DELETE.
+    true: Only the GET method is permitted.
+Default:: false +
+  
+[[_hbase.rest.threads.max]]
+.hbase.rest.threads.max
+Description:: The maximum number of threads of the REST server thread pool.
+        Threads in the pool are reused to process REST requests. This
+        controls the maximum number of requests processed concurrently.
+        It may help to control the memory used by the REST server to
+        avoid OOM issues. If the thread pool is full, incoming requests
+        will be queued up and wait for some free threads.
+Default:: 100 +
+  
+[[_hbase.rest.threads.min]]
+.hbase.rest.threads.min
+Description:: The minimum number of threads of the REST server thread pool.
+        The thread pool always has at least these number of threads so
+        the REST server is ready to serve incoming requests.
+Default:: 2 +
+  
+[[_hbase.rest.support.proxyuser]]
+.hbase.rest.support.proxyuser
+Description:: Enables running the REST server to support proxy-user mode.
+Default:: false +
+  
+[[_hbase.defaults.for.version.skip]]
+.hbase.defaults.for.version.skip
+Description:: Set to true to skip the 'hbase.defaults.for.version' check.
+    Setting this to true can be useful in contexts other than
+    the other side of a maven generation; i.e. running in an
+    ide.  You'll want to set this boolean to true to avoid
+    seeing the RuntimException complaint: "hbase-default.xml file
+    seems to be for and old version of HBase (\${hbase.version}), this
+    version is X.X.X-SNAPSHOT"
+Default:: false +
+  
+[[_hbase.coprocessor.master.classes]]
+.hbase.coprocessor.master.classes
+Description:: A comma-separated list of
+    org.apache.hadoop.hbase.coprocessor.MasterObserver coprocessors that are
+    loaded by default on the active HMaster process. For any implemented
+    coprocessor methods, the listed classes will be called in order. After
+    implementing your own MasterObserver, just put it in HBase's classpath
+    and add the fully qualified class name here.
+Default::  +
+  
+[[_hbase.coprocessor.abortonerror]]
+.hbase.coprocessor.abortonerror
+Description:: Set to true to cause the hosting server (master or regionserver)
+      to abort if a coprocessor fails to load, fails to initialize, or throws 
an
+      unexpected Throwable object. Setting this to false will allow the server 
to
+      continue execution but the system wide state of the coprocessor in 
question
+      will become inconsistent as it will be properly executing in only a 
subset
+      of servers, so this is most useful for debugging only.
+Default:: true +
+  
+[[_hbase.online.schema.update.enable]]
+.hbase.online.schema.update.enable
+Description:: Set true to enable online schema changes.
+Default:: true +
+  
+[[_hbase.table.lock.enable]]
+.hbase.table.lock.enable
+Description:: Set to true to enable locking the table in zookeeper for schema 
change operations.
+    Table locking from master prevents concurrent schema modifications to 
corrupt table
+    state.
+Default:: true +
+  
+[[_hbase.table.max.rowsize]]
+.hbase.table.max.rowsize
+Description:: 
+      Maximum size of single row in bytes (default is 1 Gb) for Get'ting
+      or Scan'ning without in-row scan flag set. If row size exceeds this limit
+      RowTooBigException is thrown to client.
+    
+Default:: 1073741824 +
+  
+[[_hbase.thrift.minWorkerThreads]]
+.hbase.thrift.minWorkerThreads
+Description:: The "core size" of the thread pool. New threads are created on 
every
+    connection until this many threads are created.
+Default:: 16 +
+  
+[[_hbase.thrift.maxWorkerThreads]]
+.hbase.thrift.maxWorkerThreads
+Description:: The maximum size of the thread pool. When the pending request 
queue
+    overflows, new threads are created until their number reaches this number.
+    After that, the server starts dropping connections.
+Default:: 1000 +
+  
+[[_hbase.thrift.maxQueuedRequests]]
+.hbase.thrift.maxQueuedRequests
+Description:: The maximum number of pending Thrift connections waiting in the 
queue. If
+     there are no idle threads in the pool, the server queues requests. Only
+     when the queue overflows, new threads are added, up to
+     hbase.thrift.maxQueuedRequests threads.
+Default:: 1000 +
+  
+[[_hbase.thrift.htablepool.size.max]]
+.hbase.thrift.htablepool.size.max
+Description:: The upper bound for the table pool used in the Thrift gateways 
server.
+      Since this is per table name, we assume a single table and so with 1000 
default
+      worker threads max this is set to a matching number. For other workloads 
this number
+      can be adjusted as needed.
+    
+Default:: 1000 +
+  
+[[_hbase.regionserver.thrift.framed]]
+.hbase.regionserver.thrift.framed
+Description:: Use Thrift TFramedTransport on the server side.
+      This is the recommended transport for thrift servers and requires a 
similar setting
+      on the client side. Changing this to false will select the default 
transport,
+      vulnerable to DoS when malformed requests are issued due to THRIFT-601.
+    
+Default:: false +
+  
+[[_hbase.regionserver.thrift.framed.max_frame_size_in_mb]]
+.hbase.regionserver.thrift.framed.max_frame_size_in_mb
+Description:: Default frame size when using framed transport
+Default:: 2 +
+  
+[[_hbase.regionserver.thrift.compact]]
+.hbase.regionserver.thrift.compact
+Description:: Use Thrift TCompactProtocol binary serialization protocol.
+Default:: false +
+  
+[[_hbase.data.umask.enable]]
+.hbase.data.umask.enable
+Description:: Enable, if true, that file permissions should be assigned
+      to the files written by the regionserver
+Default:: false +
+  
+[[_hbase.data.umask]]
+.hbase.data.umask
+Description:: File permissions that should be used to write data
+      files when hbase.data.umask.enable is true
+Default:: 000 +
+  
+[[_hbase.metrics.showTableName]]
+.hbase.metrics.showTableName
+Description:: Whether to include the prefix "tbl.tablename" in per-column 
family metrics.
+       If true, for each metric M, per-cf metrics will be reported for 
tbl.T.cf.CF.M, if false,
+       per-cf metrics will be aggregated by column-family across tables, and 
reported for cf.CF.M.
+       In both cases, the aggregated metric M across tables and cfs will be 
reported.
+Default:: true +
+  
+[[_hbase.metrics.exposeOperationTimes]]
+.hbase.metrics.exposeOperationTimes
+Description:: Whether to report metrics about time taken performing an
+      operation on the region server.  Get, Put, Delete, Increment, and Append 
can all
+      have their times exposed through Hadoop metrics per CF and per region.
+Default:: true +
+  
+[[_hbase.snapshot.enabled]]
+.hbase.snapshot.enabled
+Description:: Set to true to allow snapshots to be taken / restored / cloned.
+Default:: true +
+  
+[[_hbase.snapshot.restore.take.failsafe.snapshot]]
+.hbase.snapshot.restore.take.failsafe.snapshot
+Description:: Set to true to take a snapshot before the restore operation.
+      The snapshot taken will be used in case of failure, to restore the 
previous state.
+      At the end of the restore operation this snapshot will be deleted
+Default:: true +
+  
+[[_hbase.snapshot.restore.failsafe.name]]
+.hbase.snapshot.restore.failsafe.name
+Description:: Name of the failsafe snapshot taken by the restore operation.
+      You can use the {snapshot.name}, {table.name} and {restore.timestamp} 
variables
+      to create a name based on what you are restoring.
+Default:: hbase-failsafe-{snapshot.name}-{restore.timestamp} +
+  
+[[_hbase.server.compactchecker.interval.multiplier]]
+.hbase.server.compactchecker.interval.multiplier
+Description:: The number that determines how often we scan to see if 
compaction is necessary.
+        Normally, compactions are done after some events (such as memstore 
flush), but if
+        region didn't receive a lot of writes for some time, or due to 
different compaction
+        policies, it may be necessary to check it periodically. The interval 
between checks is
+        hbase.server.compactchecker.interval.multiplier multiplied by
+        hbase.server.thread.wakefrequency.
+Default:: 1000 +
+  
+[[_hbase.lease.recovery.timeout]]
+.hbase.lease.recovery.timeout
+Description:: How long we wait on dfs lease recovery in total before giving up.
+Default:: 900000 +
+  
+[[_hbase.lease.recovery.dfs.timeout]]
+.hbase.lease.recovery.dfs.timeout
+Description:: How long between dfs recover lease invocations. Should be larger 
than the sum of
+        the time it takes for the namenode to issue a block recovery command 
as part of
+        datanode; dfs.heartbeat.interval and the time it takes for the primary
+        datanode, performing block recovery to timeout on a dead datanode; 
usually
+        dfs.client.socket-timeout. See the end of HBASE-8389 for more.
+Default:: 64000 +
+  
+[[_hbase.column.max.version]]
+.hbase.column.max.version
+Description:: New column family descriptors will use this value as the default 
number of versions
+      to keep.
+Default:: 1 +
+  
+[[_hbase.dfs.client.read.shortcircuit.buffer.size]]
+.hbase.dfs.client.read.shortcircuit.buffer.size
+Description:: If the DFSClient configuration
+    dfs.client.read.shortcircuit.buffer.size is unset, we will
+    use what is configured here as the short circuit read default
+    direct byte buffer size. DFSClient native default is 1MB; HBase
+    keeps its HDFS files open so number of file blocks * 1MB soon
+    starts to add up and threaten OOME because of a shortage of
+    direct memory.  So, we set it down from the default.  Make
+    it > the default hbase block size set in the HColumnDescriptor
+    which is usually 64k.
+    
+Default:: 131072 +
+  
+[[_hbase.regionserver.checksum.verify]]
+.hbase.regionserver.checksum.verify
+Description:: 
+        If set to true (the default), HBase verifies the checksums for hfile
+        blocks. HBase writes checksums inline with the data when it writes out
+        hfiles. HDFS (as of this writing) writes checksums to a separate file
+        than the data file necessitating extra seeks.  Setting this flag saves
+        some on i/o.  Checksum verification by HDFS will be internally disabled
+        on hfile streams when this flag is set.  If the hbase-checksum 
verification
+        fails, we will switch back to using HDFS checksums (so do not disable 
HDFS
+        checksums!  And besides this feature applies to hfiles only, not to 
WALs).
+        If this parameter is set to false, then hbase will not verify any 
checksums,
+        instead it will depend on checksum verification being done in the HDFS 
client.  
+    
+Default:: true +
+  
+[[_hbase.hstore.bytes.per.checksum]]
+.hbase.hstore.bytes.per.checksum
+Description:: 
+        Number of bytes in a newly created checksum chunk for HBase-level
+        checksums in hfile blocks.
+    
+Default:: 16384 +
+  
+[[_hbase.hstore.checksum.algorithm]]
+.hbase.hstore.checksum.algorithm
+Description:: 
+      Name of an algorithm that is used to compute checksums. Possible values
+      are NULL, CRC32, CRC32C.
+    
+Default:: CRC32 +
+  
+[[_hbase.status.published]]
+.hbase.status.published
+Description:: 
+      This setting activates the publication by the master of the status of 
the region server.
+      When a region server dies and its recovery starts, the master will push 
this information
+      to the client application, to let them cut the connection immediately 
instead of waiting
+      for a timeout.
+    
+Default:: false +
+  
+[[_hbase.status.publisher.class]]
+.hbase.status.publisher.class
+Description:: 
+      Implementation of the status publication with a multicast message.
+    
+Default:: 
org.apache.hadoop.hbase.master.ClusterStatusPublisher$MulticastPublisher +
+  
+[[_hbase.status.listener.class]]
+.hbase.status.listener.class
+Description:: 
+      Implementation of the status listener with a multicast message.
+    
+Default:: 
org.apache.hadoop.hbase.client.ClusterStatusListener$MulticastListener +
+  
+[[_hbase.status.multicast.address.ip]]
+.hbase.status.multicast.address.ip
+Description:: 
+      Multicast address to use for the status publication by multicast.
+    
+Default:: 226.1.1.3 +
+  
+[[_hbase.status.multicast.address.port]]
+.hbase.status.multicast.address.port
+Description:: 
+      Multicast port to use for the status publication by multicast.
+    
+Default:: 16100 +
+  
+[[_hbase.dynamic.jars.dir]]
+.hbase.dynamic.jars.dir
+Description:: 
+      The directory from which the custom filter/co-processor jars can be 
loaded
+      dynamically by the region server without the need to restart. However,
+      an already loaded filter/co-processor class would not be un-loaded. See
+      HBASE-1936 for more details.
+    
+Default:: ${hbase.rootdir}/lib +
+  
+[[_hbase.security.authentication]]
+.hbase.security.authentication
+Description:: 
+      Controls whether or not secure authentication is enabled for HBase.
+      Possible values are 'simple' (no authentication), and 'kerberos'.
+    
+Default:: simple +
+  
+[[_hbase.rest.filter.classes]]
+.hbase.rest.filter.classes
+Description:: 
+      Servlet filters for REST service.
+    
+Default:: org.apache.hadoop.hbase.rest.filter.GzipFilter +
+  
+[[_hbase.master.loadbalancer.class]]
+.hbase.master.loadbalancer.class
+Description:: 
+      Class used to execute the regions balancing when the period occurs.
+      See the class comment for more on how it works
+      
http://hbase.apache.org/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.html
+      It replaces the DefaultLoadBalancer as the default (since renamed
+      as the SimpleLoadBalancer).
+    
+Default:: org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer +
+  
+[[_hbase.security.exec.permission.checks]]
+.hbase.security.exec.permission.checks
+Description:: 
+      If this setting is enabled and ACL based access control is active (the
+      AccessController coprocessor is installed either as a system coprocessor
+      or on a table as a table coprocessor) then you must grant all relevant
+      users EXEC privilege if they require the ability to execute coprocessor
+      endpoint calls. EXEC privilege, like any other permission, can be
+      granted globally to a user, or to a user on a per table or per namespace
+      basis. For more information on coprocessor endpoints, see the coprocessor
+      section of the HBase online manual. For more information on granting or
+      revoking permissions using the AccessController, see the security
+      section of the HBase online manual.
+    
+Default:: false +
+  
+[[_hbase.procedure.regionserver.classes]]
+.hbase.procedure.regionserver.classes
+Description:: A comma-separated list of 
+    org.apache.hadoop.hbase.procedure.RegionServerProcedureManager procedure 
managers that are 
+    loaded by default on the active HRegionServer process. The lifecycle 
methods (init/start/stop) 
+    will be called by the active HRegionServer process to perform the specific 
globally barriered 
+    procedure. After implementing your own RegionServerProcedureManager, just 
put it in 
+    HBase's classpath and add the fully qualified class name here.
+    
+Default::  +
+  
+[[_hbase.procedure.master.classes]]
+.hbase.procedure.master.classes
+Description:: A comma-separated list of
+    org.apache.hadoop.hbase.procedure.MasterProcedureManager procedure 
managers that are
+    loaded by default on the active HMaster process. A procedure is identified 
by its signature and
+    users can use the signature and an instant name to trigger an execution of 
a globally barriered
+    procedure. After implementing your own MasterProcedureManager, just put it 
in HBase's classpath
+    and add the fully qualified class name here.
+Default::  +
+  
+[[_hbase.coordinated.state.manager.class]]
+.hbase.coordinated.state.manager.class
+Description:: Fully qualified name of class implementing coordinated state 
manager.
+Default:: org.apache.hadoop.hbase.coordination.ZkCoordinatedStateManager +
+  
+[[_hbase.regionserver.storefile.refresh.period]]
+.hbase.regionserver.storefile.refresh.period
+Description:: 
+      The period (in milliseconds) for refreshing the store files for the 
secondary regions. 0
+      means this feature is disabled. Secondary regions sees new files (from 
flushes and
+      compactions) from primary once the secondary region refreshes the list 
of files in the
+      region (there is no notification mechanism). But too frequent refreshes 
might cause
+      extra Namenode pressure. If the files cannot be refreshed for longer 
than HFile TTL
+      (hbase.master.hfilecleaner.ttl) the requests are rejected. Configuring 
HFile TTL to a larger
+      value is also recommended with this setting.
+    
+Default:: 0 +
+  
+[[_hbase.region.replica.replication.enabled]]
+.hbase.region.replica.replication.enabled
+Description:: 
+      Whether asynchronous WAL replication to the secondary region replicas is 
enabled or not.
+      If this is enabled, a replication peer named 
"region_replica_replication" will be created
+      which will tail the logs and replicate the mutatations to region 
replicas for tables that
+      have region replication > 1. If this is enabled once, disabling this 
replication also
+      requires disabling the replication peer using shell or ReplicationAdmin 
java class.
+      Replication to secondary region replicas works over standard 
inter-cluster replication. 
+      So replication, if disabled explicitly, also has to be enabled by 
setting "hbase.replication" 
+      to true for this feature to work.
+    
+Default:: false +
+  
+[[_hbase.http.filter.initializers]]
+.hbase.http.filter.initializers
+Description:: 
+      A comma separated list of class names. Each class in the list must 
extend 
+      org.apache.hadoop.hbase.http.FilterInitializer. The corresponding Filter 
will 
+      be initialized. Then, the Filter will be applied to all user facing jsp 
+      and servlet web pages. 
+      The ordering of the list defines the ordering of the filters.
+      The default StaticUserWebFilter add a user principal as defined by the 
+      hbase.http.staticuser.user property.
+    
+Default:: org.apache.hadoop.hbase.http.lib.StaticUserWebFilter +
+  
+[[_hbase.security.visibility.mutations.checkauths]]
+.hbase.security.visibility.mutations.checkauths
+Description:: 
+      This property if enabled, will check whether the labels in the 
visibility expression are associated
+      with the user issuing the mutation
+    
+Default:: false +
+  
+[[_hbase.http.max.threads]]
+.hbase.http.max.threads
+Description:: 
+      The maximum number of threads that the HTTP Server will create in its 
+      ThreadPool.
+    
+Default:: 10 +
+  
+[[_hbase.replication.rpc.codec]]
+.hbase.replication.rpc.codec
+Description:: 
+               The codec that is to be used when replication is enabled so that
+               the tags are also replicated. This is used along with HFileV3 
which 
+               supports tags in them.  If tags are not used or if the hfile 
version used
+               is HFileV2 then KeyValueCodec can be used as the replication 
codec. Note that
+               using KeyValueCodecWithTags for replication when there are no 
tags causes no harm.
+       
+Default:: org.apache.hadoop.hbase.codec.KeyValueCodecWithTags +
+  
+[[_hbase.http.staticuser.user]]
+.hbase.http.staticuser.user
+Description:: 
+      The user name to filter as, on static web filters
+      while rendering content. An example use is the HDFS
+      web UI (user to be used for browsing files).
+    
+Default:: dr.stack +
+  
+[[_hbase.regionserver.handler.abort.on.error.percent]]
+.hbase.regionserver.handler.abort.on.error.percent
+Description:: The percent of region server RPC threads failed to abort RS.
+    -1 Disable aborting; 0 Abort if even a single handler has died;
+    0.x Abort only when this percent of handlers have died;
+    1 Abort only all of the handers have died.
+Default:: 0.5 +
+  
\ No newline at end of file

Reply via email to