jon-wei commented on a change in pull request #6127: Unified configuration doc
page
URL: https://github.com/apache/incubator-druid/pull/6127#discussion_r208771254
##########
File path: docs/content/configuration/configuration-reference.md
##########
@@ -0,0 +1,1512 @@
+---
+layout: doc_page
+---
+
+# Configuration Reference
+
+This page documents all of the configuration properties for each Druid service
type.
+
+## Table of Contents
+ * [Common configurations](#common-configurations)
+ * [Extensions](#extensions)
+ * [Modules](#modules)
+ * [Zookeeper](#zookeper)
+ * [Exhibitor](#exhibitor)
+ * [TLS](#tls)
+ * [Authentication & Authorization](#authentication-and-authorization)
+ * [Startup Logging](#startup-logging)
+ * [Request Logging](#request-logging)
+ * [Enabling Metrics](#enabling-metrics)
+ * [Emitting Metrics](#emitting-metrics)
+ * [Metadata Storage](#metadata-storage)
+ * [Deep Storage](#deep-storage)
+ * [Task Logging](#task-logging)
+ * [Indexing Service Discovery](#indexing-service-discovery)
+ * [Coordinator Discovery](#coordinator-discovery)
+ * [Announcing Segments](#announcing-segments)
+ * [JavaScript](#javascript)
+ * [Double Column Storage](#double-column-storage)
+ * [Coordinator](#coordinator)
+ * [Static Configuration](#static-configuration)
+ * [Node Config](#coordinator-node-config)
+ * [Coordinator Operation](#coordinator-operation)
+ * [Segment Management](#segment-management)
+ * [Metadata Retrieval](#metadata-retrieval)
+ * [Dynamic Configuration](#dynamic-configuration)
+ * [Lookups](#lookups-dynamic-configuration)
+ * [Compaction](#compaction-dynamic-configuration)
+ * [Overlord](#overlord)
+ * [Node Config](#overlord-node-config)
+ * [Static Configuration](#overlord-static-configuration)
+ * [Dynamic Configuration](#overlord-dynamic-configuration)
+ * [Worker Select Strategy](#worker-select-strategy)
+ * [Autoscaler](#autoscaler)
+ * [MiddleManager & Peons](#middle-manager-and-peons)
+ * [Node Config](#middlemanager-node-config)
+ * [MiddleManger Configuration](#middlemanager-configuration)
+ * [Peon Processing](#peon-processing)
+ * [Peon Query Configuration](#peon-query-configuration)
+ * [Caching](#peon-caching)
+ * [Additional Peon Configuration](#additional-peon-configuration)
+ * [Broker](#broker)
+ * [Node Config](#broker-node-configs)
+ * [Query Configuration](#broker-query-configuration)
+ * [SQL](#sql)
+ * [Caching](#broker-caching)
+ * [Segment Discovery](#segment-discovery)
+ * [Historical](#historical)
+ * [Node Configuration](#historical-node-config)
+ * [General Configuration](#historical-general-configuration)
+ * [Query Configs](#historical-query-configs)
+ * [Caching](#historical-caching)
+ * [Caching](#cache-configuration)
+ * [General Query Configuration](#general-query-configuration)
+
+## Common Configurations
+
+The properties under this section are common configurations that should be
shared across all Druid services in a cluster.
+
+### JVM Configuration Best Practices
+
+There are four JVM parameters that we set on all of our processes:
+
+1. `-Duser.timezone=UTC` This sets the default timezone of the JVM to UTC. We
always set this and do not test with other default timezones, so local
timezones might work, but they also might uncover weird and interesting bugs.
To issue queries in a non-UTC timezone, see [query
granularities](../querying/granularities.html#period-granularities)
+2. `-Dfile.encoding=UTF-8` This is similar to timezone, we test assuming
UTF-8. Local encodings might work, but they also might result in weird and
interesting bugs.
+3. `-Djava.io.tmpdir=<a path>` Various parts of the system that interact with
the file system do it via temporary files, and these files can get somewhat
large. Many production systems are set up to have small (but fast) `/tmp`
directories, which can be problematic with Druid so we recommend pointing the
JVM’s tmp directory to something with a little more meat.
+4. `-Djava.util.logging.manager=org.apache.logging.log4j.jul.LogManager` This
allows log4j2 to handle logs for non-log4j2 components (like jetty) which use
standard java logging.
+
+### Extensions
+
+Many of Druid's external dependencies can be plugged in as modules. Extensions
can be provided using the following configs:
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.extensions.directory`|The root extension directory where user can put
extensions related files. Druid will load extensions stored under this
directory.|`extensions` (This is a relative path to Druid's working directory)|
+|`druid.extensions.hadoopDependenciesDir`|The root hadoop dependencies
directory where user can put hadoop related dependencies files. Druid will load
the dependencies based on the hadoop coordinate specified in the hadoop index
task.|`hadoop-dependencies` (This is a relative path to Druid's working
directory|
+|`druid.extensions.loadList`|A JSON array of extensions to load from extension
directories by Druid. If it is not specified, its value will be `null` and
Druid will load all the extensions under `druid.extensions.directory`. If its
value is empty list `[]`, then no extensions will be loaded at all. It is also
allowed to specify absolute path of other custom extensions not stored in the
common extensions directory.|null|
+|`druid.extensions.searchCurrentClassloader`|This is a boolean flag that
determines if Druid will search the main classloader for extensions. It
defaults to true but can be turned off if you have reason to not automatically
add all modules on the classpath.|true|
+|`druid.extensions.useExtensionClassloaderFirst`|This is a boolean flag that
determines if Druid extensions should prefer loading classes from their own
jars rather than jars bundled with Druid. If false, extensions must be
compatible with classes provided by any jars bundled with Druid. If true,
extensions may depend on conflicting versions.|false|
+|`druid.extensions.hadoopContainerDruidClasspath`|Hadoop Indexing launches
hadoop jobs and this configuration provides way to explicitly set the user
classpath for the hadoop job. By default this is computed automatically by
druid based on the druid process classpath and set of extensions. However,
sometimes you might want to be explicit to resolve dependency conflicts between
druid and hadoop.|null|
+|`druid.extensions.addExtensionsToHadoopContainer`|Only applicable if
`druid.extensions.hadoopContainerDruidClasspath` is provided. If set to true,
then extensions specified in the loadList are added to hadoop container
classpath. Note that when `druid.extensions.hadoopContainerDruidClasspath` is
not provided then extensions are always added to hadoop container
classpath.|false|
+
+### Modules
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.modules.excludeList`|A JSON array of canonical class names (e. g.
`"io.druid.somepackage.SomeModule"`) of module classes which shouldn't be
loaded, even if they are found in extensions specified by
`druid.extensions.loadList`, or in the list of core modules specified to be
loaded on a particular Druid node type. Useful when some useful extension
contains some module, which shouldn't be loaded on some Druid node type because
some dependencies of that module couldn't be satisfied.|[]|
+
+### Zookeeper
+We recommend just setting the base ZK path and the ZK service host, but all ZK
paths that Druid uses can be overwritten to absolute paths.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.zk.paths.base`|Base Zookeeper path.|`/druid`|
+|`druid.zk.service.host`|The ZooKeeper hosts to connect to. This is a REQUIRED
property and therefore a host address must be supplied.|none|
+|`druid.zk.service.user`|The username to authenticate with ZooKeeper. This is
an optional property.|none|
+|`druid.zk.service.pwd`|The [Password
Provider](../operations/password-provider.html) or the string password to
authenticate with ZooKeeper. This is an optional property.|none|
+|`druid.zk.service.authScheme`|digest is the only authentication scheme
supported. |digest|
+
+#### Zookeeper Behavior
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.zk.service.sessionTimeoutMs`|ZooKeeper session timeout, in
milliseconds.|`30000`|
+|`druid.zk.service.compress`|Boolean flag for whether or not created Znodes
should be compressed.|`true`|
+|`druid.zk.service.acl`|Boolean flag for whether or not to enable ACL security
for ZooKeeper. If ACL is enabled, zNode creators will have all
permissions.|`false`|
+
+#### Path Configuration
+Druid interacts with ZK through a set of standard path configurations. We
recommend just setting the base ZK path, but all ZK paths that Druid uses can
be overwritten to absolute paths.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.zk.paths.base`|Base Zookeeper path.|`/druid`|
+|`druid.zk.paths.propertiesPath`|Zookeeper properties
path.|`${druid.zk.paths.base}/properties`|
+|`druid.zk.paths.announcementsPath`|Druid node announcement
path.|`${druid.zk.paths.base}/announcements`|
+|`druid.zk.paths.liveSegmentsPath`|Current path for where Druid nodes announce
their segments.|`${druid.zk.paths.base}/segments`|
+|`druid.zk.paths.loadQueuePath`|Entries here cause historical nodes to load
and drop segments.|`${druid.zk.paths.base}/loadQueue`|
+|`druid.zk.paths.coordinatorPath`|Used by the coordinator for leader
election.|`${druid.zk.paths.base}/coordinator`|
+|`druid.zk.paths.servedSegmentsPath`|@Deprecated. Legacy path for where Druid
nodes announce their segments.|`${druid.zk.paths.base}/servedSegments`|
+
+The indexing service also uses its own set of paths. These configs can be
included in the common configuration.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.zk.paths.indexer.base`|Base zookeeper path for
|`${druid.zk.paths.base}/indexer`|
+|`druid.zk.paths.indexer.announcementsPath`|Middle managers announce
themselves here.|`${druid.zk.paths.indexer.base}/announcements`|
+|`druid.zk.paths.indexer.tasksPath`|Used to assign tasks to middle
managers.|`${druid.zk.paths.indexer.base}/tasks`|
+|`druid.zk.paths.indexer.statusPath`|Parent path for announcement of task
statuses.|`${druid.zk.paths.indexer.base}/status`|
+
+If `druid.zk.paths.base` and `druid.zk.paths.indexer.base` are both set, and
none of the other `druid.zk.paths.*` or `druid.zk.paths.indexer.*` values are
set, then the other properties will be evaluated relative to their respective
`base`.
+For example, if `druid.zk.paths.base` is set to `/druid1` and
`druid.zk.paths.indexer.base` is set to `/druid2` then
`druid.zk.paths.announcementsPath` will default to `/druid1/announcements`
while `druid.zk.paths.indexer.announcementsPath` will default to
`/druid2/announcements`.
+
+The following path is used for service discovery. It is **not** affected by
`druid.zk.paths.base` and **must** be specified separately.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.discovery.curator.path`|Services announce themselves under this
ZooKeeper path.|`/druid/discovery`|
+
+### Exhibitor
+
+[Exhibitor](https://github.com/Netflix/exhibitor/wiki) is a supervisor system
for ZooKeeper.
+Exhibitor can dynamically scale-up/down the cluster of ZooKeeper servers.
+Druid can update self-owned list of ZooKeeper servers through Exhibitor
without restarting.
+That is, it allows Druid to keep the connections of Exhibitor-supervised
ZooKeeper servers.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.exhibitor.service.hosts`|A JSON array which contains the hostnames of
Exhibitor instances. Please specify this property if you want to use
Exhibitor-supervised cluster.|none|
+|`druid.exhibitor.service.port`|The REST port used to connect to
Exhibitor.|`8080`|
+|`druid.exhibitor.service.restUriPath`|The path of the REST call used to get
the server set.|`/exhibitor/v1/cluster/list`|
+|`druid.exhibitor.service.useSsl`|Boolean flag for whether or not to use https
protocol.|`false`|
+|`druid.exhibitor.service.pollingMs`|How ofter to poll the exhibitors for the
list|`10000`|
+
+Note that `druid.zk.service.host` is used as a backup in case an Exhibitor
instance can't be contacted and therefore should still be set.
+
+### TLS
+
+#### General Configuration
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.enablePlaintextPort`|Enable/Disable HTTP connector.|`true`|
+|`druid.enableTlsPort`|Enable/Disable HTTPS connector.|`false`|
+
+Although not recommended but both HTTP and HTTPS connectors can be enabled at
a time and respective ports are configurable using `druid.plaintextPort`
+and `druid.tlsPort` properties on each node. Please see `Configuration`
section of individual nodes to check the valid and default values for these
ports.
+
+#### Jetty Server TLS Configuration
+
+Druid uses Jetty as an embedded web server. To get familiar with TLS/SSL in
general and related concepts like Certificates etc.
+reading this [Jetty
documentation](http://www.eclipse.org/jetty/documentation/9.3.x/configuring-ssl.html)
might be helpful.
+To get more in depth knowledge of TLS/SSL support in Java in general, please
refer to this
[guide](http://docs.oracle.com/javase/8/docs/technotes/guides/security/jsse/JSSERefGuide.html).
+The documentation
[here](http://www.eclipse.org/jetty/documentation/9.3.x/configuring-ssl.html#configuring-sslcontextfactory)
+can help in understanding TLS/SSL configurations listed below. This
[document](http://docs.oracle.com/javase/8/docs/technotes/guides/security/StandardNames.html)
lists all the possible
+values for the below mentioned configs among others provided by Java
implementation.
+
+|Property|Description|Default|Required|
+|--------|-----------|-------|--------|
+|`druid.server.https.keyStorePath`|The file path or URL of the TLS/SSL Key
store.|none|yes|
+|`druid.server.https.keyStoreType`|The type of the key store.|none|yes|
+|`druid.server.https.certAlias`|Alias of TLS/SSL certificate for the
connector.|none|yes|
+|`druid.server.https.keyStorePassword`|The [Password
Provider](../operations/password-provider.html) or String password for the Key
Store.|none|yes|
+
+Following table contains non-mandatory advanced configuration options, use
caution.
+
+|Property|Description|Default|Required|
+|--------|-----------|-------|--------|
+|`druid.server.https.keyManagerFactoryAlgorithm`|Algorithm to use for creating
KeyManager, more details
[here](https://docs.oracle.com/javase/7/docs/technotes/guides/security/jsse/JSSERefGuide.html#KeyManager).|`javax.net.ssl.KeyManagerFactory.getDefaultAlgorithm()`|no|
+|`druid.server.https.keyManagerPassword`|The [Password
Provider](../operations/password-provider.html) or String password for the Key
Manager.|none|no|
+|`druid.server.https.includeCipherSuites`|List of cipher suite names to
include. You can either use the exact cipher suite name or a regular
expression.|Jetty's default include cipher list|no|
+|`druid.server.https.excludeCipherSuites`|List of cipher suite names to
exclude. You can either use the exact cipher suite name or a regular
expression.|Jetty's default exclude cipher list|no|
+|`druid.server.https.includeProtocols`|List of exact protocols names to
include.|Jetty's default include protocol list|no|
+|`druid.server.https.excludeProtocols`|List of exact protocols names to
exclude.|Jetty's default exclude protocol list|no|
+
+
+#### Internal Client TLS Configuration (requires `simple-client-sslcontext`
extension)
+
+These properties apply to the SSLContext that will be provided to the internal
HTTP client that Druid services use to communicate with each other. These
properties require the `simple-client-sslcontext` extension to be loaded.
Without it, Druid services will be unable to communicate with each other when
TLS is enabled.
+
+|Property|Description|Default|Required|
+|--------|-----------|-------|--------|
+|`druid.client.https.protocol`|SSL protocol to use.|`TLSv1.2`|no|
+|`druid.client.https.trustStoreType`|The type of the key store where trusted
root certificates are stored.|`java.security.KeyStore.getDefaultType()`|no|
+|`druid.client.https.trustStorePath`|The file path or URL of the TLS/SSL Key
store where trusted root certificates are stored.|none|yes|
+|`druid.client.https.trustStoreAlgorithm`|Algorithm to be used by TrustManager
to validate certificate
chains|`javax.net.ssl.TrustManagerFactory.getDefaultAlgorithm()`|no|
+|`druid.client.https.trustStorePassword`|The [Password
Provider](../../operations/password-provider.html) or String password for the
Trust Store.|none|yes|
+
+This
[document](http://docs.oracle.com/javase/8/docs/technotes/guides/security/StandardNames.html)
lists all the possible
+values for the above mentioned configs among others provided by Java
implementation.
+
+### Authentication and Authorization
+
+|Property|Type|Description|Default|Required|
+|--------|-----------|--------|--------|--------|
+|`druid.auth.authenticatorChain`|JSON List of Strings|List of Authenticator
type names|["allowAll"]|no|
+|`druid.escalator.type`|String|Type of the Escalator that should be used for
internal Druid communications. This Escalator must use an authentication scheme
that is supported by an Authenticator in
`druid.auth.authenticationChain`.|"noop"|no|
+|`druid.auth.authorizers`|JSON List of Strings|List of Authorizer type names
|["allowAll"]|no|
+|`druid.auth.unsecuredPaths`| List of Strings|List of paths for which security
checks will not be performed. All requests to these paths will be
allowed.|[]|no|
+|`druid.auth.allowUnauthenticatedHttpOptions`|Boolean|If true, skip
authentication checks for HTTP OPTIONS requests. This is needed for certain use
cases, such as supporting CORS pre-flight requests. Note that disabling
authentication checks for OPTIONS requests will allow unauthenticated users to
determine what Druid endpoints are valid (by checking if the OPTIONS request
returns a 200 instead of 404), so enabling this option may reveal information
about server configuration, including information about what extensions are
loaded (if those extensions add endpoints).|false|no|
+
+For more information, please see [Authentication and
Authorization](../design/auth.html).
+
+For configuration options for specific auth extensions, please refer to the
extension documentation.
+
+### Startup Logging
+
+All nodes can log debugging information on startup.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.startup.logging.logProperties`|Log all properties on startup (from
common.runtime.properties, runtime.properties, and the JVM command line).|false|
+|`druid.startup.logging.maskProperties`|Masks sensitive properties (passwords,
for example) containing theses words.|["password"]|
+
+Note that some sensitive information may be logged if these settings are
enabled.
+
+### Request Logging
+
+All nodes that can serve queries can also log the query requests they see.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.request.logging.type`|Choices: noop, file, emitter, slf4j, filtered,
composing. How to log every query request.|noop|
+
+Note that, you can enable sending all the HTTP requests to log by setting
"io.druid.jetty.RequestLog" to DEBUG level. See
[Logging](../configuration/logging.html)
+
+#### File Request Logging
+
+Daily request logs are stored on disk.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.request.logging.dir`|Historical, Realtime and Broker nodes maintain
request logs of all of the requests they get (interacton is via POST, so normal
request logs don’t generally capture information about the actual query), this
specifies the directory to store the request logs in|none|
+
+#### Emitter Request Logging
+
+Every request is emitted to some external location.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.request.logging.feed`|Feed name for requests.|none|
+
+#### SLF4J Request Logging
+
+Every request is logged via SLF4J. Queries are serialized into JSON in the log
message regardless of the SJF4J format specification. They will be logged under
the class `io.druid.server.log.LoggingRequestLogger`.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.request.logging.setMDC`|If MDC entries should be set in the log entry.
Your logging setup still has to be configured to handle MDC to format this
data|false|
+|`druid.request.logging.setContextMDC`|If the druid query `context` should be
added to the MDC entries. Has no effect unless `setMDC` is `true`|false|
+
+MDC fields populated with `setMDC`:
+
+|MDC field|Description|
+|---------|-----------|
+|`queryId` |The query ID|
+|`dataSource`|The datasource the query was against|
+|`queryType` |The type of the query|
+|`hasFilters`|If the query has any filters|
+|`remoteAddr`|The remote address of the requesting client|
+|`duration` |The duration of the query interval|
+|`resultOrdering`|The ordering of results|
+|`descending`|If the query is a descending query|
+
+#### Filtered Request Logging
+Filtered Request Logger filters requests based on a configurable query/time
threshold. Only request logs where query/time is above the threshold are
emitted.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.request.logging.queryTimeThresholdMs`|Threshold value for query/time
in milliseconds.|0 i.e no filtering|
+|`druid.request.logging.delegate.type`|Type of delegate request logger to log
requests.|none|
+
+#### Composite Request Logging
+Composite Request Logger emits request logs to multiple request loggers.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.request.logging.loggerProviders`|List of request loggers for emitting
request logs.|none|
+
+### Enabling Metrics
+
+Druid nodes periodically emit metrics and different metrics monitors can be
included. Each node can overwrite the default list of monitors.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.monitoring.emissionPeriod`|How often metrics are emitted.|PT1m|
+|`druid.monitoring.monitors`|Sets list of Druid monitors used by a node. See
below for names and more information. For example, you can specify monitors for
a Broker with
`druid.monitoring.monitors=["io.druid.java.util.metrics.SysMonitor","io.druid.java.util.metrics.JvmMonitor"]`.|none
(no monitors)|
+
+The following monitors are available:
+
+|Name|Description|
+|----|-----------|
+|`io.druid.client.cache.CacheMonitor`|Emits metrics (to logs) about the
segment results cache for Historical and Broker nodes. Reports typical cache
statistics include hits, misses, rates, and size (bytes and number of entries),
as well as timeouts and and errors.|
+|`io.druid.java.util.metrics.SysMonitor`|This uses the [SIGAR
library](http://www.hyperic.com/products/sigar) to report on various system
activities and statuses.|
+|`io.druid.server.metrics.HistoricalMetricsMonitor`|Reports statistics on
Historical nodes.|
+|`io.druid.java.util.metrics.JvmMonitor`|Reports various JVM-related
statistics.|
+|`io.druid.java.util.metrics.JvmCpuMonitor`|Reports statistics of CPU
consumption by the JVM.|
+|`io.druid.java.util.metrics.CpuAcctDeltaMonitor`|Reports consumed CPU as per
the cpuacct cgroup.|
+|`io.druid.java.util.metrics.JvmThreadsMonitor`|Reports Thread statistics in
the JVM, like numbers of total, daemon, started, died threads.|
+|`io.druid.segment.realtime.RealtimeMetricsMonitor`|Reports statistics on
Realtime nodes.|
+|`io.druid.server.metrics.EventReceiverFirehoseMonitor`|Reports how many
events have been queued in the EventReceiverFirehose.|
+|`io.druid.server.metrics.QueryCountStatsMonitor`|Reports how many queries
have been successful/failed/interrupted.|
+|`io.druid.server.emitter.HttpEmittingMonitor`|Reports internal metrics of
`http` or `parametrized` emitter (see below). Must not be used with another
emitter type. See the description of the metrics here:
https://github.com/druid-io/druid/pull/4973.|
+
+
+### Emitting Metrics
+
+The Druid servers [emit various metrics](../operations/metrics.html) and
alerts via something we call an Emitter. There are three emitter
implementations included with the code, a "noop" emitter, one that just logs to
log4j ("logging", which is used by default if no emitter is specified) and one
that does POSTs of JSON events to a server ("http"). The properties for using
the logging emitter are described below.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.emitter`|Setting this value to "noop", "logging", "http" or
"parametrized" will initialize one of the emitter modules. value "composing"
can be used to initialize multiple emitter modules. |noop|
+
+#### Logging Emitter Module
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.emitter.logging.loggerClass`|Choices: HttpPostEmitter, LoggingEmitter,
NoopServiceEmitter, ServiceEmitter. The class used for logging.|LoggingEmitter|
+|`druid.emitter.logging.logLevel`|Choices: debug, info, warn, error. The log
level at which message are logged.|info|
+
+#### Http Emitter Module
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.emitter.http.flushMillis`|How often the internal message buffer is
flushed (data is sent).|60000|
+|`druid.emitter.http.flushCount`|How many messages the internal message buffer
can hold before flushing (sending).|500|
+|`druid.emitter.http.basicAuthentication`|Login and password for
authentification in "login:password" form, e. g.
`druid.emitter.http.basicAuthentication=admin:adminpassword`|not specified = no
authentification|
+|`druid.emitter.http.flushTimeOut`|The timeout after which an event should be
sent to the endpoint, even if internal buffers are not filled, in
milliseconds.|not specified = no timeout|
+|`druid.emitter.http.batchingStrategy`|The strategy of how the batch is
formatted. "ARRAY" means `[event1,event2]`, "NEWLINES" means `event1\nevent2`,
ONLY_EVENTS means `event1event2`.|ARRAY|
+|`druid.emitter.http.maxBatchSize`|The maximum batch size, in bytes.|the
minimum of (10% of JVM heap size divided by 2) or (5191680 (i. e. 5 MB))|
+|`druid.emitter.http.batchQueueSizeLimit`|The maximum number of batches in
emitter queue, if there are problems with emitting.|the maximum of (2) or (10%
of the JVM heap size divided by 5MB)|
+|`druid.emitter.http.minHttpTimeoutMillis`|If the speed of filling batches
imposes timeout smaller than that, not even trying to send batch to endpoint,
because it will likely fail, not being able to send the data that fast.
Configure this depending based on emitter/successfulSending/minTimeMs metric.
Reasonable values are 10ms..100ms.|0|
+|`druid.emitter.http.recipientBaseUrl`|The base URL to emit messages to. Druid
will POST JSON to be consumed at the HTTP endpoint specified by this
property.|none, required config|
+
+#### Http Emitter Module TLS Overrides
+
+When emitting events to a TLS-enabled receiver, the Http Emitter will by
default use an SSLContext obtained via the process described at [Druid's
internal communication over
TLS](../operations/tls-support.html#druids-internal-communication-over-tls),
i.e., the same SSLContext that would be used for internal communications
between Druid nodes.
+
+In some use cases it may be desirable to have the Http Emitter use its own
separate truststore configuration. For example, there may be organizational
policies that prevent the TLS-enabled metrics receiver's certificate from being
added to the same truststore used by Druid's internal HTTP client.
+
+The following properties allow the Http Emitter to use its own truststore
configuration when building its SSLContext.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.emitter.http.ssl.useDefaultJavaContext`|If set to true, the
HttpEmitter will use `SSLContext.getDefault()`, the default Java SSLContext,
and all other properties below are ignored.|false|
+|`druid.emitter.http.ssl.trustStorePath`|The file path or URL of the TLS/SSL
Key store where trusted root certificates are stored. If this is unspecified,
the Http Emitter will use the same SSLContext as Druid's internal HTTP client,
as described in the beginning of this section, and all other properties below
are ignored.|null|
+|`druid.emitter.http.ssl.trustStoreType`|The type of the key store where
trusted root certificates are stored.|`java.security.KeyStore.getDefaultType()`|
+|`druid.emitter.http.ssl.trustStoreAlgorithm`|Algorithm to be used by
TrustManager to validate certificate
chains|`javax.net.ssl.TrustManagerFactory.getDefaultAlgorithm()`|
+|`druid.emitter.http.ssl.trustStorePassword`|The [Password
Provider](../../operations/password-provider.html) or String password for the
Trust Store.|none|
+|`druid.emitter.http.ssl.protocol`|TLS protocol to use.|"TLSv1.2"|
+
+#### Parametrized Http Emitter Module
+
+`druid.emitter.parametrized.httpEmitting.*` configs correspond to the configs
of Http Emitter Modules, see above.
+Except `recipientBaseUrl`. E. g.
`druid.emitter.parametrized.httpEmitting.flushMillis`,
+`druid.emitter.parametrized.httpEmitting.flushCount`,
`druid.emitter.parametrized.httpEmitting.ssl.trustStorePath`, etc.
+
+The additional configs are:
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.emitter.parametrized.recipientBaseUrlPattern`|The URL pattern to send
an event to, based on the event's feed. E. g. `http://foo.bar/{feed}`, that
will send event to `http://foo.bar/metrics` if the event's feed is
"metrics".|none, required config|
+
+#### Composing Emitter Module
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.emitter.composing.emitters`|List of emitter modules to load e.g.
["logging","http"].|[]|
+
+#### Graphite Emitter
+
+To use graphite as emitter set `druid.emitter=graphite`. For configuration
details please follow this
[link](../development/extensions-contrib/graphite.html).
+
+
+### Metadata Storage
+
+These properties specify the jdbc connection and other configuration around
the metadata storage. The only processes that connect to the metadata storage
with these properties are the [Coordinator](../design/coordinator.html),
[Indexing service](../design/indexing-service.html) and [Realtime
Nodes](../design/realtime.html).
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.metadata.storage.type`|The type of metadata storage to use. Choose
from "mysql", "postgresql", or "derby".|derby|
+|`druid.metadata.storage.connector.connectURI`|The jdbc uri for the database
to connect to|none|
+|`druid.metadata.storage.connector.user`|The username to connect with.|none|
+|`druid.metadata.storage.connector.password`|The [Password
Provider](../operations/password-provider.html) or String password used to
connect with.|none|
+|`druid.metadata.storage.connector.createTables`|If Druid requires a table and
it doesn't exist, create it?|true|
+|`druid.metadata.storage.tables.base`|The base name for tables.|druid|
+|`druid.metadata.storage.tables.segments`|The table to use to look for
segments.|druid_segments|
+|`druid.metadata.storage.tables.rules`|The table to use to look for segment
load/drop rules.|druid_rules|
+|`druid.metadata.storage.tables.config`|The table to use to look for
configs.|druid_config|
+|`druid.metadata.storage.tables.tasks`|Used by the indexing service to store
tasks.|druid_tasks|
+|`druid.metadata.storage.tables.taskLog`|Used by the indexing service to store
task logs.|druid_taskLog|
+|`druid.metadata.storage.tables.taskLock`|Used by the indexing service to
store task locks.|druid_taskLock|
+|`druid.metadata.storage.tables.supervisors`|Used by the indexing service to
store supervisor configurations.|druid_supervisors|
+|`druid.metadata.storage.tables.audit`|The table to use for audit history of
configuration changes e.g. Coordinator rules.|druid_audit|
+
+### Deep Storage
+
+The configurations concern how to push and pull
[Segments](../design/segments.html) from deep storage.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.storage.type`|Choices:local, noop, s3, hdfs, c*. The type of deep
storage to use.|local|
+
+#### Local Deep Storage
+
+Local deep storage uses the local filesystem.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.storage.storageDirectory`|Directory on disk to use as deep
storage.|/tmp/druid/localStorage|
+
+#### Noop Deep Storage
+
+This deep storage doesn't do anything. There are no configs.
+
+#### S3 Deep Storage
+
+This deep storage is used to interface with Amazon's S3.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.s3.accessKey`|The access key to use to access S3.|none|
+|`druid.s3.secretKey`|The secret key to use to access S3.|none|
+|`druid.storage.bucket`|S3 bucket name.|none|
+|`druid.storage.baseKey`|S3 object key prefix for storage.|none|
+|`druid.storage.disableAcl`|Boolean flag for ACL.|false|
+|`druid.storage.archiveBucket`|S3 bucket name for archiving when running the
indexing-service *archive task*.|none|
+|`druid.storage.archiveBaseKey`|S3 object key prefix for archiving.|none|
+
+#### HDFS Deep Storage
+
+This deep storage is used to interface with HDFS.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.storage.storageDirectory`|HDFS directory to use as deep storage.|none|
+
+#### Cassandra Deep Storage
+
+This deep storage is used to interface with Cassandra.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.storage.host`|Cassandra host.|none|
+|`druid.storage.keyspace`|Cassandra key space.|none|
+
+
+### Task Logging
+
+If you are running the indexing service in remote mode, the task logs must be
stored in S3, Azure Blob Store, Google Cloud Storage or HDFS.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.indexer.logs.type`|Choices:noop, s3, azure, google, hdfs, file. Where
to store task logs|file|
+
+You can also configure the Overlord to automatically retain the task logs only
for last x milliseconds by configuring following additional properties.
+Caution: Automatic log file deletion typically works based on log file
modification timestamp on the backing store, so large clock skews between druid
nodes and backing store nodes might result in un-intended behavior.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.indexer.logs.kill.enabled`|Boolean value for whether to enable
deletion of old task logs. |false|
+|`druid.indexer.logs.kill.durationToRetain`| Required if kill is enabled. In
milliseconds, task logs to be retained created in last x milliseconds. |None|
+|`druid.indexer.logs.kill.initialDelay`| Optional. Number of milliseconds
after overlord start when first auto kill is run. |random value less than
300000 (5 mins)|
+|`druid.indexer.logs.kill.delay`|Optional. Number of milliseconds of delay
between successive executions of auto kill run. |21600000 (6 hours)|
+
+#### File Task Logs
+
+Store task logs in the local filesystem.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.indexer.logs.directory`|Local filesystem path.|log|
+
+#### S3 Task Logs
+
+Store task logs in S3.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.indexer.logs.s3Bucket`|S3 bucket name.|none|
+|`druid.indexer.logs.s3Prefix`|S3 key prefix.|none|
+
+#### Azure Blob Store Task Logs
+Store task logs in Azure Blob Store.
+
+Note: this uses the same storage account as the deep storage module for azure.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.indexer.logs.container`|The Azure Blob Store container to write logs
to|none|
+|`druid.indexer.logs.prefix`|The path to prepend to logs|none|
+
+#### Google Cloud Storage Task Logs
+Store task logs in Google Cloud Storage.
+
+Note: this uses the same storage settings as the deep storage module for
google.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.indexer.logs.bucket`|The Google Cloud Storage bucket to write logs
to|none|
+|`druid.indexer.logs.prefix`|The path to prepend to logs|none|
+
+#### HDFS Task Logs
+
+Store task logs in HDFS.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.indexer.logs.directory`|The directory to store logs.|none|
+
+### Indexing Service Discovery
+
+This config is used to find the [Indexing
Service](../design/indexing-service.html) using Curator service discovery. Only
required if you are actually running an indexing service.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.selectors.indexing.serviceName`|The druid.service name of the indexing
service Overlord node. To start the Overlord with a different name, set it with
this property. |druid/overlord|
+
+
+### Coordinator Discovery
+
+This config is used to find the [Coordinator](../design/coordinator.html)
using Curator service discovery. This config is used by the realtime indexing
nodes to get information about the segments loaded in the cluster.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.selectors.coordinator.serviceName`|The druid.service name of the
coordinator node. To start the Coordinator with a different name, set it with
this property. |druid/coordinator|
+
+
+### Announcing Segments
+
+You can configure how to announce and unannounce Znodes in ZooKeeper (using
Curator). For normal operations you do not need to override any of these
configs.
+
+##### Batch Data Segment Announcer
+
+In current Druid, multiple data segments may be announced under the same Znode.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.announcer.segmentsPerNode`|Each Znode contains info for up to this
many segments.|50|
+|`druid.announcer.maxBytesPerNode`|Max byte size for Znode.|524288|
+|`druid.announcer.skipDimensionsAndMetrics`|Skip Dimensions and Metrics list
from segment announcements. NOTE: Enabling this will also remove the dimensions
and metrics list from coordinator and broker endpoints.|false|
+|`druid.announcer.skipLoadSpec`|Skip segment LoadSpec from segment
announcements. NOTE: Enabling this will also remove the loadspec from
coordinator and broker endpoints.|false|
+
+### JavaScript
+
+Druid supports dynamic runtime extension through JavaScript functions. This
functionality can be configured through
+the following properties.
+
+|Property|Description|Default|
+|--------|-----------|-------|
+|`druid.javascript.enabled`|Set to "true" to enable JavaScript functionality.
This affects the JavaScript parser, filter, extractionFn, aggregator,
post-aggregator, router strategy, and worker selection strategy.|false|
+
+<div class="note info">
+JavaScript-based functionality is disabled by default. Please refer to the
Druid <a href="../development/javascript.html">JavaScript programming guide</a>
for guidelines about using Druid's JavaScript functionality, including
instructions on how to enable it.
+</div>
+
+### Double Column storage
+
+Prior to version 0.13.0 Druid's storage layer used a 32-bit float
representation to store columns created by the
+doubleSum, doubleMin, and doubleMax aggregators at indexing time.
+Starting from version 0.13.0 the default will be 64-bit floats for Double
columns.
+Using 64-bit representation for double column will lead to avoid precesion
loss at the cost of doubling the storage size of such columns.
+To keep the old format set the system-wide property
`druid.indexing.doubleStorage=float`.
+You can also use floatSum, floatMin and floatMax to use 32-bit float
representation.
+Support for 64-bit floating point columns was released in Druid 0.11.0, so if
you use this feature then older versions of Druid will not be able to read your
data segments.
+|Property|Description|Default|
Review comment:
Added newline
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]
With regards,
Apache Git Services
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]