techdocsmith commented on a change in pull request #11541:
URL: https://github.com/apache/druid/pull/11541#discussion_r682129212
##########
File path: docs/ingestion/partitioning.md
##########
@@ -0,0 +1,50 @@
+---
+id: partitioning
+title: Partitioning
+sidebar_label: Partitioning
+description: Describes time chunk and secondary partitioning in Druid.
Provides guidance to choose a secondary partition dimension.
+---
+
+Optimal partitioning and sorting of segments within your Druid datasources can
have substantial impact on footprint and performance.
+
+One way to partition is to your load data into separate datasources. This is a
perfectly viable approach that works very well when the number of datasources
does not lead to excessive per-datasource overheads.
+
+This topic describes how to set up partitions within a single datasource. It
does not cover using multiple datasources. See [Multitenancy
considerations](../querying/multitenancy.md) for more details on splitting data
into separate datasources and potential operational considerations.
+
+## Time chunk partitioning
+
+Druid always partitions datasources by time into _time chunks_. Each time
chunk contains one or more segments. This partitioning happens for all
ingestion methods based on the `segmentGranularity` parameter in your ingestion
spec `dataSchema` object.
+
+## Secondary partitioning
+
+Druid can partition segments within a particular time chunk further depending
upon options that vary based on the ingestion type you have chosen. In general,
secondary partitioning on a particular dimension improves locality. This means
that rows with the same value for that dimension are stored together,
decreasing access time.
+
+To achieve the best performance and smallest overall footprint, partition your
data on a "natural"
+dimension that you often use as a filter when possible. Such partitioning
often improves compression and query performance. For example, some cases have
yielded threefold storage size decreases.
+
+## Partitioning and sorting
+
+Partitioning and sorting work well together. If you do have a "natural"
partitioning dimension, consider placing it first in the `dimensions` list of
your `dimensionsSpec`. This way Druid sorts rows within each segment by that
column. This sorting configuration frequently improves compression more than
using partitioning alone.
+
+> Note that Druid always sorts rows within a segment by timestamp first, even
before the first dimension listed in your `dimensionsSpec`. This sorting can
preclude the efficacy of dimension sorting. To work around this limitation if
necessary, set your `queryGranularity` equal to `segmentGranularity` in your
[`granularitySpec`](./ingestion-spec.md#granularityspec). Druid will set all
timestamps within the segment to the same value, and letting you identify a
[secondary timestamp](schema-design.md#secondary-timestamps) as the "real"
timestamp.
+
+## How to configure partitioning
+
+Not all ingestion methods support an explicit partitioning configuration, and
not all have equivalent levels of flexibility. If you are doing initial
ingestion through a less-flexible method like
+Kafka), you can use [reindexing](data-management.md#reingesting-data) or
[compaction](compaction.md) to repartition your data after initial ingestion.
This is a powerful technique you can use to optimally partition any data older
than a certain even while you continuously add new data from a stream.
Review comment:
```suggestion
Kafka), you can use [reindexing](data-management.md#reingesting-data) or
[compaction](compaction.md) to repartition your data after initial ingestion.
This is a powerful technique you can use to optimally partition any data older
than a certain time threshold while you continuously add new data from a stream.
```
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]