Repository: kafka Updated Branches: refs/heads/trunk b9f812491 -> be1127281
MINOR: Fix a documentation typo Author: Vahid Hashemian <vahidhashem...@us.ibm.com> Reviewers: Sriharsha Chintalapani <har...@hortonworks.com> Closes #2674 from vahidhashemian/minor/fix_typos_1703 Project: http://git-wip-us.apache.org/repos/asf/kafka/repo Commit: http://git-wip-us.apache.org/repos/asf/kafka/commit/be112728 Tree: http://git-wip-us.apache.org/repos/asf/kafka/tree/be112728 Diff: http://git-wip-us.apache.org/repos/asf/kafka/diff/be112728 Branch: refs/heads/trunk Commit: be11272818a06eea157ab7f6eee8905855f1cede Parents: b9f8124 Author: Vahid Hashemian <vahidhashem...@us.ibm.com> Authored: Wed Mar 15 11:50:57 2017 -0700 Committer: Sriharsha Chintalapani <har...@hortonworks.com> Committed: Wed Mar 15 11:50:57 2017 -0700 ---------------------------------------------------------------------- docs/ops.html | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/kafka/blob/be112728/docs/ops.html ---------------------------------------------------------------------- diff --git a/docs/ops.html b/docs/ops.html index 9232f65..7be9939 100644 --- a/docs/ops.html +++ b/docs/ops.html @@ -33,7 +33,7 @@ </pre> The replication factor controls how many servers will replicate each message that is written. If you have a replication factor of 3 then up to 2 servers can fail before you will lose access to your data. We recommend you use a replication factor of 2 or 3 so that you can transparently bounce machines without interrupting data consumption. <p> - The partition count controls how many logs the topic will be sharded into. There are several impacts of the partition count. First each partition must fit entirely on a single server. So if you have 20 partitions the full data set (and read and write load) will be handled by no more than 20 servers (noi counting replicas). Finally the partition count impacts the maximum parallelism of your consumers. This is discussed in greater detail in the <a href="#intro_consumers">concepts section</a>. + The partition count controls how many logs the topic will be sharded into. There are several impacts of the partition count. First each partition must fit entirely on a single server. So if you have 20 partitions the full data set (and read and write load) will be handled by no more than 20 servers (not counting replicas). Finally the partition count impacts the maximum parallelism of your consumers. This is discussed in greater detail in the <a href="#intro_consumers">concepts section</a>. <p> Each sharded partition log is placed into its own folder under the Kafka log directory. The name of such folders consists of the topic name, appended by a dash (-) and the partition id. Since a typical folder name can not be over 255 characters long, there will be a limitation on the length of topic names. We assume the number of partitions will not ever be above 100,000. Therefore, topic names cannot be longer than 249 characters. This leaves just enough room in the folder name for a dash and a potentially 5 digit long partition id. <p>