http://git-wip-us.apache.org/repos/asf/kafka-site/blob/970abca9/0101/ops.html
----------------------------------------------------------------------
diff --git a/0101/ops.html b/0101/ops.html
index c26f0cb..5d8c60b 100644
--- a/0101/ops.html
+++ b/0101/ops.html
@@ -15,1342 +15,1346 @@
  limitations under the License.
 -->
 
-Here is some information on actually running Kafka as a production system 
based on usage and experience at LinkedIn. Please send us any additional tips 
you know of.
-
-<h3><a id="basic_ops" href="#basic_ops">6.1 Basic Kafka Operations</a></h3>
-
-This section will review the most common operations you will perform on your 
Kafka cluster. All of the tools reviewed in this section are available under 
the <code>bin/</code> directory of the Kafka distribution and each tool will 
print details on all possible commandline options if it is run with no 
arguments.
-
-<h4><a id="basic_ops_add_topic" href="#basic_ops_add_topic">Adding and 
removing topics</a></h4>
-
-You have the option of either adding topics manually or having them be created 
automatically when data is first published to a non-existent topic. If topics 
are auto-created then you may want to tune the default <a 
href="#topic-config">topic configurations</a> used for auto-created topics.
-<p>
-Topics are added and modified using the topic tool:
-<pre>
- &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --create --topic 
my_topic_name
-       --partitions 20 --replication-factor 3 --config x=y
-</pre>
-The replication factor controls how many servers will replicate each message 
that is written. If you have a replication factor of 3 then up to 2 servers can 
fail before you will lose access to your data. We recommend you use a 
replication factor of 2 or 3 so that you can transparently bounce machines 
without interrupting data consumption.
-<p>
-The partition count controls how many logs the topic will be sharded into. 
There are several impacts of the partition count. First each partition must fit 
entirely on a single server. So if you have 20 partitions the full data set 
(and read and write load) will be handled by no more than 20 servers (no 
counting replicas). Finally the partition count impacts the maximum parallelism 
of your consumers. This is discussed in greater detail in the <a 
href="#intro_consumers">concepts section</a>.
-<p>
-Each sharded partition log is placed into its own folder under the Kafka log 
directory. The name of such folders consists of the topic name, appended by a 
dash (-) and the partition id. Since a typical folder name can not be over 255 
characters long, there will be a limitation on the length of topic names. We 
assume the number of partitions will not ever be above 100,000. Therefore, 
topic names cannot be longer than 249 characters. This leaves just enough room 
in the folder name for a dash and a potentially 5 digit long partition id.
-<p>
-The configurations added on the command line override the default settings the 
server has for things like the length of time data should be retained. The 
complete set of per-topic configurations is documented <a 
href="#topic-config">here</a>.
-
-<h4><a id="basic_ops_modify_topic" href="#basic_ops_modify_topic">Modifying 
topics</a></h4>
-
-You can change the configuration or partitioning of a topic using the same 
topic tool.
-<p>
-To add partitions you can do
-<pre>
- &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --alter --topic 
my_topic_name
-       --partitions 40
-</pre>
-Be aware that one use case for partitions is to semantically partition data, 
and adding partitions doesn't change the partitioning of existing data so this 
may disturb consumers if they rely on that partition. That is if data is 
partitioned by <code>hash(key) % number_of_partitions</code> then this 
partitioning will potentially be shuffled by adding partitions but Kafka will 
not attempt to automatically redistribute data in any way.
-<p>
-To add configs:
-<pre>
- &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --alter --topic 
my_topic_name --config x=y
-</pre>
-To remove a config:
-<pre>
- &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --alter --topic 
my_topic_name --delete-config x
-</pre>
-And finally deleting a topic:
-<pre>
- &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --delete --topic 
my_topic_name
-</pre>
-Topic deletion option is disabled by default. To enable it set the server 
config
-  <pre>delete.topic.enable=true</pre>
-<p>
-Kafka does not currently support reducing the number of partitions for a topic.
-<p>
-Instructions for changing the replication factor of a topic can be found <a 
href="#basic_ops_increase_replication_factor">here</a>.
-
-<h4><a id="basic_ops_restarting" href="#basic_ops_restarting">Graceful 
shutdown</a></h4>
-
-The Kafka cluster will automatically detect any broker shutdown or failure and 
elect new leaders for the partitions on that machine. This will occur whether a 
server fails or it is brought down intentionally for maintenance or 
configuration changes. For the latter cases Kafka supports a more graceful 
mechanism for stopping a server than just killing it.
-
-When a server is stopped gracefully it has two optimizations it will take 
advantage of:
-<ol>
-    <li>It will sync all its logs to disk to avoid needing to do any log 
recovery when it restarts (i.e. validating the checksum for all messages in the 
tail of the log). Log recovery takes time so this speeds up intentional 
restarts.
-    <li>It will migrate any partitions the server is the leader for to other 
replicas prior to shutting down. This will make the leadership transfer faster 
and minimize the time each partition is unavailable to a few milliseconds.
-</ol>
-
-Syncing the logs will happen automatically whenever the server is stopped 
other than by a hard kill, but the controlled leadership migration requires 
using a special setting:
-<pre>
-    controlled.shutdown.enable=true
-</pre>
-Note that controlled shutdown will only succeed if <i>all</i> the partitions 
hosted on the broker have replicas (i.e. the replication factor is greater than 
1 <i>and</i> at least one of these replicas is alive). This is generally what 
you want since shutting down the last replica would make that topic partition 
unavailable.
-
-<h4><a id="basic_ops_leader_balancing" 
href="#basic_ops_leader_balancing">Balancing leadership</a></h4>
-
-Whenever a broker stops or crashes leadership for that broker's partitions 
transfers to other replicas. This means that by default when the broker is 
restarted it will only be a follower for all its partitions, meaning it will 
not be used for client reads and writes.
-<p>
-To avoid this imbalance, Kafka has a notion of preferred replicas. If the list 
of replicas for a partition is 1,5,9 then node 1 is preferred as the leader to 
either node 5 or 9 because it is earlier in the replica list. You can have the 
Kafka cluster try to restore leadership to the restored replicas by running the 
command:
-<pre>
- &gt; bin/kafka-preferred-replica-election.sh --zookeeper zk_host:port/chroot
-</pre>
-
-Since running this command can be tedious you can also configure Kafka to do 
this automatically by setting the following configuration:
-<pre>
-    auto.leader.rebalance.enable=true
-</pre>
-
-<h4><a id="basic_ops_racks" href="#basic_ops_racks">Balancing Replicas Across 
Racks</a></h4>
-The rack awareness feature spreads replicas of the same partition across 
different racks. This extends the guarantees Kafka provides for broker-failure 
to cover rack-failure, limiting the risk of data loss should all the brokers on 
a rack fail at once. The feature can also be applied to other broker groupings 
such as availability zones in EC2.
-<p></p>
-You can specify that a broker belongs to a particular rack by adding a 
property to the broker config:
-<pre>   broker.rack=my-rack-id</pre>
-When a topic is <a href="#basic_ops_add_topic">created</a>, <a 
href="#basic_ops_modify_topic">modified</a> or replicas are <a 
href="#basic_ops_cluster_expansion">redistributed</a>, the rack constraint will 
be honoured, ensuring replicas span as many racks as they can (a partition will 
span min(#racks, replication-factor) different racks).
-<p></p>
-The algorithm used to assign replicas to brokers ensures that the number of 
leaders per broker will be constant, regardless of how brokers are distributed 
across racks. This ensures balanced throughput.
-<p></p>
-However if racks are assigned different numbers of brokers, the assignment of 
replicas will not be even. Racks with fewer brokers will get more replicas, 
meaning they will use more storage and put more resources into replication. 
Hence it is sensible to configure an equal number of brokers per rack.
-
-<h4><a id="basic_ops_mirror_maker" href="#basic_ops_mirror_maker">Mirroring 
data between clusters</a></h4>
-
-We refer to the process of replicating data <i>between</i> Kafka clusters 
"mirroring" to avoid confusion with the replication that happens amongst the 
nodes in a single cluster. Kafka comes with a tool for mirroring data between 
Kafka clusters. The tool consumes from a source cluster and produces to a 
destination cluster.
-
-A common use case for this kind of mirroring is to provide a replica in 
another datacenter. This scenario will be discussed in more detail in the next 
section.
-<p>
-You can run many such mirroring processes to increase throughput and for 
fault-tolerance (if one process dies, the others will take overs the additional 
load).
-<p>
-Data will be read from topics in the source cluster and written to a topic 
with the same name in the destination cluster. In fact the mirror maker is 
little more than a Kafka consumer and producer hooked together.
-<p>
-The source and destination clusters are completely independent entities: they 
can have different numbers of partitions and the offsets will not be the same. 
For this reason the mirror cluster is not really intended as a fault-tolerance 
mechanism (as the consumer position will be different); for that we recommend 
using normal in-cluster replication. The mirror maker process will, however, 
retain and use the message key for partitioning so order is preserved on a 
per-key basis.
-<p>
-Here is an example showing how to mirror a single topic (named 
<i>my-topic</i>) from an input cluster:
-<pre>
- &gt; bin/kafka-mirror-maker.sh
-       --consumer.config consumer.properties
-       --producer.config producer.properties --whitelist my-topic
-</pre>
-Note that we specify the list of topics with the <code>--whitelist</code> 
option. This option allows any regular expression using <a 
href="http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html";>Java-style
 regular expressions</a>. So you could mirror two topics named <i>A</i> and 
<i>B</i> using <code>--whitelist 'A|B'</code>. Or you could mirror <i>all</i> 
topics using <code>--whitelist '*'</code>. Make sure to quote any regular 
expression to ensure the shell doesn't try to expand it as a file path. For 
convenience we allow the use of ',' instead of '|' to specify a list of topics.
-<p>
-Sometimes it is easier to say what it is that you <i>don't</i> want. Instead 
of using <code>--whitelist</code> to say what you want
-to mirror you can use <code>--blacklist</code> to say what to exclude. This 
also takes a regular expression argument.
-However, <code>--blacklist</code> is not supported when the new consumer has 
been enabled (i.e. when <code>bootstrap.servers</code>
-has been defined in the consumer configuration).
-<p>
-Combining mirroring with the configuration 
<code>auto.create.topics.enable=true</code> makes it possible to have a replica 
cluster that will automatically create and replicate all data in a source 
cluster even as new topics are added.
-
-<h4><a id="basic_ops_consumer_lag" href="#basic_ops_consumer_lag">Checking 
consumer position</a></h4>
-Sometimes it's useful to see the position of your consumers. We have a tool 
that will show the position of all consumers in a consumer group as well as how 
far behind the end of the log they are. To run this tool on a consumer group 
named <i>my-group</i> consuming a topic named <i>my-topic</i> would look like 
this:
-<pre>
- &gt; bin/kafka-run-class.sh kafka.tools.ConsumerOffsetChecker --zookeeper 
localhost:2181 --group test
-Group           Topic                          Pid Offset          logSize     
    Lag             Owner
-my-group        my-topic                       0   0               0           
    0               test_jkreps-mn-1394154511599-60744496-0
-my-group        my-topic                       1   0               0           
    0               test_jkreps-mn-1394154521217-1a0be913-0
-</pre>
-
-
-NOTE: Since 0.9.0.0, the kafka.tools.ConsumerOffsetChecker tool has been 
deprecated. You should use the kafka.admin.ConsumerGroupCommand (or the 
bin/kafka-consumer-groups.sh script) to manage consumer groups, including 
consumers created with the <a 
href="http://kafka.apache.org/documentation.html#newconsumerapi";>new consumer 
API</a>.
-
-<h4><a id="basic_ops_consumer_group" href="#basic_ops_consumer_group">Managing 
Consumer Groups</a></h4>
-
-With the ConsumerGroupCommand tool, we can list, describe, or delete consumer 
groups. Note that deletion is only available when the group metadata is stored 
in
-ZooKeeper. When using the <a 
href="http://kafka.apache.org/documentation.html#newconsumerapi";>new consumer 
API</a> (where
-the broker handles coordination of partition handling and rebalance), the 
group is deleted when the last committed offset for that group expires.
-
-For example, to list all consumer groups across all topics:
-
-<pre>
- &gt; bin/kafka-consumer-groups.sh --bootstrap-server broker1:9092 --list
-
-test-consumer-group
-</pre>
-
-To view offsets as in the previous example with the ConsumerOffsetChecker, we 
"describe" the consumer group like this:
-
-<pre>
- &gt; bin/kafka-consumer-groups.sh --bootstrap-server broker1:9092 --describe 
--group test-consumer-group
-
-GROUP                          TOPIC                          PARTITION  
CURRENT-OFFSET  LOG-END-OFFSET  LAG             OWNER
-test-consumer-group            test-foo                       0          1     
          3               2               consumer-1_/127.0.0.1
-</pre>
-
-If you are using the old high-level consumer and storing the group metadata in 
ZooKeeper (i.e. <code>offsets.storage=zookeeper</code>), pass
-<code>--zookeeper</code> instead of <code>bootstrap-server</code>:
-
-<pre>
- &gt; bin/kafka-consumer-groups.sh --zookeeper localhost:2181 --list
-</pre>
-
-<h4><a id="basic_ops_cluster_expansion" 
href="#basic_ops_cluster_expansion">Expanding your cluster</a></h4>
-
-Adding servers to a Kafka cluster is easy, just assign them a unique broker id 
and start up Kafka on your new servers. However these new servers will not 
automatically be assigned any data partitions, so unless partitions are moved 
to them they won't be doing any work until new topics are created. So usually 
when you add machines to your cluster you will want to migrate some existing 
data to these machines.
-<p>
-The process of migrating data is manually initiated but fully automated. Under 
the covers what happens is that Kafka will add the new server as a follower of 
the partition it is migrating and allow it to fully replicate the existing data 
in that partition. When the new server has fully replicated the contents of 
this partition and joined the in-sync replica one of the existing replicas will 
delete their partition's data.
-<p>
-The partition reassignment tool can be used to move partitions across brokers. 
An ideal partition distribution would ensure even data load and partition sizes 
across all brokers. The partition reassignment tool does not have the 
capability to automatically study the data distribution in a Kafka cluster and 
move partitions around to attain an even load distribution. As such, the admin 
has to figure out which topics or partitions should be moved around.
-<p>
-The partition reassignment tool can run in 3 mutually exclusive modes:
-<ul>
-<li>--generate: In this mode, given a list of topics and a list of brokers, 
the tool generates a candidate reassignment to move all partitions of the 
specified topics to the new brokers. This option merely provides a convenient 
way to generate a partition reassignment plan given a list of topics and target 
brokers.</li>
-<li>--execute: In this mode, the tool kicks off the reassignment of partitions 
based on the user provided reassignment plan. (using the 
--reassignment-json-file option). This can either be a custom reassignment plan 
hand crafted by the admin or provided by using the --generate option</li>
-<li>--verify: In this mode, the tool verifies the status of the reassignment 
for all partitions listed during the last --execute. The status can be either 
of successfully completed, failed or in progress</li>
-</ul>
-<h5><a id="basic_ops_automigrate" href="#basic_ops_automigrate">Automatically 
migrating data to new machines</a></h5>
-The partition reassignment tool can be used to move some topics off of the 
current set of brokers to the newly added brokers. This is typically useful 
while expanding an existing cluster since it is easier to move entire topics to 
the new set of brokers, than moving one partition at a time. When used to do 
this, the user should provide a list of topics that should be moved to the new 
set of brokers and a target list of new brokers. The tool then evenly 
distributes all partitions for the given list of topics across the new set of 
brokers. During this move, the replication factor of the topic is kept 
constant. Effectively the replicas for all partitions for the input list of 
topics are moved from the old set of brokers to the newly added brokers.
-<p>
-For instance, the following example will move all partitions for topics 
foo1,foo2 to the new set of brokers 5,6. At the end of this move, all 
partitions for topics foo1 and foo2 will <i>only</i> exist on brokers 5,6.
-<p>
-Since the tool accepts the input list of topics as a json file, you first need 
to identify the topics you want to move and create the json file as follows:
-<pre>
-> cat topics-to-move.json
-{"topics": [{"topic": "foo1"},
-            {"topic": "foo2"}],
- "version":1
-}
-</pre>
-Once the json file is ready, use the partition reassignment tool to generate a 
candidate assignment:
-<pre>
-> bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--topics-to-move-json-file topics-to-move.json --broker-list "5,6" --generate
-Current partition replica assignment
-
-{"version":1,
- "partitions":[{"topic":"foo1","partition":2,"replicas":[1,2]},
-               {"topic":"foo1","partition":0,"replicas":[3,4]},
-               {"topic":"foo2","partition":2,"replicas":[1,2]},
-               {"topic":"foo2","partition":0,"replicas":[3,4]},
-               {"topic":"foo1","partition":1,"replicas":[2,3]},
-               {"topic":"foo2","partition":1,"replicas":[2,3]}]
-}
-
-Proposed partition reassignment configuration
-
-{"version":1,
- "partitions":[{"topic":"foo1","partition":2,"replicas":[5,6]},
-               {"topic":"foo1","partition":0,"replicas":[5,6]},
-               {"topic":"foo2","partition":2,"replicas":[5,6]},
-               {"topic":"foo2","partition":0,"replicas":[5,6]},
-               {"topic":"foo1","partition":1,"replicas":[5,6]},
-               {"topic":"foo2","partition":1,"replicas":[5,6]}]
-}
-</pre>
-<p>
-The tool generates a candidate assignment that will move all partitions from 
topics foo1,foo2 to brokers 5,6. Note, however, that at this point, the 
partition movement has not started, it merely tells you the current assignment 
and the proposed new assignment. The current assignment should be saved in case 
you want to rollback to it. The new assignment should be saved in a json file 
(e.g. expand-cluster-reassignment.json) to be input to the tool with the 
--execute option as follows:
-<pre>
-> bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file expand-cluster-reassignment.json --execute
-Current partition replica assignment
-
-{"version":1,
- "partitions":[{"topic":"foo1","partition":2,"replicas":[1,2]},
-               {"topic":"foo1","partition":0,"replicas":[3,4]},
-               {"topic":"foo2","partition":2,"replicas":[1,2]},
-               {"topic":"foo2","partition":0,"replicas":[3,4]},
-               {"topic":"foo1","partition":1,"replicas":[2,3]},
-               {"topic":"foo2","partition":1,"replicas":[2,3]}]
-}
-
-Save this to use as the --reassignment-json-file option during rollback
-Successfully started reassignment of partitions
-{"version":1,
- "partitions":[{"topic":"foo1","partition":2,"replicas":[5,6]},
-               {"topic":"foo1","partition":0,"replicas":[5,6]},
-               {"topic":"foo2","partition":2,"replicas":[5,6]},
-               {"topic":"foo2","partition":0,"replicas":[5,6]},
-               {"topic":"foo1","partition":1,"replicas":[5,6]},
-               {"topic":"foo2","partition":1,"replicas":[5,6]}]
-}
-</pre>
-<p>
-Finally, the --verify option can be used with the tool to check the status of 
the partition reassignment. Note that the same expand-cluster-reassignment.json 
(used with the --execute option) should be used with the --verify option:
-<pre>
-> bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file expand-cluster-reassignment.json --verify
-Status of partition reassignment:
-Reassignment of partition [foo1,0] completed successfully
-Reassignment of partition [foo1,1] is in progress
-Reassignment of partition [foo1,2] is in progress
-Reassignment of partition [foo2,0] completed successfully
-Reassignment of partition [foo2,1] completed successfully
-Reassignment of partition [foo2,2] completed successfully
-</pre>
-
-<h5><a id="basic_ops_partitionassignment" 
href="#basic_ops_partitionassignment">Custom partition assignment and 
migration</a></h5>
-The partition reassignment tool can also be used to selectively move replicas 
of a partition to a specific set of brokers. When used in this manner, it is 
assumed that the user knows the reassignment plan and does not require the tool 
to generate a candidate reassignment, effectively skipping the --generate step 
and moving straight to the --execute step
-<p>
-For instance, the following example moves partition 0 of topic foo1 to brokers 
5,6 and partition 1 of topic foo2 to brokers 2,3:
-<p>
-The first step is to hand craft the custom reassignment plan in a json file:
-<pre>
-> cat custom-reassignment.json
-{"version":1,"partitions":[{"topic":"foo1","partition":0,"replicas":[5,6]},{"topic":"foo2","partition":1,"replicas":[2,3]}]}
-</pre>
-Then, use the json file with the --execute option to start the reassignment 
process:
-<pre>
-> bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file custom-reassignment.json --execute
-Current partition replica assignment
-
-{"version":1,
- "partitions":[{"topic":"foo1","partition":0,"replicas":[1,2]},
-               {"topic":"foo2","partition":1,"replicas":[3,4]}]
-}
-
-Save this to use as the --reassignment-json-file option during rollback
-Successfully started reassignment of partitions
-{"version":1,
- "partitions":[{"topic":"foo1","partition":0,"replicas":[5,6]},
-               {"topic":"foo2","partition":1,"replicas":[2,3]}]
-}
-</pre>
-<p>
-The --verify option can be used with the tool to check the status of the 
partition reassignment. Note that the same expand-cluster-reassignment.json 
(used with the --execute option) should be used with the --verify option:
-<pre>
-bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file custom-reassignment.json --verify
-Status of partition reassignment:
-Reassignment of partition [foo1,0] completed successfully
-Reassignment of partition [foo2,1] completed successfully
-</pre>
-
-<h4><a id="basic_ops_decommissioning_brokers" 
href="#basic_ops_decommissioning_brokers">Decommissioning brokers</a></h4>
-The partition reassignment tool does not have the ability to automatically 
generate a reassignment plan for decommissioning brokers yet. As such, the 
admin has to come up with a reassignment plan to move the replica for all 
partitions hosted on the broker to be decommissioned, to the rest of the 
brokers. This can be relatively tedious as the reassignment needs to ensure 
that all the replicas are not moved from the decommissioned broker to only one 
other broker. To make this process effortless, we plan to add tooling support 
for decommissioning brokers in the future.
-
-<h4><a id="basic_ops_increase_replication_factor" 
href="#basic_ops_increase_replication_factor">Increasing replication 
factor</a></h4>
-Increasing the replication factor of an existing partition is easy. Just 
specify the extra replicas in the custom reassignment json file and use it with 
the --execute option to increase the replication factor of the specified 
partitions.
-<p>
-For instance, the following example increases the replication factor of 
partition 0 of topic foo from 1 to 3. Before increasing the replication factor, 
the partition's only replica existed on broker 5. As part of increasing the 
replication factor, we will add more replicas on brokers 6 and 7.
-<p>
-The first step is to hand craft the custom reassignment plan in a json file:
-<pre>
-> cat increase-replication-factor.json
-{"version":1,
- "partitions":[{"topic":"foo","partition":0,"replicas":[5,6,7]}]}
-</pre>
-Then, use the json file with the --execute option to start the reassignment 
process:
-<pre>
-> bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file increase-replication-factor.json --execute
-Current partition replica assignment
-
-{"version":1,
- "partitions":[{"topic":"foo","partition":0,"replicas":[5]}]}
-
-Save this to use as the --reassignment-json-file option during rollback
-Successfully started reassignment of partitions
-{"version":1,
- "partitions":[{"topic":"foo","partition":0,"replicas":[5,6,7]}]}
-</pre>
-<p>
-The --verify option can be used with the tool to check the status of the 
partition reassignment. Note that the same increase-replication-factor.json 
(used with the --execute option) should be used with the --verify option:
-<pre>
-bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file increase-replication-factor.json --verify
-Status of partition reassignment:
-Reassignment of partition [foo,0] completed successfully
-</pre>
-You can also verify the increase in replication factor with the kafka-topics 
tool:
-<pre>
-> bin/kafka-topics.sh --zookeeper localhost:2181 --topic foo --describe
-Topic:foo      PartitionCount:1        ReplicationFactor:3     Configs:
-       Topic: foo      Partition: 0    Leader: 5       Replicas: 5,6,7 Isr: 
5,6,7
-</pre>
-
-<h4><a id="rep-throttle" href="#rep-throttle">Limiting Bandwidth Usage during 
Data Migration</a></h4>
-Kafka lets you apply a throttle to replication traffic, setting an upper bound 
on the bandwidth used to move replicas from machine to machine. This is useful 
when rebalancing a cluster, bootstrapping a new broker or adding or removing 
brokers, as it limits the impact these data-intensive operations will have on 
users.
-<p></p>
-There are two interfaces that can be used to engage a throttle. The simplest, 
and safest, is to apply a throttle when invoking the 
kafka-reassign-partitions.sh, but kafka-configs.sh can also be used to view and 
alter the throttle values directly.
-<p></p>
-So for example, if you were to execute a rebalance, with the below command, it 
would move partitions at no more than 50MB/s.
-<pre>$ bin/kafka-reassign-partitions.sh --zookeeper myhost:2181--execute 
--reassignment-json-file bigger-cluster.json —throttle 50000000</pre>
-When you execute this script you will see the throttle engage:
-<pre>
-The throttle limit was set to 50000000 B/s
-Successfully started reassignment of partitions.</pre>
-<p>Should you wish to alter the throttle, during a rebalance, say to increase 
the throughput so it completes quicker, you can do this by re-running the 
execute command passing the same reassignment-json-file:</p>
-<pre>$ bin/kafka-reassign-partitions.sh --zookeeper localhost:2181  --execute 
--reassignment-json-file bigger-cluster.json --throttle 700000000
-There is an existing assignment running.
-The throttle limit was set to 700000000 B/s</pre>
-
-<p>Once the rebalance completes the administrator can check the status of the 
rebalance using the --verify option.
-    If the rebalance has completed, the throttle will be removed via the 
--verify command. It is important that
-    administrators remove the throttle in a timely manner once rebalancing 
completes by running the command with
-    the --verify option. Failure to do so could cause regular replication 
traffic to be throttled. </p>
-<p>When the --verify option is executed, and the reassignment has completed, 
the script will confirm that the throttle was removed:</p>
-
-<pre>$ bin/kafka-reassign-partitions.sh --zookeeper localhost:2181  --verify 
--reassignment-json-file bigger-cluster.json
-Status of partition reassignment:
-Reassignment of partition [my-topic,1] completed successfully
-Reassignment of partition [mytopic,0] completed successfully
-Throttle was removed.</pre>
-
-<p>The administrator can also validate the assigned configs using the 
kafka-configs.sh. There are two pairs of throttle
-    configuration used to manage the throttling process. The throttle value 
itself. This is configured, at a broker
-    level, using the dynamic properties: </p>
-
-<pre>leader.replication.throttled.rate
-follower.replication.throttled.rate</pre>
-
-<p>There is also an enumerated set of throttled replicas: </p>
-
-<pre>leader.replication.throttled.replicas
-follower.replication.throttled.replicas</pre>
-
-<p>Which are configured per topic. All four config values are automatically 
assigned by kafka-reassign-partitions.sh
-    (discussed below). </p>
-<p>To view the throttle limit configuration:</p>
-
-<pre>$ bin/kafka-configs.sh --describe --zookeeper localhost:2181 
--entity-type brokers
-Configs for brokers '2' are 
leader.replication.throttled.rate=700000000,follower.replication.throttled.rate=700000000
-Configs for brokers '1' are 
leader.replication.throttled.rate=700000000,follower.replication.throttled.rate=700000000</pre>
-
-<p>This shows the throttle applied to both leader and follower side of the 
replication protocol. By default both sides
-    are assigned the same throttled throughput value. </p>
-
-<p>To view the list of throttled replicas:</p>
-
-<pre>$ bin/kafka-configs.sh --describe --zookeeper localhost:2181 
--entity-type topics
-Configs for topic 'my-topic' are 
leader.replication.throttled.replicas=1:102,0:101,
-    follower.replication.throttled.replicas=1:101,0:102</pre>
-
-<p>Here we see the leader throttle is applied to partition 1 on broker 102 and 
partition 0 on broker 101. Likewise the
-    follower throttle is applied to partition 1 on
-    broker 101 and partition 0 on broker 102. </p>
-
-<p>By default kafka-reassign-partitions.sh will apply the leader throttle to 
all replicas that exist before the
-    rebalance, any one of which might be leader.
-    It will apply the follower throttle to all move destinations. So if there 
is a partition with replicas on brokers
-    101,102, being reassigned to 102,103, a leader throttle,
-    for that partition, would be applied to 101,102 and a follower throttle 
would be applied to 103 only. </p>
-
-
-<p>If required, you can also use the --alter switch on kafka-configs.sh to 
alter the throttle configurations manually.
-</p>
-
-<h5>Safe usage of throttled replication</h5>
-
-<p>Some care should be taken when using throttled replication. In 
particular:</p>
-
-<p><i>(1) Throttle Removal:</i></p>
-The throttle should be removed in a timely manner once reassignment completes 
(by running kafka-reassign-partitions
-—verify).
-
-<p><i>(2) Ensuring Progress:</i></p>
-<p>If the throttle is set too low, in comparison to the incoming write rate, 
it is possible for replication to not
-    make progress. This occurs when:</p>
-<pre>max(BytesInPerSec) > throttle</pre>
-<p>
-    Where BytesInPerSec is the metric that monitors the write throughput of 
producers into each broker. </p>
-<p>The administrator can monitor whether replication is making progress, 
during the rebalance, using the metric:</p>
-
-<pre>kafka.server:type=FetcherLagMetrics,name=ConsumerLag,clientId=([-.\w]+),topic=([-.\w]+),partition=([0-9]+)</pre>
-
-<p>The lag should constantly decrease during replication.  If the metric does 
not decrease the administrator should
-    increase the
-    throttle throughput as described above. </p>
-
-
-<h4><a id="quotas" href="#quotas">Setting quotas</a></h4>
-Quotas overrides and defaults may be configured at (user, client-id), user or 
client-id levels as described <a href="#design_quotas">here</a>.
-By default, clients receive an unlimited quota.
-
-It is possible to set custom quotas for each (user, client-id), user or 
client-id group.
-<p>
-Configure custom quota for (user=user1, client-id=clientA):
-<pre>
-> bin/kafka-configs.sh  --zookeeper localhost:2181 --alter --add-config 
'producer_byte_rate=1024,consumer_byte_rate=2048' --entity-type users 
--entity-name user1 --entity-type clients --entity-name clientA
-Updated config for entity: user-principal 'user1', client-id 'clientA'.
-</pre>
-
-Configure custom quota for user=user1:
-<pre>
-> bin/kafka-configs.sh  --zookeeper localhost:2181 --alter --add-config 
'producer_byte_rate=1024,consumer_byte_rate=2048' --entity-type users 
--entity-name user1
-Updated config for entity: user-principal 'user1'.
-</pre>
-
-Configure custom quota for client-id=clientA:
-<pre>
-> bin/kafka-configs.sh  --zookeeper localhost:2181 --alter --add-config 
'producer_byte_rate=1024,consumer_byte_rate=2048' --entity-type clients 
--entity-name clientA
-Updated config for entity: client-id 'clientA'.
-</pre>
-
-It is possible to set default quotas for each (user, client-id), user or 
client-id group by specifying <i>--entity-default</i> option instead of 
<i>--entity-name</i>.
-<p>
-Configure default client-id quota for user=userA:
-<pre>
-> bin/kafka-configs.sh  --zookeeper localhost:2181 --alter --add-config 
'producer_byte_rate=1024,consumer_byte_rate=2048' --entity-type users 
--entity-name user1 --entity-type clients --entity-default
-Updated config for entity: user-principal 'user1', default client-id.
-</pre>
-
-Configure default quota for user:
-<pre>
-> bin/kafka-configs.sh  --zookeeper localhost:2181 --alter --add-config 
'producer_byte_rate=1024,consumer_byte_rate=2048' --entity-type users 
--entity-default
-Updated config for entity: default user-principal.
-</pre>
-
-Configure default quota for client-id:
-<pre>
-> bin/kafka-configs.sh  --zookeeper localhost:2181 --alter --add-config 
'producer_byte_rate=1024,consumer_byte_rate=2048' --entity-type clients 
--entity-default
-Updated config for entity: default client-id.
-</pre>
-
-Here's how to describe the quota for a given (user, client-id):
-<pre>
-> bin/kafka-configs.sh  --zookeeper localhost:2181 --describe --entity-type 
users --entity-name user1 --entity-type clients --entity-name clientA
-Configs for user-principal 'user1', client-id 'clientA' are 
producer_byte_rate=1024,consumer_byte_rate=2048
-</pre>
-Describe quota for a given user:
-<pre>
-> bin/kafka-configs.sh  --zookeeper localhost:2181 --describe --entity-type 
users --entity-name user1
-Configs for user-principal 'user1' are 
producer_byte_rate=1024,consumer_byte_rate=2048
-</pre>
-Describe quota for a given client-id:
-<pre>
-> bin/kafka-configs.sh  --zookeeper localhost:2181 --describe --entity-type 
clients --entity-name clientA
-Configs for client-id 'clientA' are 
producer_byte_rate=1024,consumer_byte_rate=2048
-</pre>
-If entity name is not specified, all entities of the specified type are 
described. For example, describe all users:
-<pre>
-> bin/kafka-configs.sh  --zookeeper localhost:2181 --describe --entity-type 
users
-Configs for user-principal 'user1' are 
producer_byte_rate=1024,consumer_byte_rate=2048
-Configs for default user-principal are 
producer_byte_rate=1024,consumer_byte_rate=2048
-</pre>
-Similarly for (user, client):
-<pre>
-> bin/kafka-configs.sh  --zookeeper localhost:2181 --describe --entity-type 
users --entity-type clients
-Configs for user-principal 'user1', default client-id are 
producer_byte_rate=1024,consumer_byte_rate=2048
-Configs for user-principal 'user1', client-id 'clientA' are 
producer_byte_rate=1024,consumer_byte_rate=2048
-</pre>
-<p>
-It is possible to set default quotas that apply to all client-ids by setting 
these configs on the brokers. These properties are applied only if quota 
overrides or defaults are not configured in Zookeeper. By default, each 
client-id receives an unlimited quota. The following sets the default quota per 
producer and consumer client-id to 10MB/sec.
-<pre>
-  quota.producer.default=10485760
-  quota.consumer.default=10485760
-</pre>
-Note that these properties are being deprecated and may be removed in a future 
release. Defaults configured using kafka-configs.sh take precedence over these 
properties.
-
-<h3><a id="datacenters" href="#datacenters">6.2 Datacenters</a></h3>
-
-Some deployments will need to manage a data pipeline that spans multiple 
datacenters. Our recommended approach to this is to deploy a local Kafka 
cluster in each datacenter with application instances in each datacenter 
interacting only with their local cluster and mirroring between clusters (see 
the documentation on the <a href="#basic_ops_mirror_maker">mirror maker 
tool</a> for how to do this).
-<p>
-This deployment pattern allows datacenters to act as independent entities and 
allows us to manage and tune inter-datacenter replication centrally. This 
allows each facility to stand alone and operate even if the inter-datacenter 
links are unavailable: when this occurs the mirroring falls behind until the 
link is restored at which time it catches up.
-<p>
-For applications that need a global view of all data you can use mirroring to 
provide clusters which have aggregate data mirrored from the local clusters in 
<i>all</i> datacenters. These aggregate clusters are used for reads by 
applications that require the full data set.
-<p>
-This is not the only possible deployment pattern. It is possible to read from 
or write to a remote Kafka cluster over the WAN, though obviously this will add 
whatever latency is required to get the cluster.
-<p>
-Kafka naturally batches data in both the producer and consumer so it can 
achieve high-throughput even over a high-latency connection. To allow this 
though it may be necessary to increase the TCP socket buffer sizes for the 
producer, consumer, and broker using the <code>socket.send.buffer.bytes</code> 
and <code>socket.receive.buffer.bytes</code> configurations. The appropriate 
way to set this is documented <a 
href="http://en.wikipedia.org/wiki/Bandwidth-delay_product";>here</a>.
-<p>
-It is generally <i>not</i> advisable to run a <i>single</i> Kafka cluster that 
spans multiple datacenters over a high-latency link. This will incur very high 
replication latency both for Kafka writes and ZooKeeper writes, and neither 
Kafka nor ZooKeeper will remain available in all locations if the network 
between locations is unavailable.
-
-<h3><a id="config" href="#config">6.3 Kafka Configuration</a></h3>
-
-<h4><a id="clientconfig" href="#clientconfig">Important Client 
Configurations</a></h4>
-The most important producer configurations control
-<ul>
-    <li>compression</li>
-    <li>sync vs async production</li>
-    <li>batch size (for async producers)</li>
-</ul>
-The most important consumer configuration is the fetch size.
-<p>
-All configurations are documented in the <a 
href="#configuration">configuration</a> section.
-<p>
-<h4><a id="prodconfig" href="#prodconfig">A Production Server Config</a></h4>
-Here is our production server configuration:
-<pre>
-# Replication configurations
-num.replica.fetchers=4
-replica.fetch.max.bytes=1048576
-replica.fetch.wait.max.ms=500
-replica.high.watermark.checkpoint.interval.ms=5000
-replica.socket.timeout.ms=30000
-replica.socket.receive.buffer.bytes=65536
-replica.lag.time.max.ms=10000
-
-controller.socket.timeout.ms=30000
-controller.message.queue.size=10
-
-# Log configuration
-num.partitions=8
-message.max.bytes=1000000
-auto.create.topics.enable=true
-log.index.interval.bytes=4096
-log.index.size.max.bytes=10485760
-log.retention.hours=168
-log.flush.interval.ms=10000
-log.flush.interval.messages=20000
-log.flush.scheduler.interval.ms=2000
-log.roll.hours=168
-log.retention.check.interval.ms=300000
-log.segment.bytes=1073741824
-
-# ZK configuration
-zookeeper.connection.timeout.ms=6000
-zookeeper.sync.time.ms=2000
-
-# Socket server configuration
-num.io.threads=8
-num.network.threads=8
-socket.request.max.bytes=104857600
-socket.receive.buffer.bytes=1048576
-socket.send.buffer.bytes=1048576
-queued.max.requests=16
-fetch.purgatory.purge.interval.requests=100
-producer.purgatory.purge.interval.requests=100
-</pre>
-
-Our client configuration varies a fair amount between different use cases.
-
-<h3><a id="java" href="#java">6.4 Java Version</a></h3>
-
-From a security perspective, we recommend you use the latest released version 
of JDK 1.8 as older freely available versions have disclosed security 
vulnerabilities.
-
-LinkedIn is currently running JDK 1.8 u5 (looking to upgrade to a newer 
version) with the G1 collector. If you decide to use the G1 collector (the 
current default) and you are still on JDK 1.7, make sure you are on u51 or 
newer. LinkedIn tried out u21 in testing, but they had a number of problems 
with the GC implementation in that version.
-
-LinkedIn's tuning looks like this:
-<pre>
--Xmx6g -Xms6g -XX:MetaspaceSize=96m -XX:+UseG1GC
--XX:MaxGCPauseMillis=20 -XX:InitiatingHeapOccupancyPercent=35 
-XX:G1HeapRegionSize=16M
--XX:MinMetaspaceFreeRatio=50 -XX:MaxMetaspaceFreeRatio=80
-</pre>
-
-For reference, here are the stats on one of LinkedIn's busiest clusters (at 
peak):
-<ul>
-    <li>60 brokers</li>
-    <li>50k partitions (replication factor 2)</li>
-    <li>800k messages/sec in</li>
-    <li>300 MB/sec inbound, 1 GB/sec+ outbound</li>
-</ul>
-
-The tuning looks fairly aggressive, but all of the brokers in that cluster 
have a 90% GC pause time of about 21ms, and they're doing less than 1 young GC 
per second.
-
-<h3><a id="hwandos" href="#hwandos">6.5 Hardware and OS</a></h3>
-We are using dual quad-core Intel Xeon machines with 24GB of memory.
-<p>
-You need sufficient memory to buffer active readers and writers. You can do a 
back-of-the-envelope estimate of memory needs by assuming you want to be able 
to buffer for 30 seconds and compute your memory need as write_throughput*30.
-<p>
-The disk throughput is important. We have 8x7200 rpm SATA drives. In general 
disk throughput is the performance bottleneck, and more disks is better. 
Depending on how you configure flush behavior you may or may not benefit from 
more expensive disks (if you force flush often then higher RPM SAS drives may 
be better).
-
-<h4><a id="os" href="#os">OS</a></h4>
-Kafka should run well on any unix system and has been tested on Linux and 
Solaris.
-<p>
-We have seen a few issues running on Windows and Windows is not currently a 
well supported platform though we would be happy to change that.
-<p>
-It is unlikely to require much OS-level tuning, but there are two potentially 
important OS-level configurations:
-<ul>
-    <li>File descriptor limits: Kafka uses file descriptors for log segments 
and open connections.  If a broker hosts many partitions, consider that the 
broker needs at least (number_of_partitions)*(partition_size/segment_size) to 
track all log segments in addition to the number of connections the broker 
makes.  We recommend at least 100000 allowed file descriptors for the broker 
processes as a starting point.
-    <li>Max socket buffer size: can be increased to enable high-performance 
data transfer between data centers as <a 
href="http://www.psc.edu/index.php/networking/641-tcp-tune";>described here</a>.
-</ul>
-<p>
-
-<h4><a id="diskandfs" href="#diskandfs">Disks and Filesystem</a></h4>
-We recommend using multiple drives to get good throughput and not sharing the 
same drives used for Kafka data with application logs or other OS filesystem 
activity to ensure good latency. You can either RAID these drives together into 
a single volume or format and mount each drive as its own directory. Since 
Kafka has replication the redundancy provided by RAID can also be provided at 
the application level. This choice has several tradeoffs.
-<p>
-If you configure multiple data directories partitions will be assigned 
round-robin to data directories. Each partition will be entirely in one of the 
data directories. If data is not well balanced among partitions this can lead 
to load imbalance between disks.
-<p>
-RAID can potentially do better at balancing load between disks (although it 
doesn't always seem to) because it balances load at a lower level. The primary 
downside of RAID is that it is usually a big performance hit for write 
throughput and reduces the available disk space.
-<p>
-Another potential benefit of RAID is the ability to tolerate disk failures. 
However our experience has been that rebuilding the RAID array is so I/O 
intensive that it effectively disables the server, so this does not provide 
much real availability improvement.
-
-<h4><a id="appvsosflush" href="#appvsosflush">Application vs. OS Flush 
Management</a></h4>
-Kafka always immediately writes all data to the filesystem and supports the 
ability to configure the flush policy that controls when data is forced out of 
the OS cache and onto disk using the flush. This flush policy can be controlled 
to force data to disk after a period of time or after a certain number of 
messages has been written. There are several choices in this configuration.
-<p>
-Kafka must eventually call fsync to know that data was flushed. When 
recovering from a crash for any log segment not known to be fsync'd Kafka will 
check the integrity of each message by checking its CRC and also rebuild the 
accompanying offset index file as part of the recovery process executed on 
startup.
-<p>
-Note that durability in Kafka does not require syncing data to disk, as a 
failed node will always recover from its replicas.
-<p>
-We recommend using the default flush settings which disable application fsync 
entirely. This means relying on the background flush done by the OS and Kafka's 
own background flush. This provides the best of all worlds for most uses: no 
knobs to tune, great throughput and latency, and full recovery guarantees. We 
generally feel that the guarantees provided by replication are stronger than 
sync to local disk, however the paranoid still may prefer having both and 
application level fsync policies are still supported.
-<p>
-The drawback of using application level flush settings is that it is less 
efficient in its disk usage pattern (it gives the OS less leeway to re-order 
writes) and it can introduce latency as fsync in most Linux filesystems blocks 
writes to the file whereas the background flushing does much more granular 
page-level locking.
-<p>
-In general you don't need to do any low-level tuning of the filesystem, but in 
the next few sections we will go over some of this in case it is useful.
-
-<h4><a id="linuxflush" href="#linuxflush">Understanding Linux OS Flush 
Behavior</a></h4>
-
-In Linux, data written to the filesystem is maintained in <a 
href="http://en.wikipedia.org/wiki/Page_cache";>pagecache</a> until it must be 
written out to disk (due to an application-level fsync or the OS's own flush 
policy). The flushing of data is done by a set of background threads called 
pdflush (or in post 2.6.32 kernels "flusher threads").
-<p>
-Pdflush has a configurable policy that controls how much dirty data can be 
maintained in cache and for how long before it must be written back to disk.
-This policy is described <a 
href="http://web.archive.org/web/20160518040713/http://www.westnet.com/~gsmith/content/linux-pdflush.htm";>here</a>.
-When Pdflush cannot keep up with the rate of data being written it will 
eventually cause the writing process to block incurring latency in the writes 
to slow down the accumulation of data.
-<p>
-You can see the current state of OS memory usage by doing
-<pre>
-  &gt; cat /proc/meminfo
-</pre>
-The meaning of these values are described in the link above.
-<p>
-Using pagecache has several advantages over an in-process cache for storing 
data that will be written out to disk:
-<ul>
-  <li>The I/O scheduler will batch together consecutive small writes into 
bigger physical writes which improves throughput.
-  <li>The I/O scheduler will attempt to re-sequence writes to minimize 
movement of the disk head which improves throughput.
-  <li>It automatically uses all the free memory on the machine
-</ul>
-
-<h4><a id="filesystems" href="#filesystems">Filesystem Selection</a></h4>
-<p>Kafka uses regular files on disk, and as such it has no hard dependency on 
a specific filesystem. The two filesystems which have the most usage, however, 
are EXT4 and XFS. Historically, EXT4 has had more usage, but recent 
improvements to the XFS filesystem have shown it to have better performance 
characteristics for Kafka's workload with no compromise in stability.</p>
-<p>Comparison testing was performed on a cluster with significant message 
loads, using a variety of filesystem creation and mount options. The primary 
metric in Kafka that was monitored was the "Request Local Time", indicating the 
amount of time append operations were taking. XFS resulted in much better local 
times (160ms vs. 250ms+ for the best EXT4 configuration), as well as lower 
average wait times. The XFS performance also showed less variability in disk 
performance.</p>
-<h5><a id="generalfs" href="#generalfs">General Filesystem Notes</a></h5>
-For any filesystem used for data directories, on Linux systems, the following 
options are recommended to be used at mount time:
-<ul>
-  <li>noatime: This option disables updating of a file's atime (last access 
time) attribute when the file is read. This can eliminate a significant number 
of filesystem writes, especially in the case of bootstrapping consumers. Kafka 
does not rely on the atime attributes at all, so it is safe to disable 
this.</li>
-</ul>
-<h5><a id="xfs" href="#xfs">XFS Notes</a></h5>
-The XFS filesystem has a significant amount of auto-tuning in place, so it 
does not require any change in the default settings, either at filesystem 
creation time or at mount. The only tuning parameters worth considering are:
-<ul>
-  <li>largeio: This affects the preferred I/O size reported by the stat call. 
While this can allow for higher performance on larger disk writes, in practice 
it had minimal or no effect on performance.</li>
-  <li>nobarrier: For underlying devices that have battery-backed cache, this 
option can provide a little more performance by disabling periodic write 
flushes. However, if the underlying device is well-behaved, it will report to 
the filesystem that it does not require flushes, and this option will have no 
effect.</li>
-</ul>
-<h5><a id="ext4" href="#ext4">EXT4 Notes</a></h5>
-EXT4 is a serviceable choice of filesystem for the Kafka data directories, 
however getting the most performance out of it will require adjusting several 
mount options. In addition, these options are generally unsafe in a failure 
scenario, and will result in much more data loss and corruption. For a single 
broker failure, this is not much of a concern as the disk can be wiped and the 
replicas rebuilt from the cluster. In a multiple-failure scenario, such as a 
power outage, this can mean underlying filesystem (and therefore data) 
corruption that is not easily recoverable. The following options can be 
adjusted:
-<ul>
-  <li>data=writeback: Ext4 defaults to data=ordered which puts a strong order 
on some writes. Kafka does not require this ordering as it does very paranoid 
data recovery on all unflushed log. This setting removes the ordering 
constraint and seems to significantly reduce latency.
-  <li>Disabling journaling: Journaling is a tradeoff: it makes reboots faster 
after server crashes but it introduces a great deal of additional locking which 
adds variance to write performance. Those who don't care about reboot time and 
want to reduce a major source of write latency spikes can turn off journaling 
entirely.
-  <li>commit=num_secs: This tunes the frequency with which ext4 commits to its 
metadata journal. Setting this to a lower value reduces the loss of unflushed 
data during a crash. Setting this to a higher value will improve throughput.
-  <li>nobh: This setting controls additional ordering guarantees when using 
data=writeback mode. This should be safe with Kafka as we do not depend on 
write ordering and improves throughput and latency.
-  <li>delalloc: Delayed allocation means that the filesystem avoid allocating 
any blocks until the physical write occurs. This allows ext4 to allocate a 
large extent instead of smaller pages and helps ensure the data is written 
sequentially. This feature is great for throughput. It does seem to involve 
some locking in the filesystem which adds a bit of latency variance.
-</ul>
-
-<h3><a id="monitoring" href="#monitoring">6.6 Monitoring</a></h3>
-
-Kafka uses Yammer Metrics for metrics reporting in both the server and the 
client. This can be configured to report stats using pluggable stats reporters 
to hook up to your monitoring system.
-<p>
-The easiest way to see the available metrics is to fire up jconsole and point 
it at a running kafka client or server; this will allow browsing all metrics 
with JMX.
-<p>
-We do graphing and alerting on the following metrics:
-<table class="data-table">
-<tbody><tr>
-      <th>Description</th>
-      <th>Mbean name</th>
-      <th>Normal value</th>
-    </tr>
-    <tr>
-      <td>Message in rate</td>
-      <td>kafka.server:type=BrokerTopicMetrics,name=MessagesInPerSec</td>
-      <td></td>
-    </tr>
-    <tr>
-      <td>Byte in rate</td>
-      <td>kafka.server:type=BrokerTopicMetrics,name=BytesInPerSec</td>
-      <td></td>
-    </tr>
-    <tr>
-      <td>Request rate</td>
-      
<td>kafka.network:type=RequestMetrics,name=RequestsPerSec,request={Produce|FetchConsumer|FetchFollower}</td>
-      <td></td>
-    </tr>
-    <tr>
-      <td>Byte out rate</td>
-      <td>kafka.server:type=BrokerTopicMetrics,name=BytesOutPerSec</td>
-      <td></td>
-    </tr>
-    <tr>
-      <td>Log flush rate and time</td>
-      <td>kafka.log:type=LogFlushStats,name=LogFlushRateAndTimeMs</td>
-      <td></td>
-    </tr>
-    <tr>
-      <td># of under replicated partitions (|ISR| &lt |all replicas|)</td>
-      <td>kafka.server:type=ReplicaManager,name=UnderReplicatedPartitions</td>
-      <td>0</td>
-    </tr>
-    <tr>
-      <td>Is controller active on broker</td>
-      <td>kafka.controller:type=KafkaController,name=ActiveControllerCount</td>
-      <td>only one broker in the cluster should have 1</td>
-    </tr>
-    <tr>
-      <td>Leader election rate</td>
-      
<td>kafka.controller:type=ControllerStats,name=LeaderElectionRateAndTimeMs</td>
-      <td>non-zero when there are broker failures</td>
-    </tr>
-    <tr>
-      <td>Unclean leader election rate</td>
-      
<td>kafka.controller:type=ControllerStats,name=UncleanLeaderElectionsPerSec</td>
-      <td>0</td>
-    </tr>
-    <tr>
-      <td>Partition counts</td>
-      <td>kafka.server:type=ReplicaManager,name=PartitionCount</td>
-      <td>mostly even across brokers</td>
-    </tr>
-    <tr>
-      <td>Leader replica counts</td>
-      <td>kafka.server:type=ReplicaManager,name=LeaderCount</td>
-      <td>mostly even across brokers</td>
-    </tr>
-    <tr>
-      <td>ISR shrink rate</td>
-      <td>kafka.server:type=ReplicaManager,name=IsrShrinksPerSec</td>
-      <td>If a broker goes down, ISR for some of the partitions will
-       shrink. When that broker is up again, ISR will be expanded
-       once the replicas are fully caught up. Other than that, the
-       expected value for both ISR shrink rate and expansion rate is 0. </td>
-    </tr>
-    <tr>
-      <td>ISR expansion rate</td>
-      <td>kafka.server:type=ReplicaManager,name=IsrExpandsPerSec</td>
-      <td>See above</td>
-    </tr>
-    <tr>
-      <td>Max lag in messages btw follower and leader replicas</td>
-      
<td>kafka.server:type=ReplicaFetcherManager,name=MaxLag,clientId=Replica</td>
-      <td>lag should be proportional to the maximum batch size of a produce 
request.</td>
-    </tr>
-    <tr>
-      <td>Lag in messages per follower replica</td>
-      
<td>kafka.server:type=FetcherLagMetrics,name=ConsumerLag,clientId=([-.\w]+),topic=([-.\w]+),partition=([0-9]+)</td>
-      <td>lag should be proportional to the maximum batch size of a produce 
request.</td>
-    </tr>
-    <tr>
-      <td>Requests waiting in the producer purgatory</td>
-      
<td>kafka.server:type=DelayedOperationPurgatory,name=PurgatorySize,delayedOperation=Produce</td>
-      <td>non-zero if ack=-1 is used</td>
-    </tr>
-    <tr>
-      <td>Requests waiting in the fetch purgatory</td>
-      
<td>kafka.server:type=DelayedOperationPurgatory,name=PurgatorySize,delayedOperation=Fetch</td>
-      <td>size depends on fetch.wait.max.ms in the consumer</td>
-    </tr>
-    <tr>
-      <td>Request total time</td>
-      
<td>kafka.network:type=RequestMetrics,name=TotalTimeMs,request={Produce|FetchConsumer|FetchFollower}</td>
-      <td>broken into queue, local, remote and response send time</td>
-    </tr>
-    <tr>
-      <td>Time the request waits in the request queue</td>
-       
<td>kafka.network:type=RequestMetrics,name=RequestQueueTimeMs,request={Produce|FetchConsumer|FetchFollower}</td>
-      <td></td>
-    </tr>
-    <tr>
-      <td>Time the request is processed at the leader</td>
-      
<td>kafka.network:type=RequestMetrics,name=LocalTimeMs,request={Produce|FetchConsumer|FetchFollower}</td>
-      <td></td>
-    </tr>
-    <tr>
-      <td>Time the request waits for the follower</td>
-      
<td>kafka.network:type=RequestMetrics,name=RemoteTimeMs,request={Produce|FetchConsumer|FetchFollower}</td>
-      <td>non-zero for produce requests when ack=-1</td>
-    </tr>
-    <tr>
-        <td>Time the request waits in the response queue</td>
-        
<td>kafka.network:type=RequestMetrics,name=ResponseQueueTimeMs,request={Produce|FetchConsumer|FetchFollower}</td>
+<script id="ops-template" type="text/x-handlebars-template">
+  Here is some information on actually running Kafka as a production system 
based on usage and experience at LinkedIn. Please send us any additional tips 
you know of.
+
+  <h3><a id="basic_ops" href="#basic_ops">6.1 Basic Kafka Operations</a></h3>
+
+  This section will review the most common operations you will perform on your 
Kafka cluster. All of the tools reviewed in this section are available under 
the <code>bin/</code> directory of the Kafka distribution and each tool will 
print details on all possible commandline options if it is run with no 
arguments.
+
+  <h4><a id="basic_ops_add_topic" href="#basic_ops_add_topic">Adding and 
removing topics</a></h4>
+
+  You have the option of either adding topics manually or having them be 
created automatically when data is first published to a non-existent topic. If 
topics are auto-created then you may want to tune the default <a 
href="#topic-config">topic configurations</a> used for auto-created topics.
+  <p>
+  Topics are added and modified using the topic tool:
+  <pre>
+  &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --create --topic 
my_topic_name
+        --partitions 20 --replication-factor 3 --config x=y
+  </pre>
+  The replication factor controls how many servers will replicate each message 
that is written. If you have a replication factor of 3 then up to 2 servers can 
fail before you will lose access to your data. We recommend you use a 
replication factor of 2 or 3 so that you can transparently bounce machines 
without interrupting data consumption.
+  <p>
+  The partition count controls how many logs the topic will be sharded into. 
There are several impacts of the partition count. First each partition must fit 
entirely on a single server. So if you have 20 partitions the full data set 
(and read and write load) will be handled by no more than 20 servers (no 
counting replicas). Finally the partition count impacts the maximum parallelism 
of your consumers. This is discussed in greater detail in the <a 
href="#intro_consumers">concepts section</a>.
+  <p>
+  Each sharded partition log is placed into its own folder under the Kafka log 
directory. The name of such folders consists of the topic name, appended by a 
dash (-) and the partition id. Since a typical folder name can not be over 255 
characters long, there will be a limitation on the length of topic names. We 
assume the number of partitions will not ever be above 100,000. Therefore, 
topic names cannot be longer than 249 characters. This leaves just enough room 
in the folder name for a dash and a potentially 5 digit long partition id.
+  <p>
+  The configurations added on the command line override the default settings 
the server has for things like the length of time data should be retained. The 
complete set of per-topic configurations is documented <a 
href="#topic-config">here</a>.
+
+  <h4><a id="basic_ops_modify_topic" href="#basic_ops_modify_topic">Modifying 
topics</a></h4>
+
+  You can change the configuration or partitioning of a topic using the same 
topic tool.
+  <p>
+  To add partitions you can do
+  <pre>
+  &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --alter --topic 
my_topic_name
+        --partitions 40
+  </pre>
+  Be aware that one use case for partitions is to semantically partition data, 
and adding partitions doesn't change the partitioning of existing data so this 
may disturb consumers if they rely on that partition. That is if data is 
partitioned by <code>hash(key) % number_of_partitions</code> then this 
partitioning will potentially be shuffled by adding partitions but Kafka will 
not attempt to automatically redistribute data in any way.
+  <p>
+  To add configs:
+  <pre>
+  &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --alter --topic 
my_topic_name --config x=y
+  </pre>
+  To remove a config:
+  <pre>
+  &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --alter --topic 
my_topic_name --delete-config x
+  </pre>
+  And finally deleting a topic:
+  <pre>
+  &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --delete --topic 
my_topic_name
+  </pre>
+  Topic deletion option is disabled by default. To enable it set the server 
config
+    <pre>delete.topic.enable=true</pre>
+  <p>
+  Kafka does not currently support reducing the number of partitions for a 
topic.
+  <p>
+  Instructions for changing the replication factor of a topic can be found <a 
href="#basic_ops_increase_replication_factor">here</a>.
+
+  <h4><a id="basic_ops_restarting" href="#basic_ops_restarting">Graceful 
shutdown</a></h4>
+
+  The Kafka cluster will automatically detect any broker shutdown or failure 
and elect new leaders for the partitions on that machine. This will occur 
whether a server fails or it is brought down intentionally for maintenance or 
configuration changes. For the latter cases Kafka supports a more graceful 
mechanism for stopping a server than just killing it.
+
+  When a server is stopped gracefully it has two optimizations it will take 
advantage of:
+  <ol>
+      <li>It will sync all its logs to disk to avoid needing to do any log 
recovery when it restarts (i.e. validating the checksum for all messages in the 
tail of the log). Log recovery takes time so this speeds up intentional 
restarts.
+      <li>It will migrate any partitions the server is the leader for to other 
replicas prior to shutting down. This will make the leadership transfer faster 
and minimize the time each partition is unavailable to a few milliseconds.
+  </ol>
+
+  Syncing the logs will happen automatically whenever the server is stopped 
other than by a hard kill, but the controlled leadership migration requires 
using a special setting:
+  <pre>
+      controlled.shutdown.enable=true
+  </pre>
+  Note that controlled shutdown will only succeed if <i>all</i> the partitions 
hosted on the broker have replicas (i.e. the replication factor is greater than 
1 <i>and</i> at least one of these replicas is alive). This is generally what 
you want since shutting down the last replica would make that topic partition 
unavailable.
+
+  <h4><a id="basic_ops_leader_balancing" 
href="#basic_ops_leader_balancing">Balancing leadership</a></h4>
+
+  Whenever a broker stops or crashes leadership for that broker's partitions 
transfers to other replicas. This means that by default when the broker is 
restarted it will only be a follower for all its partitions, meaning it will 
not be used for client reads and writes.
+  <p>
+  To avoid this imbalance, Kafka has a notion of preferred replicas. If the 
list of replicas for a partition is 1,5,9 then node 1 is preferred as the 
leader to either node 5 or 9 because it is earlier in the replica list. You can 
have the Kafka cluster try to restore leadership to the restored replicas by 
running the command:
+  <pre>
+  &gt; bin/kafka-preferred-replica-election.sh --zookeeper zk_host:port/chroot
+  </pre>
+
+  Since running this command can be tedious you can also configure Kafka to do 
this automatically by setting the following configuration:
+  <pre>
+      auto.leader.rebalance.enable=true
+  </pre>
+
+  <h4><a id="basic_ops_racks" href="#basic_ops_racks">Balancing Replicas 
Across Racks</a></h4>
+  The rack awareness feature spreads replicas of the same partition across 
different racks. This extends the guarantees Kafka provides for broker-failure 
to cover rack-failure, limiting the risk of data loss should all the brokers on 
a rack fail at once. The feature can also be applied to other broker groupings 
such as availability zones in EC2.
+  <p></p>
+  You can specify that a broker belongs to a particular rack by adding a 
property to the broker config:
+  <pre>   broker.rack=my-rack-id</pre>
+  When a topic is <a href="#basic_ops_add_topic">created</a>, <a 
href="#basic_ops_modify_topic">modified</a> or replicas are <a 
href="#basic_ops_cluster_expansion">redistributed</a>, the rack constraint will 
be honoured, ensuring replicas span as many racks as they can (a partition will 
span min(#racks, replication-factor) different racks).
+  <p></p>
+  The algorithm used to assign replicas to brokers ensures that the number of 
leaders per broker will be constant, regardless of how brokers are distributed 
across racks. This ensures balanced throughput.
+  <p></p>
+  However if racks are assigned different numbers of brokers, the assignment 
of replicas will not be even. Racks with fewer brokers will get more replicas, 
meaning they will use more storage and put more resources into replication. 
Hence it is sensible to configure an equal number of brokers per rack.
+
+  <h4><a id="basic_ops_mirror_maker" href="#basic_ops_mirror_maker">Mirroring 
data between clusters</a></h4>
+
+  We refer to the process of replicating data <i>between</i> Kafka clusters 
"mirroring" to avoid confusion with the replication that happens amongst the 
nodes in a single cluster. Kafka comes with a tool for mirroring data between 
Kafka clusters. The tool consumes from a source cluster and produces to a 
destination cluster.
+
+  A common use case for this kind of mirroring is to provide a replica in 
another datacenter. This scenario will be discussed in more detail in the next 
section.
+  <p>
+  You can run many such mirroring processes to increase throughput and for 
fault-tolerance (if one process dies, the others will take overs the additional 
load).
+  <p>
+  Data will be read from topics in the source cluster and written to a topic 
with the same name in the destination cluster. In fact the mirror maker is 
little more than a Kafka consumer and producer hooked together.
+  <p>
+  The source and destination clusters are completely independent entities: 
they can have different numbers of partitions and the offsets will not be the 
same. For this reason the mirror cluster is not really intended as a 
fault-tolerance mechanism (as the consumer position will be different); for 
that we recommend using normal in-cluster replication. The mirror maker process 
will, however, retain and use the message key for partitioning so order is 
preserved on a per-key basis.
+  <p>
+  Here is an example showing how to mirror a single topic (named 
<i>my-topic</i>) from an input cluster:
+  <pre>
+  &gt; bin/kafka-mirror-maker.sh
+        --consumer.config consumer.properties
+        --producer.config producer.properties --whitelist my-topic
+  </pre>
+  Note that we specify the list of topics with the <code>--whitelist</code> 
option. This option allows any regular expression using <a 
href="http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html";>Java-style
 regular expressions</a>. So you could mirror two topics named <i>A</i> and 
<i>B</i> using <code>--whitelist 'A|B'</code>. Or you could mirror <i>all</i> 
topics using <code>--whitelist '*'</code>. Make sure to quote any regular 
expression to ensure the shell doesn't try to expand it as a file path. For 
convenience we allow the use of ',' instead of '|' to specify a list of topics.
+  <p>
+  Sometimes it is easier to say what it is that you <i>don't</i> want. Instead 
of using <code>--whitelist</code> to say what you want
+  to mirror you can use <code>--blacklist</code> to say what to exclude. This 
also takes a regular expression argument.
+  However, <code>--blacklist</code> is not supported when the new consumer has 
been enabled (i.e. when <code>bootstrap.servers</code>
+  has been defined in the consumer configuration).
+  <p>
+  Combining mirroring with the configuration 
<code>auto.create.topics.enable=true</code> makes it possible to have a replica 
cluster that will automatically create and replicate all data in a source 
cluster even as new topics are added.
+
+  <h4><a id="basic_ops_consumer_lag" href="#basic_ops_consumer_lag">Checking 
consumer position</a></h4>
+  Sometimes it's useful to see the position of your consumers. We have a tool 
that will show the position of all consumers in a consumer group as well as how 
far behind the end of the log they are. To run this tool on a consumer group 
named <i>my-group</i> consuming a topic named <i>my-topic</i> would look like 
this:
+  <pre>
+  &gt; bin/kafka-run-class.sh kafka.tools.ConsumerOffsetChecker --zookeeper 
localhost:2181 --group test
+  Group           Topic                          Pid Offset          logSize   
      Lag             Owner
+  my-group        my-topic                       0   0               0         
      0               test_jkreps-mn-1394154511599-60744496-0
+  my-group        my-topic                       1   0               0         
      0               test_jkreps-mn-1394154521217-1a0be913-0
+  </pre>
+
+
+  NOTE: Since 0.9.0.0, the kafka.tools.ConsumerOffsetChecker tool has been 
deprecated. You should use the kafka.admin.ConsumerGroupCommand (or the 
bin/kafka-consumer-groups.sh script) to manage consumer groups, including 
consumers created with the <a 
href="http://kafka.apache.org/documentation.html#newconsumerapi";>new consumer 
API</a>.
+
+  <h4><a id="basic_ops_consumer_group" 
href="#basic_ops_consumer_group">Managing Consumer Groups</a></h4>
+
+  With the ConsumerGroupCommand tool, we can list, describe, or delete 
consumer groups. Note that deletion is only available when the group metadata 
is stored in
+  ZooKeeper. When using the <a 
href="http://kafka.apache.org/documentation.html#newconsumerapi";>new consumer 
API</a> (where
+  the broker handles coordination of partition handling and rebalance), the 
group is deleted when the last committed offset for that group expires.
+
+  For example, to list all consumer groups across all topics:
+
+  <pre>
+  &gt; bin/kafka-consumer-groups.sh --bootstrap-server broker1:9092 --list
+
+  test-consumer-group
+  </pre>
+
+  To view offsets as in the previous example with the ConsumerOffsetChecker, 
we "describe" the consumer group like this:
+
+  <pre>
+  &gt; bin/kafka-consumer-groups.sh --bootstrap-server broker1:9092 --describe 
--group test-consumer-group
+
+  GROUP                          TOPIC                          PARTITION  
CURRENT-OFFSET  LOG-END-OFFSET  LAG             OWNER
+  test-consumer-group            test-foo                       0          1   
            3               2               consumer-1_/127.0.0.1
+  </pre>
+
+  If you are using the old high-level consumer and storing the group metadata 
in ZooKeeper (i.e. <code>offsets.storage=zookeeper</code>), pass
+  <code>--zookeeper</code> instead of <code>bootstrap-server</code>:
+
+  <pre>
+  &gt; bin/kafka-consumer-groups.sh --zookeeper localhost:2181 --list
+  </pre>
+
+  <h4><a id="basic_ops_cluster_expansion" 
href="#basic_ops_cluster_expansion">Expanding your cluster</a></h4>
+
+  Adding servers to a Kafka cluster is easy, just assign them a unique broker 
id and start up Kafka on your new servers. However these new servers will not 
automatically be assigned any data partitions, so unless partitions are moved 
to them they won't be doing any work until new topics are created. So usually 
when you add machines to your cluster you will want to migrate some existing 
data to these machines.
+  <p>
+  The process of migrating data is manually initiated but fully automated. 
Under the covers what happens is that Kafka will add the new server as a 
follower of the partition it is migrating and allow it to fully replicate the 
existing data in that partition. When the new server has fully replicated the 
contents of this partition and joined the in-sync replica one of the existing 
replicas will delete their partition's data.
+  <p>
+  The partition reassignment tool can be used to move partitions across 
brokers. An ideal partition distribution would ensure even data load and 
partition sizes across all brokers. The partition reassignment tool does not 
have the capability to automatically study the data distribution in a Kafka 
cluster and move partitions around to attain an even load distribution. As 
such, the admin has to figure out which topics or partitions should be moved 
around.
+  <p>
+  The partition reassignment tool can run in 3 mutually exclusive modes:
+  <ul>
+  <li>--generate: In this mode, given a list of topics and a list of brokers, 
the tool generates a candidate reassignment to move all partitions of the 
specified topics to the new brokers. This option merely provides a convenient 
way to generate a partition reassignment plan given a list of topics and target 
brokers.</li>
+  <li>--execute: In this mode, the tool kicks off the reassignment of 
partitions based on the user provided reassignment plan. (using the 
--reassignment-json-file option). This can either be a custom reassignment plan 
hand crafted by the admin or provided by using the --generate option</li>
+  <li>--verify: In this mode, the tool verifies the status of the reassignment 
for all partitions listed during the last --execute. The status can be either 
of successfully completed, failed or in progress</li>
+  </ul>
+  <h5><a id="basic_ops_automigrate" 
href="#basic_ops_automigrate">Automatically migrating data to new 
machines</a></h5>
+  The partition reassignment tool can be used to move some topics off of the 
current set of brokers to the newly added brokers. This is typically useful 
while expanding an existing cluster since it is easier to move entire topics to 
the new set of brokers, than moving one partition at a time. When used to do 
this, the user should provide a list of topics that should be moved to the new 
set of brokers and a target list of new brokers. The tool then evenly 
distributes all partitions for the given list of topics across the new set of 
brokers. During this move, the replication factor of the topic is kept 
constant. Effectively the replicas for all partitions for the input list of 
topics are moved from the old set of brokers to the newly added brokers.
+  <p>
+  For instance, the following example will move all partitions for topics 
foo1,foo2 to the new set of brokers 5,6. At the end of this move, all 
partitions for topics foo1 and foo2 will <i>only</i> exist on brokers 5,6.
+  <p>
+  Since the tool accepts the input list of topics as a json file, you first 
need to identify the topics you want to move and create the json file as 
follows:
+  <pre>
+  > cat topics-to-move.json
+  {"topics": [{"topic": "foo1"},
+              {"topic": "foo2"}],
+  "version":1
+  }
+  </pre>
+  Once the json file is ready, use the partition reassignment tool to generate 
a candidate assignment:
+  <pre>
+  > bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--topics-to-move-json-file topics-to-move.json --broker-list "5,6" --generate
+  Current partition replica assignment
+
+  {"version":1,
+  "partitions":[{"topic":"foo1","partition":2,"replicas":[1,2]},
+                {"topic":"foo1","partition":0,"replicas":[3,4]},
+                {"topic":"foo2","partition":2,"replicas":[1,2]},
+                {"topic":"foo2","partition":0,"replicas":[3,4]},
+                {"topic":"foo1","partition":1,"replicas":[2,3]},
+                {"topic":"foo2","partition":1,"replicas":[2,3]}]
+  }
+
+  Proposed partition reassignment configuration
+
+  {"version":1,
+  "partitions":[{"topic":"foo1","partition":2,"replicas":[5,6]},
+                {"topic":"foo1","partition":0,"replicas":[5,6]},
+                {"topic":"foo2","partition":2,"replicas":[5,6]},
+                {"topic":"foo2","partition":0,"replicas":[5,6]},
+                {"topic":"foo1","partition":1,"replicas":[5,6]},
+                {"topic":"foo2","partition":1,"replicas":[5,6]}]
+  }
+  </pre>
+  <p>
+  The tool generates a candidate assignment that will move all partitions from 
topics foo1,foo2 to brokers 5,6. Note, however, that at this point, the 
partition movement has not started, it merely tells you the current assignment 
and the proposed new assignment. The current assignment should be saved in case 
you want to rollback to it. The new assignment should be saved in a json file 
(e.g. expand-cluster-reassignment.json) to be input to the tool with the 
--execute option as follows:
+  <pre>
+  > bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file expand-cluster-reassignment.json --execute
+  Current partition replica assignment
+
+  {"version":1,
+  "partitions":[{"topic":"foo1","partition":2,"replicas":[1,2]},
+                {"topic":"foo1","partition":0,"replicas":[3,4]},
+                {"topic":"foo2","partition":2,"replicas":[1,2]},
+                {"topic":"foo2","partition":0,"replicas":[3,4]},
+                {"topic":"foo1","partition":1,"replicas":[2,3]},
+                {"topic":"foo2","partition":1,"replicas":[2,3]}]
+  }
+
+  Save this to use as the --reassignment-json-file option during rollback
+  Successfully started reassignment of partitions
+  {"version":1,
+  "partitions":[{"topic":"foo1","partition":2,"replicas":[5,6]},
+                {"topic":"foo1","partition":0,"replicas":[5,6]},
+                {"topic":"foo2","partition":2,"replicas":[5,6]},
+                {"topic":"foo2","partition":0,"replicas":[5,6]},
+                {"topic":"foo1","partition":1,"replicas":[5,6]},
+                {"topic":"foo2","partition":1,"replicas":[5,6]}]
+  }
+  </pre>
+  <p>
+  Finally, the --verify option can be used with the tool to check the status 
of the partition reassignment. Note that the same 
expand-cluster-reassignment.json (used with the --execute option) should be 
used with the --verify option:
+  <pre>
+  > bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file expand-cluster-reassignment.json --verify
+  Status of partition reassignment:
+  Reassignment of partition [foo1,0] completed successfully
+  Reassignment of partition [foo1,1] is in progress
+  Reassignment of partition [foo1,2] is in progress
+  Reassignment of partition [foo2,0] completed successfully
+  Reassignment of partition [foo2,1] completed successfully
+  Reassignment of partition [foo2,2] completed successfully
+  </pre>
+
+  <h5><a id="basic_ops_partitionassignment" 
href="#basic_ops_partitionassignment">Custom partition assignment and 
migration</a></h5>
+  The partition reassignment tool can also be used to selectively move 
replicas of a partition to a specific set of brokers. When used in this manner, 
it is assumed that the user knows the reassignment plan and does not require 
the tool to generate a candidate reassignment, effectively skipping the 
--generate step and moving straight to the --execute step
+  <p>
+  For instance, the following example moves partition 0 of topic foo1 to 
brokers 5,6 and partition 1 of topic foo2 to brokers 2,3:
+  <p>
+  The first step is to hand craft the custom reassignment plan in a json file:
+  <pre>
+  > cat custom-reassignment.json
+  
{"version":1,"partitions":[{"topic":"foo1","partition":0,"replicas":[5,6]},{"topic":"foo2","partition":1,"replicas":[2,3]}]}
+  </pre>
+  Then, use the json file with the --execute option to start the reassignment 
process:
+  <pre>
+  > bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file custom-reassignment.json --execute
+  Current partition replica assignment
+
+  {"version":1,
+  "partitions":[{"topic":"foo1","partition":0,"replicas":[1,2]},
+                {"topic":"foo2","partition":1,"replicas":[3,4]}]
+  }
+
+  Save this to use as the --reassignment-json-file option during rollback
+  Successfully started reassignment of partitions
+  {"version":1,
+  "partitions":[{"topic":"foo1","partition":0,"replicas":[5,6]},
+                {"topic":"foo2","partition":1,"replicas":[2,3]}]
+  }
+  </pre>
+  <p>
+  The --verify option can be used with the tool to check the status of the 
partition reassignment. Note that the same expand-cluster-reassignment.json 
(used with the --execute option) should be used with the --verify option:
+  <pre>
+  bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file custom-reassignment.json --verify
+  Status of partition reassignment:
+  Reassignment of partition [foo1,0] completed successfully
+  Reassignment of partition [foo2,1] completed successfully
+  </pre>
+
+  <h4><a id="basic_ops_decommissioning_brokers" 
href="#basic_ops_decommissioning_brokers">Decommissioning brokers</a></h4>
+  The partition reassignment tool does not have the ability to automatically 
generate a reassignment plan for decommissioning brokers yet. As such, the 
admin has to come up with a reassignment plan to move the replica for all 
partitions hosted on the broker to be decommissioned, to the rest of the 
brokers. This can be relatively tedious as the reassignment needs to ensure 
that all the replicas are not moved from the decommissioned broker to only one 
other broker. To make this process effortless, we plan to add tooling support 
for decommissioning brokers in the future.
+
+  <h4><a id="basic_ops_increase_replication_factor" 
href="#basic_ops_increase_replication_factor">Increasing replication 
factor</a></h4>
+  Increasing the replication factor of an existing partition is easy. Just 
specify the extra replicas in the custom reassignment json file and use it with 
the --execute option to increase the replication factor of the specified 
partitions.
+  <p>
+  For instance, the following example increases the replication factor of 
partition 0 of topic foo from 1 to 3. Before increasing the replication factor, 
the partition's only replica existed on broker 5. As part of increasing the 
replication factor, we will add more replicas on brokers 6 and 7.
+  <p>
+  The first step is to hand craft the custom reassignment plan in a json file:
+  <pre>
+  > cat increase-replication-factor.json
+  {"version":1,
+  "partitions":[{"topic":"foo","partition":0,"replicas":[5,6,7]}]}
+  </pre>
+  Then, use the json file with the --execute option to start the reassignment 
process:
+  <pre>
+  > bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file increase-replication-factor.json --execute
+  Current partition replica assignment
+
+  {"version":1,
+  "partitions":[{"topic":"foo","partition":0,"replicas":[5]}]}
+
+  Save this to use as the --reassignment-json-file option during rollback
+  Successfully started reassignment of partitions
+  {"version":1,
+  "partitions":[{"topic":"foo","partition":0,"replicas":[5,6,7]}]}
+  </pre>
+  <p>
+  The --verify option can be used with the tool to check the status of the 
partition reassignment. Note that the same increase-replication-factor.json 
(used with the --execute option) should be used with the --verify option:
+  <pre>
+  bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file increase-replication-factor.json --verify
+  Status of partition reassignment:
+  Reassignment of partition [foo,0] completed successfully
+  </pre>
+  You can also verify the increase in replication factor with the kafka-topics 
tool:
+  <pre>
+  > bin/kafka-topics.sh --zookeeper localhost:2181 --topic foo --describe
+  Topic:foo    PartitionCount:1        ReplicationFactor:3     Configs:
+    Topic: foo Partition: 0    Leader: 5       Replicas: 5,6,7 Isr: 5,6,7
+  </pre>
+
+  <h4><a id="rep-throttle" href="#rep-throttle">Limiting Bandwidth Usage 
during Data Migration</a></h4>
+  Kafka lets you apply a throttle to replication traffic, setting an upper 
bound on the bandwidth used to move replicas from machine to machine. This is 
useful when rebalancing a cluster, bootstrapping a new broker or adding or 
removing brokers, as it limits the impact these data-intensive operations will 
have on users.
+  <p></p>
+  There are two interfaces that can be used to engage a throttle. The 
simplest, and safest, is to apply a throttle when invoking the 
kafka-reassign-partitions.sh, but kafka-configs.sh can also be used to view and 
alter the throttle values directly.
+  <p></p>
+  So for example, if you were to execute a rebalance, with the below command, 
it would move partitions at no more than 50MB/s.
+  <pre>$ bin/kafka-reassign-partitions.sh --zookeeper myhost:2181--execute 
--reassignment-json-file bigger-cluster.json —throttle 50000000</pre>
+  When you execute this script you will see the throttle engage:
+  <pre>
+  The throttle limit was set to 50000000 B/s
+  Successfully started reassignment of partitions.</pre>
+  <p>Should you wish to alter the throttle, during a rebalance, say to 
increase the throughput so it completes quicker, you can do this by re-running 
the execute command passing the same reassignment-json-file:</p>
+  <pre>$ bin/kafka-reassign-partitions.sh --zookeeper localhost:2181  
--execute --reassignment-json-file bigger-cluster.json --throttle 700000000
+  There is an existing assignment running.
+  The throttle limit was set to 700000000 B/s</pre>
+
+  <p>Once the rebalance completes the administrator can check the status of 
the rebalance using the --verify option.
+      If the rebalance has completed, the throttle will be removed via the 
--verify command. It is important that
+      administrators remove the throttle in a timely manner once rebalancing 
completes by running the command with
+      the --verify option. Failure to do so could cause regular replication 
traffic to be throttled. </p>
+  <p>When the --verify option is executed, and the reassignment has completed, 
the script will confirm that the throttle was removed:</p>
+
+  <pre>$ bin/kafka-reassign-partitions.sh --zookeeper localhost:2181  --verify 
--reassignment-json-file bigger-cluster.json
+  Status of partition reassignment:
+  Reassignment of partition [my-topic,1] completed successfully
+  Reassignment of partition [mytopic,0] completed successfully
+  Throttle was removed.</pre>
+
+  <p>The administrator can also validate the assigned configs using the 
kafka-configs.sh. There are two pairs of throttle
+      configuration used to manage the throttling process. The throttle value 
itself. This is configured, at a broker
+      level, using the dynamic properties: </p>
+
+  <pre>leader.replication.throttled.rate
+  follower.replication.throttled.rate</pre>
+
+  <p>There is also an enumerated set of throttled replicas: </p>
+
+  <pre>leader.replication.throttled.replicas
+  follower.replication.throttled.replicas</pre>
+
+  <p>Which are configured per topic. All four config values are automatically 
assigned by kafka-reassign-partitions.sh
+      (discussed below). </p>
+  <p>To view the throttle limit configuration:</p>
+
+  <pre>$ bin/kafka-configs.sh --describe --zookeeper localhost:2181 
--entity-type brokers
+  Configs for brokers '2' are 
leader.replication.throttled.rate=700000000,follower.replication.throttled.rate=700000000
+  Configs for brokers '1' are 
leader.replication.throttled.rate=700000000,follower.replication.throttled.rate=700000000</pre>
+
+  <p>This shows the throttle applied to both leader and follower side of the 
replication protocol. By default both sides
+      are assigned the same throttled throughput value. </p>
+
+  <p>

<TRUNCATED>

Reply via email to