http://git-wip-us.apache.org/repos/asf/kafka-site/blob/ba6c994c/090/ops.html
----------------------------------------------------------------------
diff --git a/090/ops.html b/090/ops.html
new file mode 100644
index 0000000..b539b5c
--- /dev/null
+++ b/090/ops.html
@@ -0,0 +1,859 @@
+Here is some information on actually running Kafka as a production system 
based on usage and experience at LinkedIn. Please send us any additional tips 
you know of.
+
+<h3><a id="basic_ops">6.1 Basic Kafka Operations</a></h3>
+
+This section will review the most common operations you will perform on your 
Kafka cluster. All of the tools reviewed in this section are available under 
the <code>bin/</code> directory of the Kafka distribution and each tool will 
print details on all possible commandline options if it is run with no 
arguments.
+    
+<h4><a id="basic_ops_add_topic">Adding and removing topics</a></h4>
+
+You have the option of either adding topics manually or having them be created 
automatically when data is first published to a non-existent topic. If topics 
are auto-created then you may want to tune the default <a 
href="#topic-config">topic configurations</a> used for auto-created topics.
+<p>
+Topics are added and modified using the topic tool:
+<pre>
+ &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --create --topic 
my_topic_name 
+       --partitions 20 --replication-factor 3 --config x=y
+</pre>
+The replication factor controls how many servers will replicate each message 
that is written. If you have a replication factor of 3 then up to 2 servers can 
fail before you will lose access to your data. We recommend you use a 
replication factor of 2 or 3 so that you can transparently bounce machines 
without interrupting data consumption.
+<p>
+The partition count controls how many logs the topic will be sharded into. 
There are several impacts of the partition count. First each partition must fit 
entirely on a single server. So if you have 20 partitions the full data set 
(and read and write load) will be handled by no more than 20 servers (no 
counting replicas). Finally the partition count impacts the maximum parallelism 
of your consumers. This is discussed in greater detail in the <a 
href="#intro_consumers">concepts section</a>.
+<p>
+The configurations added on the command line override the default settings the 
server has for things like the length of time data should be retained. The 
complete set of per-topic configurations is documented <a 
href="#topic-config">here</a>.
+
+<h4><a id="basic_ops_modify_topic">Modifying topics</a></h4>
+
+You can change the configuration or partitioning of a topic using the same 
topic tool.
+<p>
+To add partitions you can do
+<pre>
+ &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --alter --topic 
my_topic_name 
+       --partitions 40 
+</pre>
+Be aware that one use case for partitions is to semantically partition data, 
and adding partitions doesn't change the partitioning of existing data so this 
may disturb consumers if they rely on that partition. That is if data is 
partitioned by <code>hash(key) % number_of_partitions</code> then this 
partitioning will potentially be shuffled by adding partitions but Kafka will 
not attempt to automatically redistribute data in any way.
+<p>
+To add configs:
+<pre>
+ &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --alter --topic 
my_topic_name --config x=y
+</pre>
+To remove a config:
+<pre>
+ &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --alter --topic 
my_topic_name --deleteConfig x
+</pre>
+And finally deleting a topic:
+<pre>
+ &gt; bin/kafka-topics.sh --zookeeper zk_host:port/chroot --delete --topic 
my_topic_name
+</pre>
+Topic deletion option is disabled by default. To enable it set the server 
config
+  <pre>delete.topic.enable=true</pre>
+<p>
+Kafka does not currently support reducing the number of partitions for a topic 
or changing the replication factor.
+
+<h4><a id="basic_ops_restarting">Graceful shutdown</a></h4>
+
+The Kafka cluster will automatically detect any broker shutdown or failure and 
elect new leaders for the partitions on that machine. This will occur whether a 
server fails or it is brought down intentionally for maintenance or 
configuration changes. For the later cases Kafka supports a more graceful 
mechanism for stoping a server then just killing it.
+
+When a server is stopped gracefully it has two optimizations it will take 
advantage of:
+<ol>
+    <li>It will sync all its logs to disk to avoid needing to do any log 
recovery when it restarts (i.e. validating the checksum for all messages in the 
tail of the log). Log recovery takes time so this speeds up intentional 
restarts.
+    <li>It will migrate any partitions the server is the leader for to other 
replicas prior to shutting down. This will make the leadership transfer faster 
and minimize the time each partition is unavailable to a few milliseconds.
+</ol>
+
+Syncing the logs will happen automatically happen whenever the server is 
stopped other than by a hard kill, but the controlled leadership migration 
requires using a special setting:
+<pre>
+    controlled.shutdown.enable=true
+</pre>
+Note that controlled shutdown will only succeed if <i>all</i> the partitions 
hosted on the broker have replicas (i.e. the replication factor is greater than 
1 <i>and</i> at least one of these replicas is alive). This is generally what 
you want since shutting down the last replica would make that topic partition 
unavailable.
+
+<h4><a id="basic_ops_leader_balancing">Balancing leadership</a></h4>
+
+Whenever a broker stops or crashes leadership for that broker's partitions 
transfers to other replicas. This means that by default when the broker is 
restarted it will only be a follower for all its partitions, meaning it will 
not be used for client reads and writes.
+<p>
+To avoid this imbalance, Kafka has a notion of preferred replicas. If the list 
of replicas for a partition is 1,5,9 then node 1 is preferred as the leader to 
either node 5 or 9 because it is earlier in the replica list. You can have the 
Kafka cluster try to restore leadership to the restored replicas by running the 
command:
+<pre>
+ &gt; bin/kafka-preferred-replica-election.sh --zookeeper zk_host:port/chroot
+</pre>
+
+Since running this command can be tedious you can also configure Kafka to do 
this automatically by setting the following configuration:
+<pre>
+    auto.leader.rebalance.enable=true
+</pre>
+
+<h4><a id="basic_ops_mirror_maker">Mirroring data between clusters</a></h4>
+
+We refer to the process of replicating data <i>between</i> Kafka clusters 
"mirroring" to avoid confusion with the replication that happens amongst the 
nodes in a single cluster. Kafka comes with a tool for mirroring data between 
Kafka clusters. The tool reads from one or more source clusters and writes to a 
destination cluster, like this:
+<p>
+<img src="/images/mirror-maker.png">
+<p>
+A common use case for this kind of mirroring is to provide a replica in 
another datacenter. This scenario will be discussed in more detail in the next 
section.
+<p>
+You can run many such mirroring processes to increase throughput and for 
fault-tolerance (if one process dies, the others will take overs the additional 
load).
+<p>
+Data will be read from topics in the source cluster and written to a topic 
with the same name in the destination cluster. In fact the mirror maker is 
little more than a Kafka consumer and producer hooked together.
+<p>
+The source and destination clusters are completely independent entities: they 
can have different numbers of partitions and the offsets will not be the same. 
For this reason the mirror cluster is not really intended as a fault-tolerance 
mechanism (as the consumer position will be different); for that we recommend 
using normal in-cluster replication. The mirror maker process will, however, 
retain and use the message key for partitioning so order is preserved on a 
per-key basis.
+<p>
+Here is an example showing how to mirror a single topic (named 
<i>my-topic</i>) from two input clusters:
+<pre>
+ &gt; bin/kafka-run-class.sh kafka.tools.MirrorMaker
+       --consumer.config consumer-1.properties --consumer.config 
consumer-2.properties 
+       --producer.config producer.properties --whitelist my-topic
+</pre>
+Note that we specify the list of topics with the <code>--whitelist</code> 
option. This option allows any regular expression using <a 
href="http://docs.oracle.com/javase/7/docs/api/java/util/regex/Pattern.html";>Java-style
 regular expressions</a>. So you could mirror two topics named <i>A</i> and 
<i>B</i> using <code>--whitelist 'A|B'</code>. Or you could mirror <i>all</i> 
topics using <code>--whitelist '*'</code>. Make sure to quote any regular 
expression to ensure the shell doesn't try to expand it as a file path. For 
convenience we allow the use of ',' instead of '|' to specify a list of topics.
+<p>
+Sometime it is easier to say what it is that you <i>don't</i> want. Instead of 
using <code>--whitelist</code> to say what you want to mirror you can use 
<code>--blacklist</code> to say what to exclude. This also takes a regular 
expression argument.
+<p>
+Combining mirroring with the configuration 
<code>auto.create.topics.enable=true</code> makes it possible to have a replica 
cluster that will automatically create and replicate all data in a source 
cluster even as new topics are added.
+
+<h4><a id="basic_ops_consumer_lag">Checking consumer position</a></h4>
+Sometimes it's useful to see the position of your consumers. We have a tool 
that will show the position of all consumers in a consumer group as well as how 
far behind the end of the log they are. To run this tool on a consumer group 
named <i>my-group</i> consuming a topic named <i>my-topic</i> would look like 
this:
+<pre>
+ &gt; bin/kafka-run-class.sh kafka.tools.ConsumerOffsetChecker --zkconnect 
localhost:2181 --group test
+Group           Topic                          Pid Offset          logSize     
    Lag             Owner
+my-group        my-topic                       0   0               0           
    0               test_jkreps-mn-1394154511599-60744496-0
+my-group        my-topic                       1   0               0           
    0               test_jkreps-mn-1394154521217-1a0be913-0
+</pre>
+
+<h4><a id="basic_ops_cluster_expansion">Expanding your cluster</a></h4>
+
+Adding servers to a Kafka cluster is easy, just assign them a unique broker id 
and start up Kafka on your new servers. However these new servers will not 
automatically be assigned any data partitions, so unless partitions are moved 
to them they won't be doing any work until new topics are created. So usually 
when you add machines to your cluster you will want to migrate some existing 
data to these machines.
+<p>
+The process of migrating data is manually initiated but fully automated. Under 
the covers what happens is that Kafka will add the new server as a follower of 
the partition it is migrating and allow it to fully replicate the existing data 
in that partition. When the new server has fully replicated the contents of 
this partition and joined the in-sync replica one of the existing replicas will 
delete their partition's data.
+<p>
+The partition reassignment tool can be used to move partitions across brokers. 
An ideal partition distribution would ensure even data load and partition sizes 
across all brokers. In 0.8.1, the partition reassignment tool does not have the 
capability to automatically study the data distribution in a Kafka cluster and 
move partitions around to attain an even load distribution. As such, the admin 
has to figure out which topics or partitions should be moved around. 
+<p>
+The partition reassignment tool can run in 3 mutually exclusive modes -
+<ul>
+<li>--generate: In this mode, given a list of topics and a list of brokers, 
the tool generates a candidate reassignment to move all partitions of the 
specified topics to the new brokers. This option merely provides a convenient 
way to generate a partition reassignment plan given a list of topics and target 
brokers.</li>
+<li>--execute: In this mode, the tool kicks off the reassignment of partitions 
based on the user provided reassignment plan. (using the 
--reassignment-json-file option). This can either be a custom reassignment plan 
hand crafted by the admin or provided by using the --generate option</li>
+<li>--verify: In this mode, the tool verifies the status of the reassignment 
for all partitions listed during the last --execute. The status can be either 
of successfully completed, failed or in progress</li>
+</ul>
+<h5>Automatically migrating data to new machines</h5>
+The partition reassignment tool can be used to move some topics off of the 
current set of brokers to the newly added brokers. This is typically useful 
while expanding an existing cluster since it is easier to move entire topics to 
the new set of brokers, than moving one partition at a time. When used to do 
this, the user should provide a list of topics that should be moved to the new 
set of brokers and a target list of new brokers. The tool then evenly 
distributes all partitions for the given list of topics across the new set of 
brokers. During this move, the replication factor of the topic is kept 
constant. Effectively the replicas for all partitions for the input list of 
topics are moved from the old set of brokers to the newly added brokers. 
+<p>
+For instance, the following example will move all partitions for topics 
foo1,foo2 to the new set of brokers 5,6. At the end of this move, all 
partitions for topics foo1 and foo2 will <i>only</i> exist on brokers 5,6
+<p>
+Since, the tool accepts the input list of topics as a json file, you first 
need to identify the topics you want to move and create the json file as 
follows-
+<pre>
+> cat topics-to-move.json
+{"topics": [{"topic": "foo1"},
+            {"topic": "foo2"}],
+ "version":1
+}
+</pre>
+Once the json file is ready, use the partition reassignment tool to generate a 
candidate assignment-
+<pre>
+> bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--topics-to-move-json-file topics-to-move.json --broker-list "5,6" --generate 
+Current partition replica assignment
+
+{"version":1,
+ "partitions":[{"topic":"foo1","partition":2,"replicas":[1,2]},
+               {"topic":"foo1","partition":0,"replicas":[3,4]},
+               {"topic":"foo2","partition":2,"replicas":[1,2]},
+               {"topic":"foo2","partition":0,"replicas":[3,4]},
+               {"topic":"foo1","partition":1,"replicas":[2,3]},
+               {"topic":"foo2","partition":1,"replicas":[2,3]}]
+}
+
+Proposed partition reassignment configuration
+
+{"version":1,
+ "partitions":[{"topic":"foo1","partition":2,"replicas":[5,6]},
+               {"topic":"foo1","partition":0,"replicas":[5,6]},
+               {"topic":"foo2","partition":2,"replicas":[5,6]},
+               {"topic":"foo2","partition":0,"replicas":[5,6]},
+               {"topic":"foo1","partition":1,"replicas":[5,6]},
+               {"topic":"foo2","partition":1,"replicas":[5,6]}]
+}
+</pre>
+<p>
+The tool generates a candidate assignment that will move all partitions from 
topics foo1,foo2 to brokers 5,6. Note, however, that at this point, the 
partition movement has not started, it merely tells you the current assignment 
and the proposed new assignment. The current assignment should be saved in case 
you want to rollback to it. The new assignment should be saved in a json file 
(e.g. expand-cluster-reassignment.json) to be input to the tool with the 
--execute option as follows-
+<pre>
+> bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file expand-cluster-reassignment.json --execute
+Current partition replica assignment
+
+{"version":1,
+ "partitions":[{"topic":"foo1","partition":2,"replicas":[1,2]},
+               {"topic":"foo1","partition":0,"replicas":[3,4]},
+               {"topic":"foo2","partition":2,"replicas":[1,2]},
+               {"topic":"foo2","partition":0,"replicas":[3,4]},
+               {"topic":"foo1","partition":1,"replicas":[2,3]},
+               {"topic":"foo2","partition":1,"replicas":[2,3]}]
+}
+
+Save this to use as the --reassignment-json-file option during rollback
+Successfully started reassignment of partitions
+{"version":1,
+ "partitions":[{"topic":"foo1","partition":2,"replicas":[5,6]},
+               {"topic":"foo1","partition":0,"replicas":[5,6]},
+               {"topic":"foo2","partition":2,"replicas":[5,6]},
+               {"topic":"foo2","partition":0,"replicas":[5,6]},
+               {"topic":"foo1","partition":1,"replicas":[5,6]},
+               {"topic":"foo2","partition":1,"replicas":[5,6]}]
+}
+</pre>
+<p>
+Finally, the --verify option can be used with the tool to check the status of 
the partition reassignment. Note that the same expand-cluster-reassignment.json 
(used with the --execute option) should be used with the --verify option
+<pre>
+> bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file expand-cluster-reassignment.json --verify
+Status of partition reassignment:
+Reassignment of partition [foo1,0] completed successfully
+Reassignment of partition [foo1,1] is in progress
+Reassignment of partition [foo1,2] is in progress
+Reassignment of partition [foo2,0] completed successfully
+Reassignment of partition [foo2,1] completed successfully 
+Reassignment of partition [foo2,2] completed successfully 
+</pre>
+
+<h5>Custom partition assignment and migration</h5>
+The partition reassignment tool can also be used to selectively move replicas 
of a partition to a specific set of brokers. When used in this manner, it is 
assumed that the user knows the reassignment plan and does not require the tool 
to generate a candidate reassignment, effectively skipping the --generate step 
and moving straight to the --execute step
+<p>
+For instance, the following example moves partition 0 of topic foo1 to brokers 
5,6 and partition 1 of topic foo2 to brokers 2,3
+<p>
+The first step is to hand craft the custom reassignment plan in a json file-
+<pre>
+> cat custom-reassignment.json
+{"version":1,"partitions":[{"topic":"foo1","partition":0,"replicas":[5,6]},{"topic":"foo2","partition":1,"replicas":[2,3]}]}
+</pre>
+Then, use the json file with the --execute option to start the reassignment 
process-
+<pre>
+> bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file custom-reassignment.json --execute
+Current partition replica assignment
+
+{"version":1,
+ "partitions":[{"topic":"foo1","partition":0,"replicas":[1,2]},
+               {"topic":"foo2","partition":1,"replicas":[3,4]}]
+}
+
+Save this to use as the --reassignment-json-file option during rollback
+Successfully started reassignment of partitions
+{"version":1,
+ "partitions":[{"topic":"foo1","partition":0,"replicas":[5,6]},
+               {"topic":"foo2","partition":1,"replicas":[2,3]}]
+}
+</pre>
+<p>
+The --verify option can be used with the tool to check the status of the 
partition reassignment. Note that the same expand-cluster-reassignment.json 
(used with the --execute option) should be used with the --verify option
+<pre>
+bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file custom-reassignment.json --verify
+Status of partition reassignment:
+Reassignment of partition [foo1,0] completed successfully
+Reassignment of partition [foo2,1] completed successfully 
+</pre>
+
+<h4><a id="basic_ops_decommissioning_brokers">Decommissioning brokers</a></h4>
+The partition reassignment tool does not have the ability to automatically 
generate a reassignment plan for decommissioning brokers yet. As such, the 
admin has to come up with a reassignment plan to move the replica for all 
partitions hosted on the broker to be decommissioned, to the rest of the 
brokers. This can be relatively tedious as the reassignment needs to ensure 
that all the replicas are not moved from the decommissioned broker to only one 
other broker. To make this process effortless, we plan to add tooling support 
for decommissioning brokers in 0.8.2.
+
+<h4><a id="basic_ops_increase_replication_factor">Increasing replication 
factor</a></h4>
+Increasing the replication factor of an existing partition is easy. Just 
specify the extra replicas in the custom reassignment json file and use it with 
the --execute option to increase the replication factor of the specified 
partitions. 
+<p>
+For instance, the following example increases the replication factor of 
partition 0 of topic foo from 1 to 3. Before increasing the replication factor, 
the partition's only replica existed on broker 5. As part of increasing the 
replication factor, we will add more replicas on brokers 6 and 7.
+<p>
+The first step is to hand craft the custom reassignment plan in a json file-
+<pre>
+> cat increase-replication-factor.json
+{"version":1,
+ "partitions":[{"topic":"foo","partition":0,"replicas":[5,6,7]}]}
+</pre>
+Then, use the json file with the --execute option to start the reassignment 
process-
+<pre>
+> bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file increase-replication-factor.json --execute
+Current partition replica assignment
+
+{"version":1,
+ "partitions":[{"topic":"foo","partition":0,"replicas":[5]}]}
+
+Save this to use as the --reassignment-json-file option during rollback
+Successfully started reassignment of partitions
+{"version":1,
+ "partitions":[{"topic":"foo","partition":0,"replicas":[5,6,7]}]}
+</pre>
+<p>
+The --verify option can be used with the tool to check the status of the 
partition reassignment. Note that the same increase-replication-factor.json 
(used with the --execute option) should be used with the --verify option
+<pre>
+bin/kafka-reassign-partitions.sh --zookeeper localhost:2181 
--reassignment-json-file increase-replication-factor.json --verify
+Status of partition reassignment:
+Reassignment of partition [foo,0] completed successfully
+</pre>
+You can also verify the increase in replication factor with the kafka-topics 
tool-
+<pre>
+> bin/kafka-topics.sh --zookeeper localhost:2181 --topic foo --describe
+Topic:foo      PartitionCount:1        ReplicationFactor:3     Configs:
+       Topic: foo      Partition: 0    Leader: 5       Replicas: 5,6,7 Isr: 
5,6,7
+</pre>
+
+<h3><a id="datacenters">6.2 Datacenters</a></h3>
+
+Some deployments will need to manage a data pipeline that spans multiple 
datacenters. Our recommended approach to this is to deploy a local Kafka 
cluster in each datacenter with application instances in each datacenter 
interacting only with their local cluster and mirroring between clusters (see 
the documentation on the <a href="#basic_ops_mirror_maker">mirror maker 
tool</a> for how to do this).
+<p>
+This deployment pattern allows datacenters to act as independent entities and 
allows us to manage and tune inter-datacenter replication centrally. This 
allows each facility to stand alone and operate even if the inter-datacenter 
links are unavailable: when this occurs the mirroring falls behind until the 
link is restored at which time it catches up.
+<p>
+For applications that need a global view of all data you can use mirroring to 
provide clusters which have aggregate data mirrored from the local clusters in 
<i>all</i> datacenters. These aggregate clusters are used for reads by 
applications that require the full data set.
+<p>
+This is not the only possible deployment pattern. It is possible to read from 
or write to a remote Kafka cluster over the WAN, though obviously this will add 
whatever latency is required to get the cluster.
+<p>
+Kafka naturally batches data in both the producer and consumer so it can 
achieve high-throughput even over a high-latency connection. To allow this 
though it may be necessary to increase the TCP socket buffer sizes for the 
producer, consumer, and broker using the <code>socket.send.buffer.bytes</code> 
and <code>socket.receive.buffer.bytes</code> configurations. The appropriate 
way to set this is documented <a 
href="http://en.wikipedia.org/wiki/Bandwidth-delay_product";>here</a>.    
+<p>
+It is generally <i>not</i> advisable to run a <i>single</i> Kafka cluster that 
spans multiple datacenters over a high-latency link. This will incur very high 
replication latency both for Kafka writes and ZooKeeper writes, and neither 
Kafka nor ZooKeeper will remain available in all locations if the network 
between locations is unavailable.
+
+<h3><a id="config">6.3 Kafka Configuration</a></h3>
+
+<h4><a id="clientconfig">Important Client Configurations</a></h4>
+The most important producer configurations control
+<ul>
+    <li>compression</li>
+    <li>sync vs async production</li>
+    <li>batch size (for async producers)</li>
+</ul>
+The most important consumer configuration is the fetch size.
+<p>
+All configurations are documented in the <a 
href="#configuration">configuration</a> section.
+<p>
+<h4><a id="prodconfig">A Production Server Config</a></h4>
+Here is our server production server configuration:
+<pre>
+# Replication configurations
+num.replica.fetchers=4
+replica.fetch.max.bytes=1048576
+replica.fetch.wait.max.ms=500
+replica.high.watermark.checkpoint.interval.ms=5000
+replica.socket.timeout.ms=30000
+replica.socket.receive.buffer.bytes=65536
+replica.lag.time.max.ms=10000
+
+controller.socket.timeout.ms=30000
+controller.message.queue.size=10
+
+# Log configuration
+num.partitions=8
+message.max.bytes=1000000
+auto.create.topics.enable=true
+log.index.interval.bytes=4096
+log.index.size.max.bytes=10485760
+log.retention.hours=168
+log.flush.interval.ms=10000
+log.flush.interval.messages=20000
+log.flush.scheduler.interval.ms=2000
+log.roll.hours=168
+log.retention.check.interval.ms=300000
+log.segment.bytes=1073741824
+
+# ZK configuration
+zookeeper.connection.timeout.ms=6000
+zookeeper.sync.time.ms=2000
+
+# Socket server configuration
+num.io.threads=8
+num.network.threads=8
+socket.request.max.bytes=104857600
+socket.receive.buffer.bytes=1048576
+socket.send.buffer.bytes=1048576
+queued.max.requests=16
+fetch.purgatory.purge.interval.requests=100
+producer.purgatory.purge.interval.requests=100
+</pre>
+
+Our client configuration varies a fair amount between different use cases.
+
+<h3><a id="java">Java Version</a></h3>
+We're currently running JDK 1.7 u51, and we've switched over to the G1 
collector. If you do this (and we highly recommend it), make sure you're on 
u51. We tried out u21 in testing, but we had a number of problems with the GC 
implementation in that version.
+
+Our tuning looks like this:
+<pre>
+-Xms4g -Xmx4g -XX:PermSize=48m -XX:MaxPermSize=48m -XX:+UseG1GC
+-XX:MaxGCPauseMillis=20 -XX:InitiatingHeapOccupancyPercent=35
+</pre>
+
+For reference, here are the stats on one of LinkedIn's busiest clusters (at 
peak):
+        - 15 brokers
+        - 15.5k partitions (replication factor 2)
+        - 400k messages/sec in
+        - 70 MB/sec inbound, 400 MB/sec+ outbound
+
+The tuning looks fairly aggressive, but all of the brokers in that cluster 
have a 90% GC pause time of about 21ms, and they're doing less than 1 young GC 
per second.
+       
+<h3><a id="hwandos">6.4 Hardware and OS</a></h3>
+We are using dual quad-core Intel Xeon machines with 24GB of memory.
+<p>
+You need sufficient memory to buffer active readers and writers. You can do a 
back-of-the-envelope estimate of memory needs by assuming you want to be able 
to buffer for 30 seconds and compute your memory need as write_throughput*30.
+<p>
+The disk throughput is important. We have 8x7200 rpm SATA drives. In general 
disk throughput is the performance bottleneck, and more disks is more better. 
Depending on how you configure flush behavior you may or may not benefit from 
more expensive disks (if you force flush often then higher RPM SAS drives may 
be better).
+
+<h4><a id="os">OS</a></h4>
+Kafka should run well on any unix system and has been tested on Linux and 
Solaris.
+<p>
+We have seen a few issues running on Windows and Windows is not currently a 
well supported platform though we would be happy to change that.
+<p>
+You likely don't need to do much OS-level tuning though there are a few things 
that will help performance. 
+<p>
+Two configurations that may be important:
+<ul>
+    <li>We upped the number of file descriptors since we have lots of topics 
and lots of connections.
+    <li>We upped the max socket buffer size to enable high-performance data 
transfer between data centers <a 
href="http://www.psc.edu/index.php/networking/641-tcp-tune";>described here</a>.
+</ul>
+
+<h4><a id="diskandfs">Disks and Filesystem</a></h4>
+We recommend using multiple drives to get good throughput and not sharing the 
same drives used for Kafka data with application logs or other OS filesystem 
activity to ensure good latency. As of 0.8 you can either RAID these drives 
together into a single volume or format and mount each drive as its own 
directory. Since Kafka has replication the redundancy provided by RAID can also 
be provided at the application level. This choice has several tradeoffs.
+<p>
+If you configure multiple data directories partitions will be assigned 
round-robin to data directories. Each partition will be entirely in one of the 
data directories. If data is not well balanced among partitions this can lead 
to load imbalance between disks.
+<p>
+RAID can potentially do better at balancing load between disks (although it 
doesn't always seem to) because it balances load at a lower level. The primary 
downside of RAID is that it is usually a big performance hit for write 
throughput and reduces the available disk space.
+<p>
+Another potential benefit of RAID is the ability to tolerate disk failures. 
However our experience has been that rebuilding the RAID array is so I/O 
intensive that it effectively disables the server, so this does not provide 
much real availability improvement.
+
+<h4><a id="appvsosflush">Application vs. OS Flush Management</a></h4>
+Kafka always immediately writes all data to the filesystem and supports the 
ability to configure the flush policy that controls when data is forced out of 
the OS cache and onto disk using the and flush. This flush policy can be 
controlled to force data to disk after a period of time or after a certain 
number of messages has been written. There are several choices in this 
configuration.
+<p>
+Kafka must eventually call fsync to know that data was flushed. When 
recovering from a crash for any log segment not known to be fsync'd Kafka will 
check the integrity of each message by checking its CRC and also rebuild the 
accompanying offset index file as part of the recovery process executed on 
startup.
+<p>
+Note that durability in Kafka does not require syncing data to disk, as a 
failed node will always recover from its replicas.
+<p>
+We recommend using the default flush settings which disable application fsync 
entirely. This means relying on the background flush done by the OS and Kafka's 
own background flush. This provides the best of all worlds for most uses: no 
knobs to tune, great throughput and latency, and full recovery guarantees. We 
generally feel that the guarantees provided by replication are stronger than 
sync to local disk, however the paranoid still may prefer having both and 
application level fsync policies are still supported.
+<p>
+The drawback of using application level flush settings are that this is less 
efficient in it's disk usage pattern (it gives the OS less leeway to re-order 
writes) and it can introduce latency as fsync in most Linux filesystems blocks 
writes to the file whereas the background flushing does much more granular 
page-level locking.
+<p>
+In general you don't need to do any low-level tuning of the filesystem, but in 
the next few sections we will go over some of this in case it is useful.
+
+<h4><a id="linuxflush">Understanding Linux OS Flush Behavior</a></h4>
+
+In Linux, data written to the filesystem is maintained in <a 
href="http://en.wikipedia.org/wiki/Page_cache";>pagecache</a> until it must be 
written out to disk (due to an application-level fsync or the OS's own flush 
policy). The flushing of data is done by a set of background threads called 
pdflush (or in post 2.6.32 kernels "flusher threads").
+<p>
+Pdflush has a configurable policy that controls how much dirty data can be 
maintained in cache and for how long before it must be written back to disk. 
This policy is described <a 
href="http://www.westnet.com/~gsmith/content/linux-pdflush.htm";>here</a>. When 
Pdflush cannot keep up with the rate of data being written it will eventually 
cause the writing process to block incurring latency in the writes to slow down 
the accumulation of data.
+<p>
+You can see the current state of OS memory usage by doing
+<pre>
+  &gt; cat /proc/meminfo
+</pre>
+The meaning of these values are described in the link above.
+<p>
+Using pagecache has several advantages over an in-process cache for storing 
data that will be written out to disk:
+<ul>
+  <li>The I/O scheduler will batch together consecutive small writes into 
bigger physical writes which improves throughput.
+  <li>The I/O scheduler will attempt to re-sequence writes to minimize 
movement of the disk head which improves throughput.
+  <li>It automatically uses all the free memory on the machine
+</ul>
+
+<h4><a id="ext4">Ext4 Notes</a></h4>
+Ext4 may or may not be the best filesystem for Kafka. Filesystems like XFS 
supposedly handle locking during fsync better. We have only tried Ext4, though.
+<p>
+It is not necessary to tune these settings, however those wanting to optimize 
performance have a few knobs that will help:
+<ul>
+  <li>data=writeback: Ext4 defaults to data=ordered which puts a strong order 
on some writes. Kafka does not require this ordering as it does very paranoid 
data recovery on all unflushed log. This setting removes the ordering 
constraint and seems to significantly reduce latency.
+  <li>Disabling journaling: Journaling is a tradeoff: it makes reboots faster 
after server crashes but it introduces a great deal of additional locking which 
adds variance to write performance. Those who don't care about reboot time and 
want to reduce a major source of write latency spikes can turn off journaling 
entirely.
+  <li>commit=num_secs: This tunes the frequency with which ext4 commits to its 
metadata journal. Setting this to a lower value reduces the loss of unflushed 
data during a crash. Setting this to a higher value will improve throughput.
+  <li>nobh: This setting controls additional ordering guarantees when using 
data=writeback mode. This should be safe with Kafka as we do not depend on 
write ordering and improves throughput and latency.
+  <li>delalloc: Delayed allocation means that the filesystem avoid allocating 
any blocks until the physical write occurs. This allows ext4 to allocate a 
large extent instead of smaller pages and helps ensure the data is written 
sequentially. This feature is great for throughput. It does seem to involve 
some locking in the filesystem which adds a bit of latency variance.
+</ul>
+       
+<h3><a id="monitoring">6.6 Monitoring</a></h3>
+
+Kafka uses Yammer Metrics for metrics reporting in both the server and the 
client. This can be configured to report stats using pluggable stats reporters 
to hook up to your monitoring system.
+<p>
+The easiest way to see the available metrics to fire up jconsole and point it 
at a running kafka client or server; this will all browsing all metrics with 
JMX.
+<p>
+We pay particular we do graphing and alerting on the following metrics:
+<table class="data-table">
+<tbody><tr>
+      <th>Description</th>
+      <th>Mbean name</th>
+      <th>Normal value</th>
+    </tr>
+    <tr>
+      <td>Message in rate</td>
+      <td>kafka.server:type=BrokerTopicMetrics,name=MessagesInPerSec</td>
+      <td></td>
+    </tr>
+    <tr>
+      <td>Byte in rate</td>
+      <td>kafka.server:type=BrokerTopicMetrics,name=BytesInPerSec</td>
+      <td></td>
+    </tr>
+    <tr>
+      <td>Request rate</td>
+      
<td>kafka.network:type=RequestMetrics,name=RequestsPerSec,request={Produce|FetchConsumer|FetchFollower}</td>
+      <td></td>
+    </tr>
+    <tr>
+      <td>Byte out rate</td>
+      <td>kafka.server:type=BrokerTopicMetrics,name=BytesOutPerSec</td>
+      <td></td>
+    </tr>
+    <tr>
+      <td>Log flush rate and time</td>
+      <td>kafka.log:type=LogFlushStats,name=LogFlushRateAndTimeMs</td>
+      <td></td>
+    </tr>
+    <tr>
+      <td># of under replicated partitions (|ISR| &lt |all replicas|)</td>
+      <td>kafka.server:type=ReplicaManager,name=UnderReplicatedPartitions</td>
+      <td>0</td>
+    </tr>
+    <tr>
+      <td>Is controller active on broker</td>
+      <td>kafka.controller:type=KafkaController,name=ActiveControllerCount</td>
+      <td>only one broker in the cluster should have 1</td>
+    </tr>
+    <tr>
+      <td>Leader election rate</td>
+      
<td>kafka.controller:type=ControllerStats,name=LeaderElectionRateAndTimeMs</td>
+      <td>non-zero when there are broker failures</td>
+    </tr>
+    <tr>
+      <td>Unclean leader election rate</td>
+      
<td>kafka.controller:type=ControllerStats,name=UncleanLeaderElectionsPerSec</td>
+      <td>0</td>
+    </tr>
+    <tr>
+      <td>Partition counts</td>
+      <td>kafka.server:type=ReplicaManager,name=PartitionCount</td>
+      <td>mostly even across brokers</td>
+    </tr>
+    <tr>
+      <td>Leader replica counts</td>
+      <td>kafka.server:type=ReplicaManager,name=LeaderCount</td>
+      <td>mostly even across brokers</td>
+    </tr>
+    <tr>
+      <td>ISR shrink rate</td>
+      <td>kafka.server:type=ReplicaManager,name=IsrShrinksPerSec</td>
+      <td>If a broker goes down, ISR for some of the partitions will
+       shrink. When that broker is up again, ISR will be expanded
+       once the replicas are fully caught up. Other than that, the
+       expected value for both ISR shrink rate and expansion rate is 0. </td>
+    </tr>
+    <tr>
+      <td>ISR expansion rate</td>
+      <td>kafka.server:type=ReplicaManager,name=IsrExpandsPerSec</td>
+      <td>See above</td>
+    </tr>
+    <tr>
+      <td>Max lag in messages btw follower and leader replicas</td>
+      
<td>kafka.server:type=ReplicaFetcherManager,name=MaxLag,clientId=Replica</td>
+      <td>lag should be proportional to the maximum batch size of a produce 
request.</td>
+    </tr>
+    <tr>
+      <td>Lag in messages per follower replica</td>
+      
<td>kafka.server:type=FetcherLagMetrics,name=ConsumerLag,clientId=([-.\w]+),topic=([-.\w]+),partition=([0-9]+)</td>
+      <td>lag should be proportional to the maximum batch size of a produce 
request.</td>
+    </tr>
+    <tr>
+      <td>Requests waiting in the producer purgatory</td>
+      <td>kafka.server:type=ProducerRequestPurgatory,name=PurgatorySize</td>
+      <td>non-zero if ack=-1 is used</td>
+    </tr>
+    <tr>
+      <td>Requests waiting in the fetch purgatory</td>
+      <td>kafka.server:type=FetchRequestPurgatory,name=PurgatorySize</td>
+      <td>size depends on fetch.wait.max.ms in the consumer</td>
+    </tr>
+    <tr>
+      <td>Request total time</td>
+      
<td>kafka.network:type=RequestMetrics,name=TotalTimeMs,request={Produce|FetchConsumer|FetchFollower}</td>
+      <td>broken into queue, local, remote and response send time</td>
+    </tr>
+    <tr>
+      <td>Time the request waiting in the request queue</td>
+       
<td>kafka.network:type=RequestMetrics,name=QueueTimeMs,request={Produce|FetchConsumer|FetchFollower}</td>
+      <td></td>
+    </tr>
+    <tr>
+      <td>Time the request being processed at the leader</td>
+      
<td>kafka.network:type=RequestMetrics,name=LocalTimeMs,request={Produce|FetchConsumer|FetchFollower}</td>
+      <td></td>
+    </tr>
+    <tr>
+      <td>Time the request waits for the follower</td>
+      
<td>kafka.network:type=RequestMetrics,name=RemoteTimeMs,request={Produce|FetchConsumer|FetchFollower}</td>
+      <td>non-zero for produce requests when ack=-1</td>
+    </tr>
+    <tr>
+      <td>Time to send the response</td>
+      
<td>kafka.network:type=RequestMetrics,name=ResponseSendTimeMs,request={Produce|FetchConsumer|FetchFollower}</td>
+      <td></td>
+    </tr>
+    <tr>
+      <td>Number of messages the consumer lags behind the producer by</td>
+      
<td>kafka.consumer:type=ConsumerFetcherManager,name=MaxLag,clientId=([-.\w]+)</td>
+      <td></td>
+    </tr>
+    <tr>
+      <td>The average fraction of time the network processors are idle</td>
+      
<td>kafka.network:type=SocketServer,name=NetworkProcessorAvgIdlePercent</td>
+      <td>between 0 and 1, ideally &gt 0.3</td>
+    </tr>
+    <tr>
+      <td>The average fraction of time the request handler threads are 
idle</td>
+      
<td>kafka.server:type=KafkaRequestHandlerPool,name=RequestHandlerAvgIdlePercent</td>
+      <td>between 0 and 1, ideally &gt 0.3</td>
+    </tr>
+</tbody></table>
+
+<h4><a id="new_producer_monitoring">New producer monitoring</a></h4>
+
+The following metrics are available on new producer instances.
+
+<table class="data-table">
+<tbody><tr>
+      <th>Metric/Attribute name</th>
+      <th>Description</th>
+      <th>Mbean name</th>
+    </tr>
+      <tr>
+      <td>waiting-threads</td>
+      <td>The number of user threads blocked waiting for buffer memory to 
enqueue their records</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>buffer-total-bytes</td>
+      <td>The maximum amount of buffer memory the client can use (whether or 
not it is currently used).</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>buffer-available-bytes</td>
+      <td>The total amount of buffer memory that is not being used (either 
unallocated or in the free list).</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>bufferpool-wait-time</td>
+      <td>The fraction of time an appender waits for space allocation.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>batch-size-avg</td>
+      <td>The average number of bytes sent per partition per-request.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>batch-size-max</td>
+      <td>The max number of bytes sent per partition per-request.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>compression-rate-avg</td>
+      <td>The average compression rate of record batches.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>record-queue-time-avg</td>
+      <td>The average time in ms record batches spent in the record 
accumulator.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>record-queue-time-max</td>
+      <td>The maximum time in ms record batches spent in the record 
accumulator</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>request-latency-avg</td>
+      <td>The average request latency in ms</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>request-latency-max</td>
+      <td>The maximum request latency in ms</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>record-send-rate</td>
+      <td>The average number of records sent per second.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>records-per-request-avg</td>
+      <td>The average number of records per request.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>record-retry-rate</td>
+      <td>The average per-second number of retried record sends</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>record-error-rate</td>
+      <td>The average per-second number of record sends that resulted in 
errors</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>record-size-max</td>
+      <td>The maximum record size</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>record-size-avg</td>
+      <td>The average record size</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>requests-in-flight</td>
+      <td>The current number of in-flight requests awaiting a response.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>metadata-age</td>
+      <td>The age in seconds of the current producer metadata being used.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>connection-close-rate</td>
+      <td>Connections closed per second in the window.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>connection-creation-rate</td>
+      <td>New connections established per second in the window.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>network-io-rate</td>
+      <td>The average number of network operations (reads or writes) on all 
connections per second.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>outgoing-byte-rate</td>
+      <td>The average number of outgoing bytes sent per second to all 
servers.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>request-rate</td>
+      <td>The average number of requests sent per second.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>request-size-avg</td>
+      <td>The average size of all requests in the window.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>request-size-max</td>
+      <td>The maximum size of any request sent in the window.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>incoming-byte-rate</td>
+      <td>Bytes/second read off all sockets</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>response-rate</td>
+      <td>Responses received sent per second.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>select-rate</td>
+      <td>Number of times the I/O layer checked for new I/O to perform per 
second</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>io-wait-time-ns-avg</td>
+      <td>The average length of time the I/O thread spent waiting for a socket 
ready for reads or writes in nanoseconds.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>io-wait-ratio</td>
+      <td>The fraction of time the I/O thread spent waiting.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>io-time-ns-avg</td>
+      <td>The average length of time for I/O per select call in 
nanoseconds.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>io-ratio</td>
+      <td>The fraction of time the I/O thread spent doing I/O</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>connection-count</td>
+      <td>The current number of active connections.</td>
+      <td>kafka.producer:type=producer-metrics,client-id=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>outgoing-byte-rate</td>
+      <td>The average number of outgoing bytes sent per second for a node.</td>
+      
<td>kafka.producer:type=producer-node-metrics,client-id=([-.\w]+),node-id=([0-9]+)</td>
+    </tr>
+    <tr>
+      <td>request-rate</td>
+      <td>The average number of requests sent per second for a node.</td>
+      
<td>kafka.producer:type=producer-node-metrics,client-id=([-.\w]+),node-id=([0-9]+)</td>
+    </tr>
+    <tr>
+      <td>request-size-avg</td>
+      <td>The average size of all requests in the window for a node.</td>
+      
<td>kafka.producer:type=producer-node-metrics,client-id=([-.\w]+),node-id=([0-9]+)</td>
+    </tr>
+    <tr>
+      <td>request-size-max</td>
+      <td>The maximum size of any request sent in the window for a node.</td>
+      
<td>kafka.producer:type=producer-node-metrics,client-id=([-.\w]+),node-id=([0-9]+)</td>
+    </tr>
+    <tr>
+      <td>incoming-byte-rate</td>
+      <td>The average number of responses received per second for a node.</td>
+      
<td>kafka.producer:type=producer-node-metrics,client-id=([-.\w]+),node-id=([0-9]+)</td>
+    </tr>
+    <tr>
+      <td>request-latency-avg</td>
+      <td>The average request latency in ms for a node.</td>
+      
<td>kafka.producer:type=producer-node-metrics,client-id=([-.\w]+),node-id=([0-9]+)</td>
+    </tr>
+    <tr>
+      <td>request-latency-max</td>
+      <td>The maximum request latency in ms for a node.</td>
+      
<td>kafka.producer:type=producer-node-metrics,client-id=([-.\w]+),node-id=([0-9]+)</td>
+    </tr>
+    <tr>
+      <td>response-rate</td>
+      <td>Responses received sent per second for a node.</td>
+      
<td>kafka.producer:type=producer-node-metrics,client-id=([-.\w]+),node-id=([0-9]+)</td>
+    </tr>
+    <tr>
+      <td>record-send-rate</td>
+      <td>The average number of records sent per second for a topic.</td>
+      
<td>kafka.producer:type=producer-topic-metrics,client-id=([-.\w]+),topic=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>byte-rate</td>
+      <td>The average number of bytes sent per second for a topic.</td>
+      
<td>kafka.producer:type=producer-topic-metrics,client-id=([-.\w]+),topic=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>compression-rate</td>
+      <td>The average compression rate of record batches for a topic.</td>
+      
<td>kafka.producer:type=producer-topic-metrics,client-id=([-.\w]+),topic=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>record-retry-rate</td>
+      <td>The average per-second number of retried record sends for a 
topic</td>
+      
<td>kafka.producer:type=producer-topic-metrics,client-id=([-.\w]+),topic=([-.\w]+)</td>
+    </tr>
+    <tr>
+      <td>record-error-rate</td>
+      <td>The average per-second number of record sends that resulted in 
errors for a topic.</td>
+      
<td>kafka.producer:type=producer-topic-metrics,client-id=([-.\w]+),topic=([-.\w]+)</td>
+    </tr>
+</tbody></table>
+
+We recommend monitor GC time and other stats and various server stats such as 
CPU utilization, I/O service time, etc.
+
+On the client side, we recommend monitor the message/byte rate (global and per 
topic), request rate/size/time, and on the consumer side, max lag in messages 
among all partitions and min fetch request rate. For a consumer to keep up, max 
lag needs to be less than a threshold and min fetch rate needs to be larger 
than 0.
+
+<h4>Audit</h4>
+The final alerting we do is on the correctness of the data delivery. We audit 
that every message that is sent is consumed by all consumers and measure the 
lag for this to occur. For important topics we alert if a certain completeness 
is not achieved in a certain time period. The details of this are discussed in 
KAFKA-260.
+
+<h3><a id="zk">6.7 ZooKeeper</a></h3>
+
+<h4><a id="zkversion">Stable version</a></h4>
+At LinkedIn, we are running ZooKeeper 3.3.*. Version 3.3.3 has known serious 
issues regarding ephemeral node deletion and session expirations. After running 
into those issues in production, we upgraded to 3.3.4 and have been running 
that smoothly for over a year now.
+
+<h4><a id="zkops">Operationalizing ZooKeeper</a></h4>
+Operationally, we do the following for a healthy ZooKeeper installation:
+<ul>
+  <li>Redundancy in the physical/hardware/network layout: try not to put them 
all in the same rack, decent (but don't go nuts) hardware, try to keep 
redundant power and network paths, etc.</li>
+  <li>I/O segregation: if you do a lot of write type traffic you'll almost 
definitely want the transaction logs on a different disk group than application 
logs and snapshots (the write to the ZooKeeper service has a synchronous write 
to disk, which can be slow).</li>
+  <li>Application segregation: Unless you really understand the application 
patterns of other apps that you want to install on the same box, it can be a 
good idea to run ZooKeeper in isolation (though this can be a balancing act 
with the capabilities of the hardware).</li>
+  <li>Use care with virtualization: It can work, depending on your cluster 
layout and read/write patterns and SLAs, but the tiny overheads introduced by 
the virtualization layer can add up and throw off ZooKeeper, as it can be very 
time sensitive</li>
+  <li>ZooKeeper configuration and monitoring: It's java, make sure you give it 
'enough' heap space (We usually run them with 3-5G, but that's mostly due to 
the data set size we have here). Unfortunately we don't have a good formula for 
it. As far as monitoring, both JMX and the 4 letter words (4lw) commands are 
very useful, they do overlap in some cases (and in those cases we prefer the 4 
letter commands, they seem more predictable, or at the very least, they work 
better with the LI monitoring infrastructure)</li>
+  <li>Don't overbuild the cluster: large clusters, especially in a write heavy 
usage pattern, means a lot of intracluster communication (quorums on the writes 
and subsequent cluster member updates), but don't underbuild it (and risk 
swamping the cluster).</li>
+  <li>Try to run on a 3-5 node cluster: ZooKeeper writes use quorums and 
inherently that means having an odd number of machines in a cluster. Remember 
that a 5 node cluster will cause writes to slow down compared to a 3 node 
cluster, but will allow more fault tolerance.</li>
+</ul>
+Overall, we try to keep the ZooKeeper system as small as will handle the load 
(plus standard growth capacity planning) and as simple as possible. We try not 
to do anything fancy with the configuration or application layout as compared 
to the official release as well as keep it as self contained as possible. For 
these reasons, we tend to skip the OS packaged versions, since it has a 
tendency to try to put things in the OS standard hierarchy, which can be 
'messy', for want of a better way to word it.

http://git-wip-us.apache.org/repos/asf/kafka-site/blob/ba6c994c/090/quickstart.html
----------------------------------------------------------------------
diff --git a/090/quickstart.html b/090/quickstart.html
new file mode 100644
index 0000000..1e5c3ab
--- /dev/null
+++ b/090/quickstart.html
@@ -0,0 +1,172 @@
+<h3><a id="quickstart">1.3 Quick Start</a></h3>
+
+This tutorial assumes you are starting fresh and have no existing Kafka or 
ZooKeeper data.
+
+<h4> Step 1: Download the code </h4>
+
+<a 
href="https://www.apache.org/dyn/closer.cgi?path=/kafka/0.8.2.0/kafka_2.10-0.8.2.0.tgz";
 title="Kafka downloads">Download</a> the 0.8.2.0 release and un-tar it.
+
+<pre>
+&gt; <b>tar -xzf kafka_2.10-0.8.2.0.tgz</b>
+&gt; <b>cd kafka_2.10-0.8.2.0</b>
+</pre>
+
+<h4>Step 2: Start the server</h4>
+
+<p>
+Kafka uses ZooKeeper so you need to first start a ZooKeeper server if you 
don't already have one. You can use the convenience script packaged with kafka 
to get a quick-and-dirty single-node ZooKeeper instance.
+
+<pre>
+&gt; <b>bin/zookeeper-server-start.sh config/zookeeper.properties</b>
+[2013-04-22 15:01:37,495] INFO Reading configuration from: 
config/zookeeper.properties 
(org.apache.zookeeper.server.quorum.QuorumPeerConfig)
+...
+</pre>
+
+Now start the Kafka server:
+<pre>
+&gt; <b>bin/kafka-server-start.sh config/server.properties</b>
+[2013-04-22 15:01:47,028] INFO Verifying properties 
(kafka.utils.VerifiableProperties)
+[2013-04-22 15:01:47,051] INFO Property socket.send.buffer.bytes is overridden 
to 1048576 (kafka.utils.VerifiableProperties)
+...
+</pre>
+
+<h4>Step 3: Create a topic</h4>
+
+Let's create a topic named "test" with a single partition and only one replica:
+<pre>
+&gt; <b>bin/kafka-topics.sh --create --zookeeper localhost:2181 
--replication-factor 1 --partitions 1 --topic test</b>
+</pre>
+
+We can now see that topic if we run the list topic command:
+<pre>
+&gt; <b>bin/kafka-topics.sh --list --zookeeper localhost:2181</b>
+test
+</pre>
+Alternatively, instead of manually creating topics you can also configure your 
brokers to auto-create topics when a non-existent topic is published to.
+
+<h4>Step 4: Send some messages</h4>
+
+Kafka comes with a command line client that will take input from a file or 
from standard input and send it out as messages to the Kafka cluster. By 
default each line will be sent as a separate message.
+<p>
+Run the producer and then type a few messages into the console to send to the 
server.
+
+<pre>
+&gt; <b>bin/kafka-console-producer.sh --broker-list localhost:9092 --topic 
test</b> 
+<b>This is a message</b>
+<b>This is another message</b>
+</pre>
+
+<h4>Step 5: Start a consumer</h4>
+
+Kafka also has a command line consumer that will dump out messages to standard 
output.
+
+<pre>
+&gt; <b>bin/kafka-console-consumer.sh --zookeeper localhost:2181 --topic test 
--from-beginning</b>
+This is a message
+This is another message
+</pre>
+<p>
+If you have each of the above commands running in a different terminal then 
you should now be able to type messages into the producer terminal and see them 
appear in the consumer terminal.
+</p>
+<p>
+All of the command line tools have additional options; running the command 
with no arguments will display usage information documenting them in more 
detail.   
+</p>
+
+<h4>Step 6: Setting up a multi-broker cluster</h4>
+
+So far we have been running against a single broker, but that's no fun. For 
Kafka, a single broker is just a cluster of size one, so nothing much changes 
other than starting a few more broker instances. But just to get feel for it, 
let's expand our cluster to three nodes (still all on our local machine).
+<p>
+First we make a config file for each of the brokers:
+<pre>
+&gt; <b>cp config/server.properties config/server-1.properties</b> 
+&gt; <b>cp config/server.properties config/server-2.properties</b>
+</pre>
+
+Now edit these new files and set the following properties:
+<pre>
+ 
+config/server-1.properties:
+    broker.id=1
+    port=9093
+    log.dir=/tmp/kafka-logs-1
+ 
+config/server-2.properties:
+    broker.id=2
+    port=9094
+    log.dir=/tmp/kafka-logs-2
+</pre>
+The <code>broker.id</code> property is the unique and permanent name of each 
node in the cluster. We have to override the port and log directory only 
because we are running these all on the same machine and we want to keep the 
brokers from all trying to register on the same port or overwrite each others 
data.
+<p>
+We already have Zookeeper and our single node started, so we just need to 
start the two new nodes:
+<pre>
+&gt; <b>bin/kafka-server-start.sh config/server-1.properties &amp;</b>
+...
+&gt; <b>bin/kafka-server-start.sh config/server-2.properties &amp;</b>
+...
+</pre>
+
+Now create a new topic with a replication factor of three:
+<pre>
+&gt; <b>bin/kafka-topics.sh --create --zookeeper localhost:2181 
--replication-factor 3 --partitions 1 --topic my-replicated-topic</b>
+</pre>
+
+Okay but now that we have a cluster how can we know which broker is doing 
what? To see that run the "describe topics" command:
+<pre>
+&gt; <b>bin/kafka-topics.sh --describe --zookeeper localhost:2181 --topic 
my-replicated-topic</b>
+Topic:my-replicated-topic      PartitionCount:1        ReplicationFactor:3     
Configs:
+       Topic: my-replicated-topic      Partition: 0    Leader: 1       
Replicas: 1,2,0 Isr: 1,2,0
+</pre>
+Here is an explanation of output. The first line gives a summary of all the 
partitions, each additional line gives information about one partition. Since 
we have only one partition for this topic there is only one line.
+<ul>
+  <li>"leader" is the node responsible for all reads and writes for the given 
partition. Each node will be the leader for a randomly selected portion of the 
partitions.
+  <li>"replicas" is the list of nodes that replicate the log for this 
partition regardless of whether they are the leader or even if they are 
currently alive.
+  <li>"isr" is the set of "in-sync" replicas. This is the subset of the 
replicas list that is currently alive and caught-up to the leader.
+</ul> 
+Note that in my example node 1 is the leader for the only partition of the 
topic.
+<p>
+We can run the same command on the original topic we created to see where it 
is:
+<pre>
+&gt; <b>bin/kafka-topics.sh --describe --zookeeper localhost:2181 --topic 
test</b>
+Topic:test     PartitionCount:1        ReplicationFactor:1     Configs:
+       Topic: test     Partition: 0    Leader: 0       Replicas: 0     Isr: 0
+</pre>
+So there is no surprise there&mdash;the original topic has no replicas and is 
on server 0, the only server in our cluster when we created it.
+<p>
+Let's publish a few messages to our new topic:
+<pre>
+&gt; <b>bin/kafka-console-producer.sh --broker-list localhost:9092 --topic 
my-replicated-topic</b>
+...
+<b>my test message 1</b>
+<b>my test message 2</b>
+<b>^C</b> 
+</pre>
+Now let's consume these messages:
+<pre>
+&gt; <b>bin/kafka-console-consumer.sh --zookeeper localhost:2181 
--from-beginning --topic my-replicated-topic</b>
+...
+my test message 1
+my test message 2
+<b>^C</b>
+</pre>
+
+Now let's test out fault-tolerance. Broker 1 was acting as the leader so let's 
kill it:
+<pre>
+&gt; <b>ps | grep server-1.properties</b>
+<i>7564</i> ttys002    0:15.91 
/System/Library/Frameworks/JavaVM.framework/Versions/1.6/Home/bin/java...
+&gt; <b>kill -9 7564</b>
+</pre>
+
+Leadership has switched to one of the slaves and node 1 is no longer in the 
in-sync replica set:
+<pre>
+&gt; <b>bin/kafka-topics.sh --describe --zookeeper localhost:2181 --topic 
my-replicated-topic</b>
+Topic:my-replicated-topic      PartitionCount:1        ReplicationFactor:3     
Configs:
+       Topic: my-replicated-topic      Partition: 0    Leader: 2       
Replicas: 1,2,0 Isr: 2,0
+</pre>
+But the messages are still be available for consumption even though the leader 
that took the writes originally is down:
+<pre>
+&gt; <b>bin/kafka-console-consumer.sh --zookeeper localhost:2181 
--from-beginning --topic my-replicated-topic</b>
+...
+my test message 1
+my test message 2
+<b>^C</b>
+</pre>

http://git-wip-us.apache.org/repos/asf/kafka-site/blob/ba6c994c/090/upgrade.html
----------------------------------------------------------------------
diff --git a/090/upgrade.html b/090/upgrade.html
new file mode 100644
index 0000000..2be08dd
--- /dev/null
+++ b/090/upgrade.html
@@ -0,0 +1,27 @@
+<h3><a id="upgrade">1.5 Upgrading From Previous Versions</a></h3>
+
+<h4>Upgrading from 0.8.0, 0.8.1.X or 0.8.2.X to 0.8.3.0</h4>
+
+0.8.3.0 has an inter-broker protocol change from previous versions. For a 
rolling upgrade:
+<ol>
+       <li> Update server.properties file on all brokers and add the following 
property: inter.broker.protocol.version=0.8.2.X </li>
+       <li> Upgrade the brokers. This can be done a broker at a time by simply 
bringing it down, updating the code, and restarting it. </li>
+       <li> Once the entire cluster is upgraded, bump the protocol version by 
editing inter.broker.protocol.version and setting it to 0.8.3.0.</li>
+       <li> Restart the brokers one by one for the new protocol version to 
take effect </li>
+</ol>
+
+Note: If you are willing to accept downtime, you can simply take all the 
brokers down, update the code and start all of them. They will start with the 
new protocol by default.
+
+Note: Bumping the protocol version and restarting can be done any time after 
the brokers were upgraded. It does not have to be immediately after.
+
+<h4>Upgrading from 0.8.1 to 0.8.2.0</h4>
+
+0.8.2.0 is fully compatible with 0.8.1. The upgrade can be done one broker at 
a time by simply bringing it down, updating the code, and restarting it.
+
+<h4>Upgrading from 0.8.0 to 0.8.1</h4>
+
+0.8.1 is fully compatible with 0.8. The upgrade can be done one broker at a 
time by simply bringing it down, updating the code, and restarting it.
+
+<h4>Upgrading from 0.7</h4>
+
+0.8, the release in which added replication, was our first 
backwards-incompatible release: major changes were made to the API, ZooKeeper 
data structures, and protocol, and configuration. The upgrade from 0.7 to 0.8.x 
requires a <a 
href="https://cwiki.apache.org/confluence/display/KAFKA/Migrating+from+0.7+to+0.8";>special
 tool</a> for migration. This migration can be done without downtime.

http://git-wip-us.apache.org/repos/asf/kafka-site/blob/ba6c994c/090/uses.html
----------------------------------------------------------------------
diff --git a/090/uses.html b/090/uses.html
new file mode 100644
index 0000000..6b73ddd
--- /dev/null
+++ b/090/uses.html
@@ -0,0 +1,39 @@
+<h3><a id="uses">1.2 Use Cases</a></h3>
+
+Here is a description of a few of the popular use cases for Apache Kafka. For 
an overview of a number of these areas in action, see <a 
href="http://engineering.linkedin.com/distributed-systems/log-what-every-software-engineer-should-know-about-real-time-datas-unifying";>this
 blog post</a>.
+
+<h4>Messaging</h4>
+
+Kafka works well as a replacement for a more traditional message broker. 
Message brokers are used for a variety of reasons (to decouple processing from 
data producers, to buffer unprocessed messages, etc). In comparison to most 
messaging systems Kafka has better throughput, built-in partitioning, 
replication, and fault-tolerance which makes it a good solution for large scale 
message processing applications.
+<p>
+In our experience messaging uses are often comparatively low-throughput, but 
may require low end-to-end latency and often depend on the strong durability 
guarantees Kafka provides.
+<p>
+In this domain Kafka is comparable to traditional messaging systems such as <a 
href="http://activemq.apache.org";>ActiveMQ</a> or <a 
href="https://www.rabbitmq.com";>RabbitMQ</a>.
+
+<h4>Website Activity Tracking</h4>
+
+The original use case for Kafka was to be able to rebuild a user activity 
tracking pipeline as a set of real-time publish-subscribe feeds. This means 
site activity (page views, searches, or other actions users may take) is 
published to central topics with one topic per activity type. These feeds are 
available for subscription for a range of use cases including real-time 
processing, real-time monitoring, and loading into Hadoop or offline data 
warehousing systems for offline processing and reporting.
+<p>
+Activity tracking is often very high volume as many activity messages are 
generated for each user page view.
+
+<h4>Metrics</h4>
+
+Kafka is often used for operational monitoring data. This involves aggregating 
statistics from distributed applications to produce centralized feeds of 
operational data.
+
+<h4>Log Aggregation</h4>
+
+Many people use Kafka as a replacement for a log aggregation solution. Log 
aggregation typically collects physical log files off servers and puts them in 
a central place (a file server or HDFS perhaps) for processing. Kafka abstracts 
away the details of files and gives a cleaner abstraction of log or event data 
as a stream of messages. This allows for lower-latency processing and easier 
support for multiple data sources and distributed data consumption.
+
+In comparison to log-centric systems like Scribe or Flume, Kafka offers 
equally good performance, stronger durability guarantees due to replication, 
and much lower end-to-end latency.
+
+<h4>Stream Processing</h4>
+
+Many users end up doing stage-wise processing of data where data is consumed 
from topics of raw data and then aggregated, enriched, or otherwise transformed 
into new Kafka topics for further consumption. For example a processing flow 
for article recommendation might crawl article content from RSS feeds and 
publish it to an "articles" topic; further processing might help normalize or 
deduplicate this content to a topic of cleaned article content; a final stage 
might attempt to match this content to users. This creates a graph of real-time 
data flow out of the individual topics. <a 
href="https://storm.apache.org/";>Storm</a> and <a 
href="http://samza.apache.org/";>Samza</a> are popular frameworks for 
implementing these kinds of transformations.
+
+<h4>Event Sourcing</h4>
+
+<a href="http://martinfowler.com/eaaDev/EventSourcing.html";>Event sourcing</a> 
is a style of application design where state changes are logged as a 
time-ordered sequence of records. Kafka's support for very large stored log 
data makes it an excellent backend for an application built in this style.
+
+<h4>Commit Log</h4>
+
+Kafka can serve as a kind of external commit-log for a distributed system. The 
log helps replicate data between nodes and acts as a re-syncing mechanism for 
failed nodes to restore their data. The <a 
href="/documentation.html#compaction">log compaction</a> feature in Kafka helps 
support this usage. In this usage Kafka is similar to <a 
href="http://zookeeper.apache.org/bookkeeper/";>Apache BookKeeper</a> project.

http://git-wip-us.apache.org/repos/asf/kafka-site/blob/ba6c994c/KEYS
----------------------------------------------------------------------
diff --git a/KEYS b/KEYS
new file mode 100644
index 0000000..57c139e
--- /dev/null
+++ b/KEYS
@@ -0,0 +1,294 @@
+This file contains the PGP keys of various developers.
+
+Users: pgp < KEYS
+       gpg --import KEYS
+Developers:
+        pgp -kxa <your name> and append it to this file.
+        (pgpk -ll <your name> && pgpk -xa <your name>) >> this file.
+        (gpg --list-sigs <your name>
+             && gpg --armor --export <your name>) >> this file.
+
+pub   4096R/99369B56 2011-10-06
+uid                  Neha Narkhede (Key for signing code and releases) 
<[email protected]>
+sig 3        99369B56 2011-10-06  Neha Narkhede (Key for signing code and 
releases) <[email protected]>
+sub   4096R/A71D126A 2011-10-06
+sig          99369B56 2011-10-06  Neha Narkhede (Key for signing code and 
releases) <[email protected]>
+
+-----BEGIN PGP PUBLIC KEY BLOCK-----
+Version: GnuPG/MacGPG2 v2.0.17 (Darwin)
+Comment: GPGTools - http://gpgtools.org
+
+mQENBEt9wioBCADh0bdDopK7wdLLt6YIEA3KWdXmRhhmY2PDikKZq5EQlwkAmdZF
+/CTcheArdAXXuxfPN8kZp4MJE01mrgyZA9S8tsYG1GarPTpYUDXxZJJgswSKNAbU
+j4sL1sYm89pzsm57Mjt+4ek9F+GJeEBOiog3/4oaaVeuzFT2LhLbD2PY5CS/5MdZ
+t4KaqPAHoAQQGwzGSzaXxXbGKlQUDm0W33jjyWyON/eHHLUHbFwlb9f+du8DJLQ6
+WQjFR/xSzSxztJFUF1JUBm6l9ZMMMeAojTEBKXeh9Bo7iWcVU/nJ+VebAE8IpgU4
+9TnNf/o2iJnKKTyKLW5QDR83gIFsml/6sja3ABEBAAGJAR8EIAEKAAkFAkyaTuQC
+HQEACgkQaCVaMFMDgyhOlAgAixphPIQp1z/5fuuiXLBTPzLLRsR3UQ9jeAowUdrA
+utnn7uVHvAqcUy/WLGJyRMVQl3y8oNAQXmNSfeLzp2Bs/uRuOf4B2b2XBX3F6AqX
+1UI7ASzyG1cowHNZ+Oq3Edg6YwQCt+5QwrlLCcp4eo3J0/NwxrRqYM5TdFVnvN3L
+PmHMPlOhLfZJRf2/g9dcWpIcLEYaGgJMD4uUogaN2CT2GXjntFsdgRG7jYasTK6+
+TF3ML/rB/tbNcMo23IiQ8GKaGO04uQUlyo2b3ix4uUZGtIIIFtZMOcgSLTKSavn/
+hPMe6wPOkB/Onno0zDUBCrpACxjcW9fyTALBlrJTCrYpz7QjQ2hyaXMgRG91Z2xh
+cyA8Y2RvdWdsYXNAYXBhY2hlLm9yZz6JATgEEwECACIFAkt9wioCGwMGCwkIBwMC
+BhUIAgkKCwQWAgMBAh4BAheAAAoJEGglWjBTA4MoN1QH+gJLOaIUhIqECvwz77js
+9HfadKmKqwc0SGDpoZm2tMeUsp8+26fhqEdprNcfKV530+M4e4l/ka4Y96p7C+EW
+kAuMQ41F9YnvxX9dATc+RgLueVkaUxMjNLPAIwQkUyQoeR/tLcMpNI/lbWdwEJRK
+unufbxDGIXVjnN8y9dkmlZ5NGsAWa2ZoYDrpKHFwTKvNv+J24rYuFbuBktlKk1L7
+f0N/4dCeFlHjoaVykS0BuQrEA8/ZPzE4qPw7FFWUb3M+gZH7xulSQnc+Q4oCVdYw
+FKsoB5iz924R4g09Yk9l8wXIkKvhCFwkioapZLqFQyse1hcsEUmBJzLBjjXWkw9+
+pXe5AQ0ES33CKgEIALqimXtgSq6zB21b2z4mVIAgSHFRQOa3q5jhWYvHALnwSIIc
+AvPHnPfWUuRyaGjJZbVgLdB0o+ooxYT8SdZtPVFMpVvoAaLH/4t1DJvQTXYUW+vE
+Z6cWeC5rr+JXcwLa8xhSYYhv3y7Lf3AwdKq6bzF5+7/rwDu7K5LhzS+hVrhln4Uq
+yFNQdnWoXIsBIt0wxut3KuRyNICjYs6Jr0zkFR7azCJCDml6+NnGX/jCtK5HE0AZ
+f11VOgds1kH1bohI0FkzRXk5PbMgvnxRZNlSGxzTUceYystJq/iIuzvSH3ixMKqL
+Se0KAOj1FrcpGsP+RF17sooIMjntg2MZKr7STTsAEQEAAYkBHwQYAQIACQUCS33C
+KgIbDAAKCRBoJVowUwODKP9GB/9TsH6lBFz/ueae0AD6gwGGRySkc+zilFYYBAbR
+jEpPpoyj5KCVYcQ1Q5SPeLzy7/4hQJlf0F0gg8w3G4Axjnaej4oDuNIGBKGKTLsl
+SaBI2P44wlVdDOaUhgAVB/zzCqUidLYVwoOmT3OLz01lZgPXKihw+I5JtNvdpOa+
+xXHQ7dwezmUMChHZADF198/2pRIamqwMD0MyApaMFeb5posM0KYHj+k6x8uMQB8m
+V6IdsgqZpglXO7gfuvz3mQqye50e0xIhCLdX6JwIOnLGzUPJW9ds33rWkzk9MW+h
+JnjuiZeYbtlLzCzuvRVjnSblrUUCaOm1RFVWrw6yVVTGHErmmQINBE6OPIgBEAC5
+XC3CGy1peReJPUk2J5iCvmdSIOEaiOfULaz+1WiKgBMovUAKCbIuCB9ZMjfT0gce
+Agqj4UeCeRhPi2dVxVl/r1LCyGMJ1hZyEGDk7No1QkCemnCD4yiBD+BX+9V1Zuqa
+r1goV5tMXoE0kq0cyoih/c6t8qrP5Wf/BTS118TBFAHRql/yYle+g3YEk0uZ4yue
+dCVQIrjTWe46fSAacF8eGluGQVTbNj2aRacfu+UYPj8G241F8gTOTjZ3fJ80KTOI
+tjiV8Gy+I6HDIWd7XMwNCO8z2G2lto0CBoDuaUukXgB5/CNFQL1S8Zxl2idxEQSt
+ZzKTKmZZcom+GNMvL8xCE8sMgMSiZ8Q54fRevTQrDLPpIhl0g6V1v7vOfokWxf7K
+6t+KZXB7oZbc1YrW/kcjDUAMlS4gAQ4K7ngEK3qIEVquvXx4nLVM++zn0/0PVzES
+2gMnacP9mdsVsPbJx+Jwk6hZi1mYBS7DyZkoVJYgXqLdCtWebH+VqdHiEHBrWYxH
+eoyHWjViJKZYh3ojcKuZMicptEzA25hz5HaKEkKv1PEYeKEDG8tLVf09dU+HHdK7
+Sbca08hXFhDIiDlksFa67Zy7MriJRTun8rTCD1DFH3wyuubRu8bDo7UobBgSK3td
+cBRKM8i3e3JQZejhKlROhJ56xIT3ivO3Coe0hTUDXQARAQABtEtOZWhhIE5hcmto
+ZWRlIChLZXkgZm9yIHNpZ25pbmcgY29kZSBhbmQgcmVsZWFzZXMpIDxuZWhhbmFy
+a2hlZGVAYXBhY2hlLm9yZz6JAjgEEwECACIFAk6OPIgCGwMGCwkIBwMCBhUIAgkK
+CwQWAgMBAh4BAheAAAoJEGx0AWyZNptWXSsP/i75oSJ50HJBRSni0oT6SxDG0Ybs
+Gj3Ojwn2FppOOpVW7kjSRPTb/gSTHYlxNvBxI+nkyKx4Z2+IloDaH2LUsxdAkBor
+9a58ykW9BD5Yd/5nRvduHp71XV28b3/xnN5H6kbCUr5yWZPVZ4//o8L12PS3Jy0i
+c9cQQF7vzuqQOvPpncruBChikEnSrwFmZI+UMRBBduCck6PNyeWTjb8ipfJtvwfO
+vQHX5+AOSz6zwICkzVfOC/nVgSgJtOW/5sF+aZRGylQNpduwl3hB/fqFGuPqPFQD
+NoVmCen4YQkzXcKF/cWum89GQgrH5sPlUqI0FFbeieSWV6fXeOjk6Jeyccd6VBFm
+3jSM6GVEwUqxe0pCBkmkL8QOWBmwhgVdb9e+Sonq6RTcbZlZEOu4tmAuLmdlYMry
+Vdw/zIe4qoLufbgKljo9467Kp+yWdXGp8kk55bLyqDnxzyUzYGwVcCiAthP4cWKN
+Aygibby4eyHe+D2GjlQNBMfl3lzcFD4SpJBHMgfF4/7eOv79/estNiAI8GXaPdVd
+bHFNwNalSsdrL4V6rsDuCmjAqpRe3blhYTg133bMAXMbvah0w4O/zvZKcD8tS+oy
+enMEezN/n1yl+33BL6VPbslnNx+rJ6ZXeuFq1QuXdwpcnUGL2hzJH7RVlGDcZpar
+goB5Qax+bx0f4uZhuQINBE6OPIgBEADnVeGW0oQWDB+redSEpyA2VdgTidGhxLcF
+qIO6e4HnK5K6XpiCe4TE9tCBYpHz+IQLk2MtkkiNh1PlFIRWVZUqlxUD07kTvxaB
+Oy9eS5P7SGIzbnvVcpyaZkKcvGf5mRAf6Wm+RVCUraGhPg2Gzwua3vcaalIEYzyU
+3jfjS4ieUtyRdKxZ7y62mzOT5bL0YZLgiTt4HgfMhMxxOD/JOVZkoCvVmOghGLok
+F08eFC8UqXA0tkyV5AOt0V+Beea30IGbcKDdzZGfaniyYJG6qRvblZnHFw3/+2mp
+kj30AtOdPbmKmWgc+UQQrwEwn2WhOBHnovEKDdhTGSEH1tg3EcFpsZw8YRDSH7pH
+Bpb6kzjDn1UQ8fFkoPEA62OmPBeE9x8n8XlMJsxMh8p18kxRQDTYuOmS/7UK5ty2
+EU7RibhBTjz125Ta5w7V+fb12m1XMrdj1vP4o9Dnjb2f3LqoiPG3lyCgyNqpIKoc
+YX0cgFEiUzFpWY4gnFuHNi5ie6Tl5FAZz5chAew5ItlwatSGoEx9DS//cSWc9b+w
+djV3uvhqWaTX5zGVFu9pOvR08BrB7vgR6Rl8lrEoZzu5x0r7KjJgfb9efnuLadn+
+kL/4+f43XMAQFQBGHBDxREOhokIXkQzzbad+SQu7Fz4Zkfnj1A8Xvf5TopIC8Ruk
+6lLByh9chQARAQABiQIfBBgBAgAJBQJOjjyIAhsMAAoJEGx0AWyZNptWw54P/iey
++h3QvPN5BQwlp44YC/oMQrbB9mH3oq6z0vqC/MainKHnclGDxkgzmq6IBaE8npgr
+JqcoL7WZ4Kid6StGBCA09Ah+ZhwCYl+0Nob8/jZbqO6ghTb/SJibWkYjwRpSfouh
+bAT20HAkC1H4Aw0HQ2a0vehBJZ6chPksKSYl9lh2Q+wLCXDz+VaYOvkaJ6reMe1g
+TiuCb20Xn+N7BVIQbcVb8XJ6thlaeKK03r+Sum8FOmvcbLME9VsXnseVCaqQQa49
+54CtSFmiarnu6xN1AGzuSHRzlhZaCksFRDjs5Qw1EvVpgUckM8CdfRj2wtyTNB1e
+eFj6gT0Q9k1ZQv58whR6ZY1uavObiq/CemGc7ZpXZTbVHicW6WNpYTz9aziHzUYx
+zVed8gfTmcPi7Mi+FZC2NjX4BAA6ICeK33wWO0OBq/rcXLsUhCprrKZY4SUsKrtV
+YxbBRTWJFMsBVuXTX+uPdey7WxA0XCmKIcp2wDpptkq25yvu2UgAwrXyTpg0StuV
+XN0Wj3YqFrse3tXiE+Vqe5/3H6Rc2+Rjh8ysMkmg/dKCAuys6Is2vAyfpXgZzcQO
+kyHxwitfr4d8mm6GW3/YluGi7oPCnH3FEsIoUZMyfGs8XW/zwH26mCIpYpu9rDHh
+V/dDwg+iqlaqtN6rIS4E1gML3K33OVsdUvlcodhE
+=gNdQ
+-----END PGP PUBLIC KEY BLOCK-----
+-----BEGIN PGP PUBLIC KEY BLOCK-----
+Version: GnuPG v2.0.14 (GNU/Linux)
+
+mQINBE6OPIgBEAC5XC3CGy1peReJPUk2J5iCvmdSIOEaiOfULaz+1WiKgBMovUAK
+CbIuCB9ZMjfT0gceAgqj4UeCeRhPi2dVxVl/r1LCyGMJ1hZyEGDk7No1QkCemnCD
+4yiBD+BX+9V1Zuqar1goV5tMXoE0kq0cyoih/c6t8qrP5Wf/BTS118TBFAHRql/y
+Yle+g3YEk0uZ4yuedCVQIrjTWe46fSAacF8eGluGQVTbNj2aRacfu+UYPj8G241F
+8gTOTjZ3fJ80KTOItjiV8Gy+I6HDIWd7XMwNCO8z2G2lto0CBoDuaUukXgB5/CNF
+QL1S8Zxl2idxEQStZzKTKmZZcom+GNMvL8xCE8sMgMSiZ8Q54fRevTQrDLPpIhl0
+g6V1v7vOfokWxf7K6t+KZXB7oZbc1YrW/kcjDUAMlS4gAQ4K7ngEK3qIEVquvXx4
+nLVM++zn0/0PVzES2gMnacP9mdsVsPbJx+Jwk6hZi1mYBS7DyZkoVJYgXqLdCtWe
+bH+VqdHiEHBrWYxHeoyHWjViJKZYh3ojcKuZMicptEzA25hz5HaKEkKv1PEYeKED
+G8tLVf09dU+HHdK7Sbca08hXFhDIiDlksFa67Zy7MriJRTun8rTCD1DFH3wyuubR
+u8bDo7UobBgSK3tdcBRKM8i3e3JQZejhKlROhJ56xIT3ivO3Coe0hTUDXQARAQAB
+tEtOZWhhIE5hcmtoZWRlIChLZXkgZm9yIHNpZ25pbmcgY29kZSBhbmQgcmVsZWFz
+ZXMpIDxuZWhhbmFya2hlZGVAYXBhY2hlLm9yZz6JAjgEEwECACIFAk6OPIgCGwMG
+CwkIBwMCBhUIAgkKCwQWAgMBAh4BAheAAAoJEGx0AWyZNptWXSsP/i75oSJ50HJB
+RSni0oT6SxDG0YbsGj3Ojwn2FppOOpVW7kjSRPTb/gSTHYlxNvBxI+nkyKx4Z2+I
+loDaH2LUsxdAkBor9a58ykW9BD5Yd/5nRvduHp71XV28b3/xnN5H6kbCUr5yWZPV
+Z4//o8L12PS3Jy0ic9cQQF7vzuqQOvPpncruBChikEnSrwFmZI+UMRBBduCck6PN
+yeWTjb8ipfJtvwfOvQHX5+AOSz6zwICkzVfOC/nVgSgJtOW/5sF+aZRGylQNpduw
+l3hB/fqFGuPqPFQDNoVmCen4YQkzXcKF/cWum89GQgrH5sPlUqI0FFbeieSWV6fX
+eOjk6Jeyccd6VBFm3jSM6GVEwUqxe0pCBkmkL8QOWBmwhgVdb9e+Sonq6RTcbZlZ
+EOu4tmAuLmdlYMryVdw/zIe4qoLufbgKljo9467Kp+yWdXGp8kk55bLyqDnxzyUz
+YGwVcCiAthP4cWKNAygibby4eyHe+D2GjlQNBMfl3lzcFD4SpJBHMgfF4/7eOv79
+/estNiAI8GXaPdVdbHFNwNalSsdrL4V6rsDuCmjAqpRe3blhYTg133bMAXMbvah0
+w4O/zvZKcD8tS+oyenMEezN/n1yl+33BL6VPbslnNx+rJ6ZXeuFq1QuXdwpcnUGL
+2hzJH7RVlGDcZpargoB5Qax+bx0f4uZhuQINBE6OPIgBEADnVeGW0oQWDB+redSE
+pyA2VdgTidGhxLcFqIO6e4HnK5K6XpiCe4TE9tCBYpHz+IQLk2MtkkiNh1PlFIRW
+VZUqlxUD07kTvxaBOy9eS5P7SGIzbnvVcpyaZkKcvGf5mRAf6Wm+RVCUraGhPg2G
+zwua3vcaalIEYzyU3jfjS4ieUtyRdKxZ7y62mzOT5bL0YZLgiTt4HgfMhMxxOD/J
+OVZkoCvVmOghGLokF08eFC8UqXA0tkyV5AOt0V+Beea30IGbcKDdzZGfaniyYJG6
+qRvblZnHFw3/+2mpkj30AtOdPbmKmWgc+UQQrwEwn2WhOBHnovEKDdhTGSEH1tg3
+EcFpsZw8YRDSH7pHBpb6kzjDn1UQ8fFkoPEA62OmPBeE9x8n8XlMJsxMh8p18kxR
+QDTYuOmS/7UK5ty2EU7RibhBTjz125Ta5w7V+fb12m1XMrdj1vP4o9Dnjb2f3Lqo
+iPG3lyCgyNqpIKocYX0cgFEiUzFpWY4gnFuHNi5ie6Tl5FAZz5chAew5ItlwatSG
+oEx9DS//cSWc9b+wdjV3uvhqWaTX5zGVFu9pOvR08BrB7vgR6Rl8lrEoZzu5x0r7
+KjJgfb9efnuLadn+kL/4+f43XMAQFQBGHBDxREOhokIXkQzzbad+SQu7Fz4Zkfnj
+1A8Xvf5TopIC8Ruk6lLByh9chQARAQABiQIfBBgBAgAJBQJOjjyIAhsMAAoJEGx0
+AWyZNptWw54P/iey+h3QvPN5BQwlp44YC/oMQrbB9mH3oq6z0vqC/MainKHnclGD
+xkgzmq6IBaE8npgrJqcoL7WZ4Kid6StGBCA09Ah+ZhwCYl+0Nob8/jZbqO6ghTb/
+SJibWkYjwRpSfouhbAT20HAkC1H4Aw0HQ2a0vehBJZ6chPksKSYl9lh2Q+wLCXDz
++VaYOvkaJ6reMe1gTiuCb20Xn+N7BVIQbcVb8XJ6thlaeKK03r+Sum8FOmvcbLME
+9VsXnseVCaqQQa4954CtSFmiarnu6xN1AGzuSHRzlhZaCksFRDjs5Qw1EvVpgUck
+M8CdfRj2wtyTNB1eeFj6gT0Q9k1ZQv58whR6ZY1uavObiq/CemGc7ZpXZTbVHicW
+6WNpYTz9aziHzUYxzVed8gfTmcPi7Mi+FZC2NjX4BAA6ICeK33wWO0OBq/rcXLsU
+hCprrKZY4SUsKrtVYxbBRTWJFMsBVuXTX+uPdey7WxA0XCmKIcp2wDpptkq25yvu
+2UgAwrXyTpg0StuVXN0Wj3YqFrse3tXiE+Vqe5/3H6Rc2+Rjh8ysMkmg/dKCAuys
+6Is2vAyfpXgZzcQOkyHxwitfr4d8mm6GW3/YluGi7oPCnH3FEsIoUZMyfGs8XW/z
+wH26mCIpYpu9rDHhV/dDwg+iqlaqtN6rIS4E1gML3K33OVsdUvlcodhEmQENBE7i
+WaIBCADX/FrkgaxVLnDRU3hGfPdt7grEHinVgIct9PXNve1L/MhLq7stpVzJNUVz
+0TKC0njU3sJyYp5Imyojx423huOBTOTtn8KyPTZN5o8hBnr5FfbzBwZj6HcuUYc3
+UWCw09SzB1LyrunO+s8n/1jeMzVSFrzlLQrD62jmjVDq3FSieo6EdrQHjU9tUXeb
+mLk/VfwNcITuCL3X2Gs6+RutRgqM4OZJaQGo43U8gGRA5ZHvXm+RCbS5F7oItqSK
+JalnHxLzjwr4dM8Cc+uGAetcpbj3Vycxt/p0oehRUjdBWe69dVMic6XykwFpgP+9
+d/ExrJ8kbYYEqtdEwtVU/l02Cn+bABEBAAG0R05laGEgTmFya2hlZGUgKEtleSB1
+c2VkIGZvciBzaWduaW5nIHJlbGVhc2VzKSA8bmVoYS5uYXJraGVkZUBnbWFpbC5j
+b20+iQE4BBMBAgAiBQJO4lmiAhsDBgsJCAcDAgYVCAIJCgsEFgIDAQIeAQIXgAAK
+CRCEGNwlV1omRQ1kCACkbgPATgZ8WpDG9/jzZHh4Y0k5MHy8wsBEnMaGHmjLiKf8
+g/1dJgLiSeizsH/idBGBix4qu/5tZ0ayd2jOE15LaVK1NI1ja57y5wWa6/LFIBTN
+wY1ODTiEGlJ/TIDQTRmBWACfv6nb7VBFVQzETLm8RfRQuqYs617E9RY95u26/Rx2
+Sxw6QlXUzH/BwrQ6H0eR4rl32P90amIpGd8sHuMhMRI4kcTmzlJxSiokXceU76Fq
+6ACghjwrnaYy18wS2ZNq78Fye2kEabHdTtcXYiw0x5sDSdPwWLG+WrcoLcHVJTPb
+QhCftCzTjUZh8ybFJIr8zn7Fy7P6oQpNYdZpROFEuQENBE7iWaIBCADDlUDcfzdR
+RjxsyrYnP5Zo/e6Xcr3+1kiGUoxpf73VOsz7F9NqYqiR4ZJr1PwQPKscbQlD23PZ
+81A4p/nRhDAQQA2vcvijDECKu5T7wscIxrE1DYqRuZhlXreGQvhpH75XpDde+vIR
+9xmOz/uY30J4u2YoQaq9/ZKSU3RKzQ4HIOSJlvM9SibV/OFL8eT/kjPNXswczNk3
+XLDTJQFJDynMLwdBUH71BeC8q3WMWhQjmv0gVzegNAkRdW3CqNWx2by1v2V0Nkb2
+AkdHZQbSo8Q5vMz6flntWqbLWOEmrFa5xYX8QZUgo0oRWcI/0Lw6z+C/caqVJ015
+zv1lyJ0Q3ZDvABEBAAGJAR8EGAECAAkFAk7iWaICGwwACgkQhBjcJVdaJkWIfwgA
+rY4RXWny/twz3e5HK1p/rCDAwB5cVzXSoITNRBmGDXbyNBievfOr5UGi8TKWVR/o
+UxnG87ato8QxjU5s/9np0cgPGvOptN5dgrrzis9iSUhsgGs1UrR4UOo7xLW+jeX6
+KDAe0IIx25rOscycKv1inIY7IW/4bg/pmWUXYHUfIJGkChkzx+fdpEygvUFS0dSb
+MAlpYHsVMBhlW8cSk7uhWLeH6Abp7++ZdDqhnIfL4XG1IMUslkjQ5TTEfGxETCm2
+oe/dVCR4blsExaub6gUxyRiRzxaWaCYt+AT0vuTbBqDQduJqcry0AyTJe7O4+yo4
+bEnlTfUDA6C/gOJPKv4JXg==
+=EIZj
+-----END PGP PUBLIC KEY BLOCK-----
+pub   4096R/7F1183BD 2012-06-13
+uid                  Joe Stein (CODE SIGNING KEY) <[email protected]>
+sig 3        7F1183BD 2012-06-13  Joe Stein (CODE SIGNING KEY) 
<[email protected]>
+sub   4096R/3384EA8C 2012-06-13
+sig          7F1183BD 2012-06-13  Joe Stein (CODE SIGNING KEY) 
<[email protected]>
+
+-----BEGIN PGP PUBLIC KEY BLOCK-----
+Version: GnuPG/MacGPG2 v2.0.18 (Darwin)
+Comment: GPGTools - http://gpgtools.org
+
+mQINBE/X+B0BEADHcoRCi6OiC3TNh6xDkiE40YdDBM1g/YyF5r1Kim3+09bgGlm4
+RLIgdbhh+bQyqPJGQSRal6ypRFRuzlXskJcxkVZdU8XDzCBk6f5d1AgNbgD9YFLo
+NjEocYPAE9X10MDQ5a8eI7KBSrFYL4tDI7q8O+ZayaFFdkfgjxw5+Mg81H93pm77
+h9rHNOMCxQxCJusrkh5MIqcttJhWSVxukRkYoty3qNLzDjR5C3z5v6Xwdg+pii8g
+SEC1DvMXxhJVRPO24uXHKbrL2zbyh2GP0WgSuogcg1smGNh+6EGFhZNJj3uKCljh
+738uE6dmQsDwXwgXtoQfUi4cnMWmRWhxA7XSdt46XdeoWGblFp3hKTWGKHX1AVYY
+avUMajg0ISodtW6XIHoxHZzfH5J1LQbDWbWrkXZvIoPI13dqeJQqvqlpmGZtsYYY
+7vH3Tu8FTXLNUx638b8iBvGYSvEbm4GEdtqVTnOC80fYrflRMdWUvVPWIDvGat1F
+ucTg71QDmoEU4ZMaSRvn/8gX2DWjTooxh/k1YtJJAGV21P3GwxHPqNaa0IMLwBlV
+589cJMLoPGmXYcn6v6CD+A92ZltUOowx94vlFTZ5W2u/MPUMDQwTi9CpRsue+yyR
+Huz/yTi3KEyUTNRn/oXFhvW/f6UZmsLAHbt5TalyWrFIEHz64IRNfZhDcwARAQAB
+tDJKb2UgU3RlaW4gKENPREUgU0lHTklORyBLRVkpIDxqb2VzdGVpbkBhcGFjaGUu
+b3JnPokCMQQTAQIAGwIbAwIeAQIXgAUCT9f41AULCQgHAwUVCgkICwAKCRCGUYeW
+fxGDve+2D/sFNDDpjZA73+ISPoojtitMiFiYA+9CZ9uLdskqS+liTGwNBaD3ypV1
+RPdrKwjlb03ELFP7EH48ktUQZBKootVHkuyBQBI8WMEJEaAUYutuEDZgfoICTZ5L
+I/XTIW3Ih2wErBcp281nU9lkZ91Kte7gMfjsA1bZPPtL6w4p8YhxD1zMerMQ0HP0
+H2jYJopbyTn+BrrkAc9I7NpPvt+XiCewXZGmDPGl4fBTloUy+xxAOWVzUhX8hwmh
+N9E77L9gLKBi+IQNn3jWs4Kck7QH5iULXAGZ1YdDoaMOdFO8nwmeDF9/AdyhTj37
+zF4CvsoEBnIxZPkxnH56orD1R2MmZd1P8g2yoYrAF9XdIRPnO30JZDXP6czZX9PP
+eJcKwXofx5ldzLb6og1ZZGwY742u6SKvc7rj+CryaPqixu4OFTWCtpDkewCiNHwL
+OUnhBXDSqxaFgZT3diOypCOzKE1PtygbeVUcxCMtGdF6s1EGMfamRZc+SlXhoG9i
+f9nZRrKjSo/PeFxPMTI7s8ACpISplpiaHLx6s/GK+AR51XwG65miHX9Z6empNFXk
+iLYxnZjo0nnHU8qM+XSjs5ARFlvGbK5Iw1wUhTVX31nInT9OAojuY8MgPwyPH413
+W+qMWRUziWwEmeXLNDPq3BNksfxp3Ac+ztCkzYfKIDa8L4uNGyIAu7kCDQRP1/gd
+ARAAxCKGUjsUAqsXKA2TNzOAfJDGpeQOzHFmADJiPXvThJ8hsc8DgfGOTQZfwLPh
+P4MYlt1E4dvTukOYIXC/ORCtUdGNaZNh0JNkpGggOUcCJBRjqwA9TXeKUESzdKFG
+BX26GZZtGfvzYTz8hSyEU0VnVQbK1ZrcUiMh86oFGOUH6SG3ZFDx+QoMcC5felZx
+LzNtOOLOwRKW3XXXsuwEFsBi/Nre/0jLehWgpGzSq2Sx8uIRQ1gn08fA999DEFOk
+MqGHs26FijrqUm99pvIgYM1LU0WgvdjQY5wnoopMvYDhtOIIXjgOlnd/D8mBAbei
+zt54ag7RvrugjwTKmrtaSd0OhYDbZDQOPVQNGveug1pEl8l/dS6Og2t824DRi7zu
+cZyUd6y05FIJpHf+Pjt7M2bijJ7GJZm04iBfG4TZ+zaOwe6OImvC3DWxhFCEJqxq
+v5oPpaoqby/R/tKbo48Dq1lZsbt9Rw2moDndYJNuV43XWNfGgwz9PTpM56BG5NIB
+j/g4EQgwvy60dNzGQqU2FSIBLcLGGTEDnDJpYStflOgTBNdX6uPceHKoNQ1/qoRg
+z+AqcHSSxEAs4MaLiYzFQ5+1UScrCka6Ox6boiMQ4K7PJlOv12jTT6TNEU6SyFlV
+LPLhK/XC+qrsmkK3fx6j1ZbwZfkue/74CqwydcVRQeGAF3sAEQEAAYkCHwQYAQIA
+CQUCT9f4HQIbDAAKCRCGUYeWfxGDvfEaD/kB5qzWV/Pq9wWz0acrNEGnrdc9s2uB
+1MkB2p+2LeZzgmvAeHTKQy0y3KEd4VY1JQrB+HPOyJARf6fpXkZIOWlDJ6wbHxnZ
+6pAAtpPKIVghkv3TPAvTuyqxMpkBAhHm/hvpj/3OmLPitBOFK3p9F/FAmqHADato
+DuzNdkF4v/Xl1LbaBI+lyHA8BM3pQnci+Hq3ozxW5JEqk4HYliPpJe77ifHukJn3
+cwlnIz0SHMV7+hzRrJ/yaAmf0cfVZHb7iKBarh3SWm7rN9g1LJlhgkRmP4njkctf
+kSWLSAYbjfU/5yOfhEPrD2w5Y3qjErHAmBafYkMp0J1a8Acg2Unb1Q5xh23DFr1m
+7VSjLTu/Qe9XTs+0EMvvVsp8vvQSqs7VP5GRfuA6f2DXlGy1cSdbS8m5mhMMCTei
+Rsm1N+68cmON1stRrtpqxktokwNf5GIY1AxC2KT4HLPjS5AQHuJBIO1vmfVnKF3Q
+pVzipF+8VFilrN0IerddThBYgcAUQM+9ZzuPX0yCORpEMwlq0WJTn0KlJc/jJjJb
+VdyQ2DksBCpN7T+i8tTM9moq/RSieRIjEXKeEfM6EoNVX5x/f4HQKv7YJ9mr1gDJ
+v/W3AOBR+NSWhBudk0N6Jroga5DI8X9NGz0I98Z4Nvme5I6jikcjE0NnKuJ/APLU
+KzfEt4aFC/3Gzw==
+=qUcq
+-----END PGP PUBLIC KEY BLOCK-----
+pub   4096R/E0A61EEA 2015-01-14
+uid                  Jun Rao <[email protected]>
+sig 3        E0A61EEA 2015-01-14  Jun Rao <[email protected]>
+sub   4096R/942424A8 2015-01-14
+sig          E0A61EEA 2015-01-14  Jun Rao <[email protected]>
+
+-----BEGIN PGP PUBLIC KEY BLOCK-----
+Version: GnuPG/MacGPG2 v2.0.22 (Darwin)
+
+mQINBFS1uQIBEADnz+VGKABlkZae60UPIFqliX71KupZH6JnYEOZQtYwwGGAvuGk
+Ws+mDRIfLcE31+C1sDj87pmHfONUbPMI9AhjHVCPwc6rVLfAT6dVdGrmjZ5kRerT
+S25+a3O4njZxekMt6oWr0PcZoEkuhN6EoJ5u/BanWwtayWe2poZp+VVF7MG9vheH
+nsiM+Wfbvt1UiwIvpJ1U9HfnKL8te8d5FlCuxCJTvv8PYfUpGn2MpcvmrjNknDMc
+T2sReeU+6gD+f69OcU2SrvYUB5fyhKmeU42byImLCF0Kzz2dwvFMNzDHTKv42N7Y
+zyFEN6r7VrugrAyQiDOAdy8PBrJBCtg309XrakMgWleeH2PfkMpEMVVcR2sOCVAJ
+wHfFrXOtrg6mFJpphqIPydRYKCxy6XQyvy91igK8W1fS/iKVVpd+WFw0HF1x9p4L
+3UC+0BYyf2wsQSpE1CRftnhxog8xbkDQJmy1qfoBuYXFXS4ARpewW5JxZbuSsqLC
+fUU8rWGn3yngzOQmYURFVjfP0bJmOcwIGmDathX+8CvxmX6J4qsHlrfJEZ9aBhXA
+FYGfIK7OmRU/ZgtcOELyT18EHF3mOzQVNcmweKHz1QvJW/lsPRPDy3yqWmrivmVR
+/Sw0yTuOiNRj28qSStEJahA1seRVS9upwEcMU4bhN0D/MtLR0s4/9SDqYwARAQAB
+tBpKdW4gUmFvIDxqdW5yYW9AZ21haWwuY29tPokCOAQTAQIAIgUCVLW5AgIbAwYL
+CQgHAwIGFQgCCQoLBBYCAwECHgECF4AACgkQq2H3D+CmHuqv3RAA4FBD0uHIZj1H
+nZd4ddU7YxPPcuuoEaIzEZ3LJFtvyA8vjW50ZRvp1oa95D0qu90lo2Tf3vDaZoxR
+isGI1jrcjwjXfkgKqcMOWsGwJfqlTnE54nIzJ5x9C7ePzAQCdlq4WE/3OHc6V/zK
+KadAlJajSR24yT/oZUeu8W2oA86HF42fJUacoERATv3XBRtxClDdYkQfYxumOdSn
+/hbxaT+SUfoLZKfuWaRCFzvvDpudGUCnnCwSOXhDQjsYxy7Y7WYqirh6LhkJIvy3
+ejyZebM9IDk6bAm9VfOKLGGtlAQ8JVkLGidk7+1NMt5eAfxV5/9X6AuiOPdYV+cq
+4p4ZuqPjVbDDmQwsGt15EVZ92kGcVtidhttZ60znslhSrn9HVjYSEi9zT1YdbDSW
+r2rwTq1apXfsvS42sT0Xj1x2ALR5Db4cFStpY48QBkg453fmnIBE70oPWwy9wWWY
+YniLyYiwSnZAyocaFq/xCY74I6S1FEB7pZwEo2aeR61hmiv0cC33MY3b36HP3LaQ
+eSZGEyMjw3Cus3o5BUmrFirokyQQSl1p2dojTppAYWzpxT1D10JfecfLlMmCcres
+HhyxWE8xuuEciZLMvWFjKJCSxrbCU9BbJ6JAVp/xzDkfmxiImJS4XQDa0hhiIAMg
+ustPi9r867y/nSmpmmuhBHwJzYCdOKm5Ag0EVLW5AgEQALpRBfJ7qvOuZjg4VNZR
+lI9n2IzSIFLwKhQibGTl9i5oQD/g1VoumjwZm6N/QWbxoU8qFHWou1KGGEibnTsC
+xJJKg1YspC2WLRRqzWzsgHNrF4DAwtRQTolbj/c0c30iNgZVUbguVsZIwUv8Yj90
+v1wdVxUrcyHF5KJrNz0I+6qy15z/T97OpQ3y7S2WVrGvb1rY03NpwQcOA6lam5b1
+jSQKNNUs5Fldr27YY9PDufOI37trCwLXtveW2N+clO2aKpcrtYC9m2WaFilXbZ3z
+1UdfjLHbMCBal+UabebLM/IQ882fhzXTk8GA/54ILW1kOaMQ8FoqjXqzXiVo3rgm
+IqpXvoy0GKL7T8Dawf+AoiyHvNvG7YyEczrNYy2CelHXz7rMfzk60n7rE7dMVt1s
+/behN9KIdXKYMUcw/wByiScTbSGagJhZcyApdTYlTvyy8U82DIEPcO46+eUnQRcv
+JRPLgLsY21jrXzqUkL1MEESYMerc8nsO+FdkBLrMcuBgQpxlGtLeODG3tnXakBdD
+qfdOorDs13HGyzclc26ACwPXofd5qV61npBR3zNGro1Vb1hCDRUOEHBimj/amj/Q
+J0CyeDWt00meChlTWefhV7AjAIRxYnX9gcs6FFlyYKt8m4BUQEWBicS6jhVxM1Sa
+ucEKSEF/PP+9AfuxBmToiohRABEBAAGJAh8EGAECAAkFAlS1uQICGwwACgkQq2H3
+D+CmHuqeaA//cHM2oONqiLtk0oRYa577I0uZj2Pddlq38cv2rhRNXAEpIXTBZRrV
+3PwFVOqpysOC9A9TQe3rlDU/AvFw965UDJ+nFdgN5bc6+O58wr//wjTf3zGrxjGT
+LINmRSU7RoiKXdZ9At9/Ra1HDIzGqCu5nHqdYE+EH1ucu1CWIj/caCjgoZ6yRacI
+Viawor4b+T3Yov/hWfJzawDWbV9COpeRU4XPfEAaWvuX4ZmZYr2PGI/KEVIbXXUb
+RyggTu6s8JdU4K+rytzlB0RkC75yjfWuY+4lD8Dv/ASPD+R8NmP7xuLsGS8YUd7t
+H0+U4Y0GI7/ZqHeKw8QjJ+IBkvUuOW1zFRi7utKygueP/jMAhk3HhB/aSoxU7xyC
+CTLeS2/KDSdm3BjgiSak7/2sr7SHPzxyR8FgK3AWDrupkx8o92DHh/2i2/47d/iy
+uG/+sPuvsx2/C8NIPUke7augQefOWGLUuTnOj5H/oINYu2AUy4ceJajgOKrJcE/3
+XsBXGoqGwgGjI7OER1vnPZGoby46vmcMi6MFjA/eITzLzCcGT01F8jb7Rh5kkn/J
+el67DxjpxG2aTeBWgmqnk+YfdEIleRKflSyuk9azhoHD4leD2L7i7d9LQGEe0M3R
+LyTaTKeMBhCrhjTTJ4dmMozlivQx6gqOwtE14osBHAzoCYtcWjp6GKA=
+=npff
+-----END PGP PUBLIC KEY BLOCK-----

Reply via email to