This is an automated email from the ASF dual-hosted git repository.
divijv pushed a commit to branch trunk
in repository https://gitbox.apache.org/repos/asf/kafka.git
The following commit(s) were added to refs/heads/trunk by this push:
new 49c1697ab08 MINOR: fix typos for doc (#13883)
49c1697ab08 is described below
commit 49c1697ab08189c9707d04a6078aa9d5b69ed3aa
Author: minjian.cai <[email protected]>
AuthorDate: Wed Jun 21 15:57:43 2023 +0800
MINOR: fix typos for doc (#13883)
Reviewers: Divij Vaidya <[email protected]>
---
docs/design.html | 2 +-
docs/ops.html | 14 +++++++-------
docs/security.html | 6 +++---
docs/streams/developer-guide/dsl-api.html | 6 +++---
docs/streams/developer-guide/dsl-topology-naming.html | 2 +-
docs/streams/upgrade-guide.html | 8 ++++----
6 files changed, 19 insertions(+), 19 deletions(-)
diff --git a/docs/design.html b/docs/design.html
index 9485ab9c500..139eb0c1e3b 100644
--- a/docs/design.html
+++ b/docs/design.html
@@ -339,7 +339,7 @@
and mark the broker offline.
<p>
We refer to nodes satisfying these two conditions as being "in sync" to
avoid the vagueness of "alive" or "failed". The leader keeps track of the set
of "in sync" replicas,
- which is known as the ISR. If either of these conditions fail to be
satisified, then the broker will be removed from the ISR. For example,
+ which is known as the ISR. If either of these conditions fail to be
satisfied, then the broker will be removed from the ISR. For example,
if a follower dies, then the controller will notice the failure through
the loss of its session, and will remove the broker from the ISR.
On the other hand, if the follower lags too far behind the leader but
still has an active session, then the leader can also remove it from the ISR.
The determination of lagging replicas is controlled through the
<code>replica.lag.time.max.ms</code> configuration.
diff --git a/docs/ops.html b/docs/ops.html
index 60a9bd515e3..b506b286064 100644
--- a/docs/ops.html
+++ b/docs/ops.html
@@ -892,7 +892,7 @@ A->B.topics = bar
</p>
<p>
- It is therefore important to keep the MirrorMaker configration consistent
across replication flows to the same target cluster. This can be achieved, for
example, through automation tooling or by using a single, shared MirrorMaker
configuration file for your entire organization.
+ It is therefore important to keep the MirrorMaker configuration consistent
across replication flows to the same target cluster. This can be achieved, for
example, through automation tooling or by using a single, shared MirrorMaker
configuration file for your entire organization.
</p>
<h5 class="anchor-heading"><a id="georeplication-best-practice"
class="anchor-link"></a><a href="#georeplication-best-practice">Best Practice:
Consume from Remote, Produce to Local</a></h5>
@@ -1415,7 +1415,7 @@ $ bin/kafka-acls.sh \
<h4 class="anchor-heading"><a id="replace_disk" class="anchor-link"></a><a
href="#replace_disk">Replace KRaft Controller Disk</a></h4>
<p>When Kafka is configured to use KRaft, the controllers store the cluster
metadata in the directory specified in <code>metadata.log.dir</code> -- or the
first log directory, if <code>metadata.log.dir</code> is not configured. See
the documentation for <code>metadata.log.dir</code> for details.</p>
- <p>If the data in the cluster metdata directory is lost either because of
hardware failure or the hardware needs to be replaced, care should be taken
when provisioning the new controller node. The new controller node should not
be formatted and started until the majority of the controllers have all of the
committed data. To determine if the majority of the controllers have the
committed data, run the <code>kafka-metadata-quorum.sh</code> tool to describe
the replication status:</p>
+ <p>If the data in the cluster metadata directory is lost either because of
hardware failure or the hardware needs to be replaced, care should be taken
when provisioning the new controller node. The new controller node should not
be formatted and started until the majority of the controllers have all of the
committed data. To determine if the majority of the controllers have the
committed data, run the <code>kafka-metadata-quorum.sh</code> tool to describe
the replication status:</p>
<pre class="line-numbers"><code class="language-bash"> >
bin/kafka-metadata-quorum.sh --bootstrap-server broker_host:port describe
--replication
NodeId LogEndOffset Lag LastFetchTimestamp LastCaughtUpTimestamp
Status
@@ -1427,7 +1427,7 @@ $ bin/kafka-acls.sh \
<pre class="line-numbers"><code class="language-bash"> >
bin/kafka-storage.sh format --cluster-id uuid --config
server_properties</code></pre>
- <p>It is possible for the <code>bin/kafka-storage.sh format</code> command
above to fail with a message like <code>Log directory ... is already
formatted</code>. This can happend when combined mode is used and only the
metadata log directory was lost but not the others. In that case and only in
that case, can you run the <code>kafka-storage.sh format</code> command with
the <code>--ignore-formatted</code> option.</p>
+ <p>It is possible for the <code>bin/kafka-storage.sh format</code> command
above to fail with a message like <code>Log directory ... is already
formatted</code>. This can happen when combined mode is used and only the
metadata log directory was lost but not the others. In that case and only in
that case, can you run the <code>kafka-storage.sh format</code> command with
the <code>--ignore-formatted</code> option.</p>
<p>Start the KRaft controller after formatting the log directories.</p>
@@ -1779,7 +1779,7 @@ $ bin/kafka-acls.sh \
<tr>
<td>ZooKeeper client request latency</td>
<td>kafka.server:type=ZooKeeperClientMetrics,name=ZooKeeperRequestLatencyMs</td>
- <td>Latency in millseconds for ZooKeeper requests from broker.</td>
+ <td>Latency in milliseconds for ZooKeeper requests from broker.</td>
</tr>
<tr>
<td>ZooKeeper connection status</td>
@@ -2447,7 +2447,7 @@ These metrics are reported on both Controllers and
Brokers in a KRaft Cluster
<td>kafka.consumer:type=consumer-metrics,client-id=([-.\w]+)</td>
</tr>
<tr>
- <td>commited-time-ns-total</td>
+ <td>committed-time-ns-total</td>
<td>The total time the Consumer spent in committed in nanoseconds.</td>
<td>kafka.consumer:type=consumer-metrics,client-id=([-.\w]+)</td>
</tr>
@@ -3461,7 +3461,7 @@ for built-in state stores, currently we have:
</tr>
<tr>
<td>hit-ratio-min</td>
- <td>The mininum cache hit ratio.</td>
+ <td>The minimum cache hit ratio.</td>
<td>kafka.streams:type=stream-record-cache-metrics,thread-id=([-.\w]+),task-id=([-.\w]+),record-cache-id=([-.\w]+)</td>
</tr>
<tr>
@@ -3562,7 +3562,7 @@ CurrentObservers: [0,1,2]</code></pre>
<pre class="line-numbers"><code class="language-bash"> >
bin/kafka-dump-log.sh --cluster-metadata-decoder --files
metadata_log_dir/__cluster_metadata-0/00000000000000000000.log</code></pre>
- <p>This command decodes and prints the recrods in the a cluster metadata
snapshot:</p>
+ <p>This command decodes and prints the records in the a cluster metadata
snapshot:</p>
<pre class="line-numbers"><code class="language-bash"> >
bin/kafka-dump-log.sh --cluster-metadata-decoder --files
metadata_log_dir/__cluster_metadata-0/00000000000000000100-0000000001.checkpoint</code></pre>
diff --git a/docs/security.html b/docs/security.html
index 94c0ad0b935..a6cece897c0 100644
--- a/docs/security.html
+++ b/docs/security.html
@@ -262,7 +262,7 @@ default_ca = CA_default # The default ca section
[ CA_default ]
base_dir = .
-certificate = $base_dir/cacert.pem # The CA certifcate
+certificate = $base_dir/cacert.pem # The CA certificate
private_key = $base_dir/cakey.pem # The CA private key
new_certs_dir = $base_dir # Location for new certs after signing
database = $base_dir/index.txt # Database index file
@@ -437,7 +437,7 @@ keyUsage = digitalSignature,
keyEncipherment</code></pre>
<li><b>Failure to copy extension fields</b><br>
CA operators are often hesitant to copy and requested
extension fields from CSRs and prefer to specify these themselves as this makes
it
harder for a malicious party to obtain certificates with
potentially misleading or fraudulent values.
- It is adviseable to double check signed certificates,
whether these contain all requested SAN fields to enable proper hostname
verification.
+ It is advisable to double check signed certificates,
whether these contain all requested SAN fields to enable proper hostname
verification.
The following command can be used to print certificate
details to the console, which should be compared with what was originally
requested:
<pre class="line-numbers"><code class="language-bash">>
openssl x509 -in certificate.crt -text -noout</code></pre>
</li>
@@ -1269,7 +1269,7 @@ sasl.mechanism.inter.broker.protocol=GSSAPI (or one of
the other enabled mechani
</ol>
<h3 class="anchor-heading"><a id="security_authz"
class="anchor-link"></a><a href="#security_authz">7.5 Authorization and
ACLs</a></h3>
- Kafka ships with a pluggable authorization framework, which is configured
with the <tt>authorizer.class.name</tt> property in the server confgiuration.
+ Kafka ships with a pluggable authorization framework, which is configured
with the <tt>authorizer.class.name</tt> property in the server configuration.
Configured implementations must extend
<code>org.apache.kafka.server.authorizer.Authorizer</code>.
Kafka provides default implementations which store ACLs in the cluster
metadata (either Zookeeper or the KRaft metadata log).
diff --git a/docs/streams/developer-guide/dsl-api.html
b/docs/streams/developer-guide/dsl-api.html
index 867272603da..564f878736d 100644
--- a/docs/streams/developer-guide/dsl-api.html
+++ b/docs/streams/developer-guide/dsl-api.html
@@ -3324,8 +3324,8 @@ SlidingWindows.ofTimeDifferenceAndGrace(timeDifference,
gracePeriod);</code></pr
<div class="figure align-center" id="id35">
<img class="centered"
src="/{{version}}/images/streams-sliding-windows.png">
<p class="caption"><span
class="caption-text">This diagram shows windowing a stream of data records with
sliding windows. The overlap of
- the sliding window snapshots varies depending
on the record times. In this diagram, the time numbers represent miliseconds.
For example,
- t=5 means “at the five milisecond
mark“.</span></p>
+ the sliding window snapshots varies depending
on the record times. In this diagram, the time numbers represent milliseconds.
For example,
+ t=5 means “at the five millisecond
mark“.</span></p>
</div>
<p>Sliding windows are aligned to the data record
timestamps, not to the epoch. In contrast to hopping and tumbling windows,
the lower and upper window time interval bounds
of sliding windows are both inclusive.</p>
@@ -3385,7 +3385,7 @@ t=5 (blue), which lead to a merge of sessions and an
extension of a session, res
Common examples of this are sending alerts or
delivering results to a system that doesn't support updates.
</p>
<p>Suppose that you have an hourly windowed count
of events per user.
- If you want to send an alert when a user has
<em>less than</em> three events in an hour, you have a real challange.
+ If you want to send an alert when a user has
<em>less than</em> three events in an hour, you have a real challenge.
All users would match this condition at first,
until they accrue enough events, so you cannot simply
send an alert when someone matches the
condition; you have to wait until you know you won't see any more events for a
particular window
and <em>then</em> send the alert.
diff --git a/docs/streams/developer-guide/dsl-topology-naming.html
b/docs/streams/developer-guide/dsl-topology-naming.html
index cd11c132bb7..72e0c2887a1 100644
--- a/docs/streams/developer-guide/dsl-topology-naming.html
+++ b/docs/streams/developer-guide/dsl-topology-naming.html
@@ -66,7 +66,7 @@
<p>
By saying there is a readability trade-off, we are
referring to viewing a description of the topology.
- When you render the string description of your topology
via the <code>Topology#desribe()</code>
+ When you render the string description of your topology
via the <code>Topology#describe()</code>
method, you can see what the processor is, but you
don't have any context for its business purpose.
For example, consider the following simple topology:
diff --git a/docs/streams/upgrade-guide.html b/docs/streams/upgrade-guide.html
index c9c689e36c1..9979105a18e 100644
--- a/docs/streams/upgrade-guide.html
+++ b/docs/streams/upgrade-guide.html
@@ -158,8 +158,8 @@
<p>
<a
href="https://cwiki.apache.org/confluence/display/KAFKA/KIP-904%3A+Kafka+Streams+-+Guarantee+subtractor+is+called+before+adder+if+key+has+not+changed">KIP-904</a>
- improves the implemenation of KTable aggregations. In general, an input
KTable update triggers a result refinent for two rows;
- however, prior to KIP-904, if both refinements happend to the same
result row, two independent updates to the same row are applied, resulting in
spurious itermediate results.
+ improves the implementation of KTable aggregations. In general, an input
KTable update triggers a result refinent for two rows;
+ however, prior to KIP-904, if both refinements happen to the same result
row, two independent updates to the same row are applied, resulting in spurious
itermediate results.
KIP-904 allows us to detect this case, and to only apply a single update
avoiding spurious intermediate results.
</p>
@@ -284,7 +284,7 @@
<p>
The public <code>topicGroupId</code> and <code>partition</code> fields
on TaskId have been deprecated and replaced with getters. Please migrate to
using the new <code>TaskId.subtopology()</code>
(which replaces <code>topicGroupId</code>) and
<code>TaskId.partition()</code> APIs instead. Also, the
<code>TaskId#readFrom</code> and <code>TaskId#writeTo</code> methods have been
deprecated
- and will be removed, as they were never intended for public use. We
have also deprecated the
<code>org.apache.kafka.streams.processsor.TaskMetadata</code> class and
introduced a new interface
+ and will be removed, as they were never intended for public use. We
have also deprecated the
<code>org.apache.kafka.streams.processor.TaskMetadata</code> class and
introduced a new interface
<code>org.apache.kafka.streams.TaskMetadata</code> to be used instead.
This change was introduced to better reflect the fact that
<code>TaskMetadata</code> was not meant to be instantiated outside
of Kafka codebase.
Please note that the new <code>TaskMetadata</code> offers APIs that
better represent the task id as an actual <code>TaskId</code> object instead of
a String. Please migrate to the new
@@ -733,7 +733,7 @@
Also, window sizes and retention times are now specified as
<code>Duration</code> type in <code>Stores</code> class.
The <code>Window</code> class has new methods
<code>#startTime()</code> and <code>#endTime()</code> that return window
start/end timestamp as <code>Instant</code>.
For interactive queries, there are new <code>#fetch(...)</code>
overloads taking <code>Instant</code> arguments.
- Additionally, punctuations are now registerd via
<code>ProcessorContext#schedule(Duration interval, ...)</code>.
+ Additionally, punctuations are now registered via
<code>ProcessorContext#schedule(Duration interval, ...)</code>.
For more details, see <a
href="https://cwiki.apache.org/confluence/display/KAFKA/KIP-358%3A+Migrate+Streams+API+to+Duration+instead+of+long+ms+times">KIP-358</a>.
</p>