This is an automated email from the ASF dual-hosted git repository.

frankvicky pushed a commit to branch trunk
in repository https://gitbox.apache.org/repos/asf/kafka.git


The following commit(s) were added to refs/heads/trunk by this push:
     new d31885d33c0 MINOR: Use <code> block instead of backtick (#20107)
d31885d33c0 is described below

commit d31885d33c06588b934b68791ffb968b2ed4105e
Author: Ken Huang <s7133...@gmail.com>
AuthorDate: Sun Jul 6 14:49:51 2025 +0800

    MINOR: Use <code> block instead of backtick (#20107)
    
    When writing HTML, it's recommended to use the <code> element instead of
    backticks for inline code formatting.
    
    Reviewers: Chia-Ping Tsai <chia7...@gmail.com>, TengYao Chi
    <frankvi...@apache.org>
---
 .../kafka/clients/consumer/ConsumerConfig.java       |  2 +-
 .../org/apache/kafka/common/config/SslConfigs.java   |  8 ++++----
 .../org/apache/kafka/common/config/TopicConfig.java  | 14 +++++++-------
 .../apache/kafka/connect/runtime/WorkerConfig.java   |  8 ++++----
 .../connect/transforms/field/FieldSyntaxVersion.java |  8 ++++----
 docs/ops.html                                        |  2 +-
 .../kafka/server/config/ReplicationConfigs.java      |  2 +-
 .../log/remote/storage/RemoteLogManagerConfig.java   | 20 ++++++++++----------
 .../java/org/apache/kafka/streams/StreamsConfig.java |  6 +++---
 9 files changed, 35 insertions(+), 35 deletions(-)

diff --git 
a/clients/src/main/java/org/apache/kafka/clients/consumer/ConsumerConfig.java 
b/clients/src/main/java/org/apache/kafka/clients/consumer/ConsumerConfig.java
index 2e0c186186d..09ce436c81e 100644
--- 
a/clients/src/main/java/org/apache/kafka/clients/consumer/ConsumerConfig.java
+++ 
b/clients/src/main/java/org/apache/kafka/clients/consumer/ConsumerConfig.java
@@ -376,7 +376,7 @@ public class ConsumerConfig extends AbstractConfig {
     public static final String ALLOW_AUTO_CREATE_TOPICS_CONFIG = 
"allow.auto.create.topics";
     private static final String ALLOW_AUTO_CREATE_TOPICS_DOC = "Allow 
automatic topic creation on the broker when" +
             " subscribing to or assigning a topic. A topic being subscribed to 
will be automatically created only if the" +
-            " broker allows for it using `auto.create.topics.enable` broker 
configuration.";
+            " broker allows for it using 
<code>auto.create.topics.enable</code> broker configuration.";
     public static final boolean DEFAULT_ALLOW_AUTO_CREATE_TOPICS = true;
 
     /**
diff --git 
a/clients/src/main/java/org/apache/kafka/common/config/SslConfigs.java 
b/clients/src/main/java/org/apache/kafka/common/config/SslConfigs.java
index 5ca4980fc22..0b88f44f310 100644
--- a/clients/src/main/java/org/apache/kafka/common/config/SslConfigs.java
+++ b/clients/src/main/java/org/apache/kafka/common/config/SslConfigs.java
@@ -34,7 +34,7 @@ public class SslConfigs {
         + "this config are dependent on the JVM. "
         + "Clients using the defaults for this config and 
'ssl.enabled.protocols' will downgrade to 'TLSv1.2' if "
         + "the server does not support 'TLSv1.3'. If this config is set to 
'TLSv1.2', however, clients will not use 'TLSv1.3' even "
-        + "if it is one of the values in `ssl.enabled.protocols` and the 
server only supports 'TLSv1.3'.";
+        + "if it is one of the values in <code>ssl.enabled.protocols</code> 
and the server only supports 'TLSv1.3'.";
 
     public static final String DEFAULT_SSL_PROTOCOL = "TLSv1.3";
 
@@ -49,12 +49,12 @@ public class SslConfigs {
     public static final String SSL_ENABLED_PROTOCOLS_DOC = "The list of 
protocols enabled for SSL connections. "
         + "The default is 'TLSv1.2,TLSv1.3'. This means that clients and 
servers will prefer TLSv1.3 if both support it "
         + "and fallback to TLSv1.2 otherwise (assuming both support at least 
TLSv1.2). This default should be fine for most use "
-        + "cases. Also see the config documentation for `ssl.protocol` to 
understand how it can impact the TLS version negotiation behavior.";
+        + "cases. Also see the config documentation for 
<code>ssl.protocol</code> to understand how it can impact the TLS version 
negotiation behavior.";
     public static final String DEFAULT_SSL_ENABLED_PROTOCOLS = 
"TLSv1.2,TLSv1.3";
 
     public static final String SSL_KEYSTORE_TYPE_CONFIG = "ssl.keystore.type";
     public static final String SSL_KEYSTORE_TYPE_DOC = "The file format of the 
key store file. "
-            + "This is optional for client. The values currently supported by 
the default `ssl.engine.factory.class` are [JKS, PKCS12, PEM].";
+            + "This is optional for client. The values currently supported by 
the default <code>ssl.engine.factory.class</code> are [JKS, PKCS12, PEM].";
     public static final String DEFAULT_SSL_KEYSTORE_TYPE = "JKS";
 
     public static final String SSL_KEYSTORE_KEY_CONFIG = "ssl.keystore.key";
@@ -84,7 +84,7 @@ public class SslConfigs {
         + "the PEM key specified in 'ssl.keystore.key'.";
 
     public static final String SSL_TRUSTSTORE_TYPE_CONFIG = 
"ssl.truststore.type";
-    public static final String SSL_TRUSTSTORE_TYPE_DOC = "The file format of 
the trust store file. The values currently supported by the default 
`ssl.engine.factory.class` are [JKS, PKCS12, PEM].";
+    public static final String SSL_TRUSTSTORE_TYPE_DOC = "The file format of 
the trust store file. The values currently supported by the default 
<code>ssl.engine.factory.class</code> are [JKS, PKCS12, PEM].";
     public static final String DEFAULT_SSL_TRUSTSTORE_TYPE = "JKS";
 
     public static final String SSL_TRUSTSTORE_LOCATION_CONFIG = 
"ssl.truststore.location";
diff --git 
a/clients/src/main/java/org/apache/kafka/common/config/TopicConfig.java 
b/clients/src/main/java/org/apache/kafka/common/config/TopicConfig.java
index 08fad453568..9f5a9a16a5e 100755
--- a/clients/src/main/java/org/apache/kafka/common/config/TopicConfig.java
+++ b/clients/src/main/java/org/apache/kafka/common/config/TopicConfig.java
@@ -21,7 +21,7 @@ package org.apache.kafka.common.config;
  * <p>Keys that can be used to configure a topic. These keys are useful when 
creating or reconfiguring a
  * topic using the AdminClient.
  *
- * <p>The intended pattern is for broker configs to include a 
<code>`log.`</code> prefix. For example, to set the default broker
+ * <p>The intended pattern is for broker configs to include a 
<code>log.</code> prefix. For example, to set the default broker
  * cleanup policy, one would set <code>log.cleanup.policy</code> instead of 
<code>cleanup.policy</code>. Unfortunately, there are many cases
  * where this pattern is not followed.
  */
@@ -87,13 +87,13 @@ public class TopicConfig {
 
     public static final String LOCAL_LOG_RETENTION_MS_CONFIG = 
"local.retention.ms";
     public static final String LOCAL_LOG_RETENTION_MS_DOC = "The number of 
milliseconds to keep the local log segment before it gets deleted. " +
-            "Default value is -2, it represents `retention.ms` value is to be 
used. The effective value should always be less than or equal " +
-            "to `retention.ms` value.";
+            "Default value is -2, it represents <code>retention.ms</code> 
value is to be used. The effective value should always be less than or equal " +
+            "to <code>retention.ms</code> value.";
 
     public static final String LOCAL_LOG_RETENTION_BYTES_CONFIG = 
"local.retention.bytes";
     public static final String LOCAL_LOG_RETENTION_BYTES_DOC = "The maximum 
size of local log segments that can grow for a partition before it " +
-            "deletes the old segments. Default value is -2, it represents 
`retention.bytes` value to be used. The effective value should always be " +
-            "less than or equal to `retention.bytes` value.";
+            "deletes the old segments. Default value is -2, it represents 
<code>retention.bytes</code> value to be used. The effective value should 
always be " +
+            "less than or equal to <code>retention.bytes</code> value.";
 
     public static final String REMOTE_LOG_COPY_DISABLE_CONFIG = 
"remote.log.copy.disable";
     public static final String REMOTE_LOG_COPY_DISABLE_DOC = "Determines 
whether tiered data for a topic should become read only," +
@@ -104,7 +104,7 @@ public class TopicConfig {
     public static final String REMOTE_LOG_DELETE_ON_DISABLE_CONFIG = 
"remote.log.delete.on.disable";
     public static final String REMOTE_LOG_DELETE_ON_DISABLE_DOC = "Determines 
whether tiered data for a topic should be " +
             "deleted after tiered storage is disabled on a topic. This 
configuration should be enabled when trying to " +
-            "set `remote.storage.enable` from true to false";
+            "set <code>remote.storage.enable</code> from true to false";
 
     public static final String MAX_MESSAGE_BYTES_CONFIG = "max.message.bytes";
     public static final String MAX_MESSAGE_BYTES_DOC =
@@ -165,7 +165,7 @@ public class TopicConfig {
     public static final String UNCLEAN_LEADER_ELECTION_ENABLE_DOC = "Indicates 
whether to enable replicas " +
         "not in the ISR set to be elected as leader as a last resort, even 
though doing so may result in data " +
         "loss.<p>Note: In KRaft mode, when enabling this config dynamically, 
it needs to wait for the unclean leader election" +
-        "thread to trigger election periodically (default is 5 minutes). 
Please run `kafka-leader-election.sh` with `unclean` option " +
+        "thread to trigger election periodically (default is 5 minutes). 
Please run <code>kafka-leader-election.sh</code> with <code>unclean</code> 
option " +
          "to trigger the unclean leader election immediately if needed.</p>";
 
     public static final String MIN_IN_SYNC_REPLICAS_CONFIG = 
"min.insync.replicas";
diff --git 
a/connect/runtime/src/main/java/org/apache/kafka/connect/runtime/WorkerConfig.java
 
b/connect/runtime/src/main/java/org/apache/kafka/connect/runtime/WorkerConfig.java
index e859accb527..434ab33b99e 100644
--- 
a/connect/runtime/src/main/java/org/apache/kafka/connect/runtime/WorkerConfig.java
+++ 
b/connect/runtime/src/main/java/org/apache/kafka/connect/runtime/WorkerConfig.java
@@ -160,9 +160,9 @@ public class WorkerConfig extends AbstractConfig {
     public static final String CONNECTOR_CLIENT_POLICY_CLASS_CONFIG = 
"connector.client.config.override.policy";
     public static final String CONNECTOR_CLIENT_POLICY_CLASS_DOC =
         "Class name or alias of implementation of 
<code>ConnectorClientConfigOverridePolicy</code>. Defines what client 
configurations can be "
-        + "overridden by the connector. The default implementation is `All`, 
meaning connector configurations can override all client properties. "
-        + "The other possible policies in the framework include `None` to 
disallow connectors from overriding client properties, "
-        + "and `Principal` to allow connectors to override only client 
principals.";
+        + "overridden by the connector. The default implementation is 
<code>All</code>, meaning connector configurations can override all client 
properties. "
+        + "The other possible policies in the framework include 
<code>None</code> to disallow connectors from overriding client properties, "
+        + "and <code>Principal</code> to allow connectors to override only 
client principals.";
     public static final String CONNECTOR_CLIENT_POLICY_CLASS_DEFAULT = "All";
 
 
@@ -187,7 +187,7 @@ public class WorkerConfig extends AbstractConfig {
     public static final String TOPIC_CREATION_ENABLE_CONFIG = 
"topic.creation.enable";
     protected static final String TOPIC_CREATION_ENABLE_DOC = "Whether to 
allow "
             + "automatic creation of topics used by source connectors, when 
source connectors "
-            + "are configured with `" + TOPIC_CREATION_PREFIX + "` properties. 
Each task will use an "
+            + "are configured with <code>" + TOPIC_CREATION_PREFIX + "</code> 
properties. Each task will use an "
             + "admin client to create its topics and will not depend on the 
Kafka brokers "
             + "to create topics automatically.";
     protected static final boolean TOPIC_CREATION_ENABLE_DEFAULT = true;
diff --git 
a/connect/transforms/src/main/java/org/apache/kafka/connect/transforms/field/FieldSyntaxVersion.java
 
b/connect/transforms/src/main/java/org/apache/kafka/connect/transforms/field/FieldSyntaxVersion.java
index 514ebb425eb..4b585d663b8 100644
--- 
a/connect/transforms/src/main/java/org/apache/kafka/connect/transforms/field/FieldSyntaxVersion.java
+++ 
b/connect/transforms/src/main/java/org/apache/kafka/connect/transforms/field/FieldSyntaxVersion.java
@@ -44,13 +44,13 @@ public enum FieldSyntaxVersion {
     public static final String FIELD_SYNTAX_VERSION_CONFIG = 
"field.syntax.version";
     public static final String FIELD_SYNTAX_VERSION_DOC =
             "Defines the version of the syntax to access fields. "
-                    + "If set to `V1`, then the field paths are limited to 
access the elements at the root level of the struct or map. "
-                    + "If set to `V2`, the syntax will support accessing 
nested elements. "
+                    + "If set to <code>V1</code>, then the field paths are 
limited to access the elements at the root level of the struct or map. "
+                    + "If set to <code>V2</code>, the syntax will support 
accessing nested elements. "
                     + "To access nested elements, dotted notation is used. "
                     + "If dots are already included in the field name, "
                     + "then backtick pairs can be used to wrap field names 
containing dots. "
-                    + "E.g. to access the subfield `baz` from a field named 
\"foo.bar\" in a struct/map "
-                    + "the following format can be used to access its 
elements: \"`foo.bar`.baz\".";
+                    + "E.g. to access the subfield <code>baz</code> from a 
field named \"foo.bar\" in a struct/map "
+                    + "the following format can be used to access its 
elements: \"<code>foo.bar</code>.baz\".";
 
     public static final String FIELD_SYNTAX_VERSION_DEFAULT_VALUE = V1.name();
 
diff --git a/docs/ops.html b/docs/ops.html
index 09842ee2084..a356bb36be2 100644
--- a/docs/ops.html
+++ b/docs/ops.html
@@ -4220,7 +4220,7 @@ foo
   implemented for integration test can be used, which will create a temporary 
directory in local storage to simulate the remote storage.
 </p>
 
-<p>To adopt the `LocalTieredStorage`, the test library needs to be built 
locally</p>
+<p>To adopt the <code>LocalTieredStorage</code>, the test library needs to be 
built locally</p>
 <pre><code class="language-bash"># please checkout to the specific version tag 
you're using before building it
 # ex: `git checkout {{fullDotVersion}}`
 $ ./gradlew clean :storage:testJar</code></pre>
diff --git 
a/server/src/main/java/org/apache/kafka/server/config/ReplicationConfigs.java 
b/server/src/main/java/org/apache/kafka/server/config/ReplicationConfigs.java
index dcd46c8cf33..43453e33a8b 100644
--- 
a/server/src/main/java/org/apache/kafka/server/config/ReplicationConfigs.java
+++ 
b/server/src/main/java/org/apache/kafka/server/config/ReplicationConfigs.java
@@ -118,7 +118,7 @@ public class ReplicationConfigs {
     public static final String UNCLEAN_LEADER_ELECTION_ENABLE_CONFIG = 
ServerTopicConfigSynonyms.serverSynonym(TopicConfig.UNCLEAN_LEADER_ELECTION_ENABLE_CONFIG);
     public static final String UNCLEAN_LEADER_ELECTION_ENABLE_DOC = "Indicates 
whether to enable replicas not in the ISR set to be elected as leader as a last 
resort, even though doing so may result in data loss" +
             "<p>Note: In KRaft mode, when enabling this config dynamically, it 
needs to wait for the unclean leader election " +
-            "thread to trigger election periodically (default is 5 minutes). 
Please run `kafka-leader-election.sh` with `unclean` option " +
+            "thread to trigger election periodically (default is 5 minutes). 
Please run <code>kafka-leader-election.sh</code> with <code>unclean</code> 
option " +
             "to trigger the unclean leader election immediately if 
needed.</p>";
 
     public static final String INTER_BROKER_SECURITY_PROTOCOL_CONFIG = 
"security.inter.broker.protocol";
diff --git 
a/storage/src/main/java/org/apache/kafka/server/log/remote/storage/RemoteLogManagerConfig.java
 
b/storage/src/main/java/org/apache/kafka/server/log/remote/storage/RemoteLogManagerConfig.java
index 8e0acbb16ee..586816dae29 100644
--- 
a/storage/src/main/java/org/apache/kafka/server/log/remote/storage/RemoteLogManagerConfig.java
+++ 
b/storage/src/main/java/org/apache/kafka/server/log/remote/storage/RemoteLogManagerConfig.java
@@ -40,7 +40,7 @@ public final class RemoteLogManagerConfig {
      */
     public static final String REMOTE_STORAGE_MANAGER_CONFIG_PREFIX_PROP = 
"remote.log.storage.manager.impl.prefix";
     public static final String REMOTE_STORAGE_MANAGER_CONFIG_PREFIX_DOC = 
"Prefix used for properties to be passed to RemoteStorageManager " +
-            "implementation. For example this value can be `rsm.config.`.";
+            "implementation. For example this value can be 
<code>rsm.config.</code>.";
     public static final String DEFAULT_REMOTE_STORAGE_MANAGER_CONFIG_PREFIX = 
"rsm.config.";
 
     /**
@@ -49,7 +49,7 @@ public final class RemoteLogManagerConfig {
      */
     public static final String REMOTE_LOG_METADATA_MANAGER_CONFIG_PREFIX_PROP 
= "remote.log.metadata.manager.impl.prefix";
     public static final String REMOTE_LOG_METADATA_MANAGER_CONFIG_PREFIX_DOC = 
"Prefix used for properties to be passed to RemoteLogMetadataManager " +
-            "implementation. For example this value can be `rlmm.config.`.";
+            "implementation. For example this value can be 
<code>rlmm.config.</code>.";
     public static final String 
DEFAULT_REMOTE_LOG_METADATA_MANAGER_CONFIG_PREFIX = "rlmm.config.";
 
     public static final String REMOTE_LOG_STORAGE_SYSTEM_ENABLE_PROP = 
"remote.log.storage.system.enable";
@@ -58,20 +58,20 @@ public final class RemoteLogManagerConfig {
     public static final boolean DEFAULT_REMOTE_LOG_STORAGE_SYSTEM_ENABLE = 
false;
 
     public static final String REMOTE_STORAGE_MANAGER_CLASS_NAME_PROP = 
"remote.log.storage.manager.class.name";
-    public static final String REMOTE_STORAGE_MANAGER_CLASS_NAME_DOC = "Fully 
qualified class name of `RemoteStorageManager` implementation.";
+    public static final String REMOTE_STORAGE_MANAGER_CLASS_NAME_DOC = "Fully 
qualified class name of <code>RemoteStorageManager</code> implementation.";
 
     public static final String REMOTE_STORAGE_MANAGER_CLASS_PATH_PROP = 
"remote.log.storage.manager.class.path";
-    public static final String REMOTE_STORAGE_MANAGER_CLASS_PATH_DOC = "Class 
path of the `RemoteStorageManager` implementation. " +
+    public static final String REMOTE_STORAGE_MANAGER_CLASS_PATH_DOC = "Class 
path of the <code>RemoteStorageManager</code> implementation. " +
             "If specified, the RemoteStorageManager implementation and its 
dependent libraries will be loaded by a dedicated " +
             "classloader which searches this class path before the Kafka 
broker class path. The syntax of this parameter is same " +
             "as the standard Java class path string.";
 
     public static final String REMOTE_LOG_METADATA_MANAGER_CLASS_NAME_PROP = 
"remote.log.metadata.manager.class.name";
-    public static final String REMOTE_LOG_METADATA_MANAGER_CLASS_NAME_DOC = 
"Fully qualified class name of `RemoteLogMetadataManager` implementation.";
+    public static final String REMOTE_LOG_METADATA_MANAGER_CLASS_NAME_DOC = 
"Fully qualified class name of <code>RemoteLogMetadataManager</code> 
implementation.";
     public static final String DEFAULT_REMOTE_LOG_METADATA_MANAGER_CLASS_NAME 
= 
"org.apache.kafka.server.log.remote.metadata.storage.TopicBasedRemoteLogMetadataManager";
 
     public static final String REMOTE_LOG_METADATA_MANAGER_CLASS_PATH_PROP = 
"remote.log.metadata.manager.class.path";
-    public static final String REMOTE_LOG_METADATA_MANAGER_CLASS_PATH_DOC = 
"Class path of the `RemoteLogMetadataManager` implementation. " +
+    public static final String REMOTE_LOG_METADATA_MANAGER_CLASS_PATH_DOC = 
"Class path of the <code>RemoteLogMetadataManager</code> implementation. " +
             "If specified, the RemoteLogMetadataManager implementation and its 
dependent libraries will be loaded by a dedicated " +
             "classloader which searches this class path before the Kafka 
broker class path. The syntax of this parameter is same " +
             "as the standard Java class path string.";
@@ -149,14 +149,14 @@ public final class RemoteLogManagerConfig {
 
     public static final String LOG_LOCAL_RETENTION_MS_PROP = 
"log.local.retention.ms";
     public static final String LOG_LOCAL_RETENTION_MS_DOC = "The number of 
milliseconds to keep the local log segments before it gets eligible for 
deletion. " +
-            "Default value is -2, it represents `log.retention.ms` value is to 
be used. The effective value should always be less than or equal " +
-            "to `log.retention.ms` value.";
+            "Default value is -2, it represents <code>log.retention.ms</code> 
value is to be used. The effective value should always be less than or equal " +
+            "to <code>log.retention.ms</code> value.";
     public static final Long DEFAULT_LOG_LOCAL_RETENTION_MS = -2L;
 
     public static final String LOG_LOCAL_RETENTION_BYTES_PROP = 
"log.local.retention.bytes";
     public static final String LOG_LOCAL_RETENTION_BYTES_DOC = "The maximum 
size of local log segments that can grow for a partition before it gets 
eligible for deletion. " +
-            "Default value is -2, it represents `log.retention.bytes` value to 
be used. The effective value should always be " +
-            "less than or equal to `log.retention.bytes` value.";
+            "Default value is -2, it represents 
<code>log.retention.bytes</code> value to be used. The effective value should 
always be " +
+            "less than or equal to <code>log.retention.bytes</code> value.";
     public static final Long DEFAULT_LOG_LOCAL_RETENTION_BYTES = -2L;
 
     public static final String 
REMOTE_LOG_MANAGER_COPY_MAX_BYTES_PER_SECOND_PROP = 
"remote.log.manager.copy.max.bytes.per.second";
diff --git a/streams/src/main/java/org/apache/kafka/streams/StreamsConfig.java 
b/streams/src/main/java/org/apache/kafka/streams/StreamsConfig.java
index 1043e5fb6f0..5e1eaff1162 100644
--- a/streams/src/main/java/org/apache/kafka/streams/StreamsConfig.java
+++ b/streams/src/main/java/org/apache/kafka/streams/StreamsConfig.java
@@ -647,9 +647,9 @@ public class StreamsConfig extends AbstractConfig {
                                                               " the task 
available on one instance while it is warming up on another instance it has 
been reassigned to. Used to throttle how much extra broker " +
                                                               " traffic and 
cluster state can be used for high availability. Must be at least 1." +
                                                               "Note that one 
warmup replica corresponds to one Stream Task. Furthermore, note that each 
warmup replica can only be promoted to an active task " +
-                                                              "during a 
rebalance (normally during a so-called probing rebalance, which occur at a 
frequency specified by the `probing.rebalance.interval.ms` config). This means 
" +
+                                                              "during a 
rebalance (normally during a so-called probing rebalance, which occur at a 
frequency specified by the <code>probing.rebalance.interval.ms</code> config). 
This means " +
                                                               "that the 
maximum rate at which active tasks can be migrated from one Kafka Streams 
Instance to another instance can be determined by " +
-                                                              
"(`max.warmup.replicas` / `probing.rebalance.interval.ms`).";
+                                                              
"(<code>max.warmup.replicas</code> / 
<code>probing.rebalance.interval.ms</code>).";
 
     /** {@code metadata.max.age.ms} */
     @SuppressWarnings("WeakerAccess")
@@ -821,7 +821,7 @@ public class StreamsConfig extends AbstractConfig {
     /** {@code upgrade.from} */
     @SuppressWarnings("WeakerAccess")
     public static final String UPGRADE_FROM_CONFIG = "upgrade.from";
-    private static final String UPGRADE_FROM_DOC = "Allows live upgrading (and 
downgrading in some cases -- see upgrade guide) in a backward compatible way. 
Default is `null`. " +
+    private static final String UPGRADE_FROM_DOC = "Allows live upgrading (and 
downgrading in some cases -- see upgrade guide) in a backward compatible way. 
Default is <code>null</code>. " +
         "Please refer to the Kafka Streams upgrade guide for instructions on 
how and when to use this config. " +
         "Note that when upgrading from 3.5 to a newer version it is never 
required to specify this config, " +
         "while upgrading live directly to 4.0+ from 2.3 or below is no longer 
supported even with this config. " +

Reply via email to