This is an automated email from the ASF dual-hosted git repository.

acosentino pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/camel.git

commit a2851212679af28d81f91024cb00b0defe0fe01f
Author: Andrea Cosentino <[email protected]>
AuthorDate: Tue Jul 11 09:57:16 2023 +0200

    Regen
    
    Signed-off-by: Andrea Cosentino <[email protected]>
---
 ...neCacheProducerMultiOperationSameCacheTest.java |  3 -
 ...adCacheProducerMultiOperationSameCacheTest.java |  3 -
 .../camel/component/kafka/KafkaComponent.java      | 10 ++--
 .../camel/component/kafka/KafkaConfiguration.java  | 26 ++++-----
 .../others/examples/json/platform-http-main.json   |  1 +
 docs/components/modules/others/nav.adoc            |  1 +
 .../modules/others/pages/platform-http-main.adoc   |  1 +
 .../dsl/KafkaComponentBuilderFactory.java          | 38 ++++++-------
 .../endpoint/dsl/KafkaEndpointBuilderFactory.java  | 64 +++++++++++-----------
 9 files changed, 72 insertions(+), 75 deletions(-)

diff --git 
a/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/cache/CaffeineCacheProducerMultiOperationSameCacheTest.java
 
b/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/cache/CaffeineCacheProducerMultiOperationSameCacheTest.java
index abcebc693ab..fec5aeefcc1 100644
--- 
a/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/cache/CaffeineCacheProducerMultiOperationSameCacheTest.java
+++ 
b/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/cache/CaffeineCacheProducerMultiOperationSameCacheTest.java
@@ -16,9 +16,6 @@
  */
 package org.apache.camel.component.caffeine.cache;
 
-import java.util.HashMap;
-import java.util.Map;
-
 import org.apache.camel.builder.RouteBuilder;
 import org.apache.camel.component.caffeine.CaffeineConstants;
 import org.apache.camel.component.mock.MockEndpoint;
diff --git 
a/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/loadcache/CaffeineLoadCacheProducerMultiOperationSameCacheTest.java
 
b/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/loadcache/CaffeineLoadCacheProducerMultiOperationSameCacheTest.java
index 68eec560449..3a955336dc2 100644
--- 
a/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/loadcache/CaffeineLoadCacheProducerMultiOperationSameCacheTest.java
+++ 
b/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/loadcache/CaffeineLoadCacheProducerMultiOperationSameCacheTest.java
@@ -16,9 +16,6 @@
  */
 package org.apache.camel.component.caffeine.loadcache;
 
-import java.util.HashMap;
-import java.util.Map;
-
 import org.apache.camel.builder.RouteBuilder;
 import org.apache.camel.component.caffeine.CaffeineConstants;
 import org.apache.camel.component.mock.MockEndpoint;
diff --git 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaComponent.java
 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaComponent.java
index 2f2618939f7..7b8b5322a2e 100644
--- 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaComponent.java
+++ 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaComponent.java
@@ -175,8 +175,8 @@ public class KafkaComponent extends DefaultComponent 
implements SSLContextParame
      * invalid kafka configurations. Unfortunately, kafka-client does not 
separate this kind of errors.
      *
      * Camel will by default retry forever, and therefore never give up. If 
you want to give up after many attempts then
-     * set this option and Camel will then when giving up terminate the 
consumer. To try again, you can manually restart the consumer
-     * by stopping, and starting the route.
+     * set this option and Camel will then when giving up terminate the 
consumer. To try again, you can manually restart
+     * the consumer by stopping, and starting the route.
      */
     public void setCreateConsumerBackoffMaxAttempts(int 
createConsumerBackoffMaxAttempts) {
         this.createConsumerBackoffMaxAttempts = 
createConsumerBackoffMaxAttempts;
@@ -204,9 +204,9 @@ public class KafkaComponent extends DefaultComponent 
implements SSLContextParame
      * Error during subscribing the consumer to the kafka topic could be 
temporary errors due to network issues, and
      * could potentially be recoverable.
      *
-     * Camel will by default retry forever, and therefore never give up. If 
you want to give up after many attempts, then
-     * set this option and Camel will then when giving up terminate the 
consumer. You can manually restart the consumer
-     * by stopping and starting the route, to try again.
+     * Camel will by default retry forever, and therefore never give up. If 
you want to give up after many attempts,
+     * then set this option and Camel will then when giving up terminate the 
consumer. You can manually restart the
+     * consumer by stopping and starting the route, to try again.
      */
     public void setSubscribeConsumerBackoffMaxAttempts(int 
subscribeConsumerBackoffMaxAttempts) {
         this.subscribeConsumerBackoffMaxAttempts = 
subscribeConsumerBackoffMaxAttempts;
diff --git 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaConfiguration.java
 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaConfiguration.java
index a695f2028fb..06aad479337 100755
--- 
a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaConfiguration.java
+++ 
b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaConfiguration.java
@@ -824,8 +824,8 @@ public class KafkaConfiguration implements Cloneable, 
HeaderFilterStrategyAware
      * <tt>false</tt> then the consumer continues to the next message and 
processes it. If the option is <tt>true</tt>
      * then the consumer breaks out, and will seek back to offset of the 
message that caused a failure, and then
      * re-attempt to process this message. However this can lead to endless 
processing of the same message if its bound
-     * to fail every time, eg a poison message. Therefore it is recommended to 
deal with that for example by using Camel's
-     * error handler.
+     * to fail every time, eg a poison message. Therefore it is recommended to 
deal with that for example by using
+     * Camel's error handler.
      */
     public void setBreakOnFirstError(boolean breakOnFirstError) {
         this.breakOnFirstError = breakOnFirstError;
@@ -889,8 +889,8 @@ public class KafkaConfiguration implements Cloneable, 
HeaderFilterStrategyAware
 
     /**
      * Before each retry, the producer refreshes the metadata of relevant 
topics to see if a new leader has been
-     * elected. Since the leader election takes a bit of time, this property 
specifies the amount of time that the producer
-     * waits before refreshing the metadata.
+     * elected. Since the leader election takes a bit of time, this property 
specifies the amount of time that the
+     * producer waits before refreshing the metadata.
      */
     public void setRetryBackoffMs(Integer retryBackoffMs) {
         this.retryBackoffMs = retryBackoffMs;
@@ -1220,8 +1220,8 @@ public class KafkaConfiguration implements Cloneable, 
HeaderFilterStrategyAware
     }
 
     /**
-     * The location of the key store file. This is optional for the client and 
can be used for two-way authentication for
-     * the client.
+     * The location of the key store file. This is optional for the client and 
can be used for two-way authentication
+     * for the client.
      */
     public void setSslKeystoreLocation(String sslKeystoreLocation) {
         this.sslKeystoreLocation = sslKeystoreLocation;
@@ -1232,8 +1232,8 @@ public class KafkaConfiguration implements Cloneable, 
HeaderFilterStrategyAware
     }
 
     /**
-     * The store password for the key store file. This is optional for the 
client and only needed if sslKeystoreLocation' is
-     * configured. Key store password is not supported for PEM format.
+     * The store password for the key store file. This is optional for the 
client and only needed if
+     * sslKeystoreLocation' is configured. Key store password is not supported 
for PEM format.
      */
     public void setSslKeystorePassword(String sslKeystorePassword) {
         this.sslKeystorePassword = sslKeystorePassword;
@@ -1399,9 +1399,9 @@ public class KafkaConfiguration implements Cloneable, 
HeaderFilterStrategyAware
      * request. Normally this occurs only under load when records arrive 
faster than they can be sent out. However in
      * some circumstances the client may want to reduce the number of requests 
even under moderate load. This setting
      * accomplishes this by adding a small amount of artificial delay that is, 
rather than immediately sending out a
-     * record the producer will wait for up to the given delay to allow other 
records to be sent so that they can
-     * be batched together. This can be thought of as analogous to Nagle's 
algorithm in TCP. This setting gives the
-     * upper bound on the delay for batching: once we get batch.size worth of 
records for a partition it will be sent
+     * record the producer will wait for up to the given delay to allow other 
records to be sent so that they can be
+     * batched together. This can be thought of as analogous to Nagle's 
algorithm in TCP. This setting gives the upper
+     * bound on the delay for batching: once we get batch.size worth of 
records for a partition it will be sent
      * immediately regardless of this setting, however if we have fewer than 
this many bytes accumulated for this
      * partition we will 'linger' for the specified time waiting for more 
records to show up. This setting defaults to 0
      * (i.e. no delay). Setting linger.ms=5, for example, would have the 
effect of reducing the number of requests sent
@@ -1663,8 +1663,8 @@ public class KafkaConfiguration implements Cloneable, 
HeaderFilterStrategyAware
     }
 
     /**
-     * Set if KafkaConsumer will read from the beginning or the end on 
startup: SeekPolicy.BEGINNING: read from the beginning.
-     * SeekPolicy.END: read from the end.
+     * Set if KafkaConsumer will read from the beginning or the end on 
startup: SeekPolicy.BEGINNING: read from the
+     * beginning. SeekPolicy.END: read from the end.
      */
     public void setSeekTo(SeekPolicy seekTo) {
         this.seekTo = seekTo;
diff --git 
a/docs/components/modules/others/examples/json/platform-http-main.json 
b/docs/components/modules/others/examples/json/platform-http-main.json
new file mode 120000
index 00000000000..cabf7d49547
--- /dev/null
+++ b/docs/components/modules/others/examples/json/platform-http-main.json
@@ -0,0 +1 @@
+../../../../../../components/camel-platform-http-main/src/generated/resources/platform-http-main.json
\ No newline at end of file
diff --git a/docs/components/modules/others/nav.adoc 
b/docs/components/modules/others/nav.adoc
index 6f03eacacad..fd975c765b7 100644
--- a/docs/components/modules/others/nav.adoc
+++ b/docs/components/modules/others/nav.adoc
@@ -36,6 +36,7 @@
 ** xref:microprofile-health.adoc[Microprofile Health]
 ** xref:openapi-java.adoc[Openapi Java]
 ** xref:opentelemetry.adoc[OpenTelemetry]
+** xref:platform-http-main.adoc[Platform Http Main]
 ** xref:platform-http-vertx.adoc[Platform Http Vertx]
 ** xref:reactive-executor-tomcat.adoc[Reactive Executor Tomcat]
 ** xref:reactive-executor-vertx.adoc[Reactive Executor Vert.x]
diff --git a/docs/components/modules/others/pages/platform-http-main.adoc 
b/docs/components/modules/others/pages/platform-http-main.adoc
new file mode 120000
index 00000000000..2ac31070d90
--- /dev/null
+++ b/docs/components/modules/others/pages/platform-http-main.adoc
@@ -0,0 +1 @@
+../../../../../components/camel-platform-http-main/src/main/docs/platform-http-main.adoc
\ No newline at end of file
diff --git 
a/dsl/camel-componentdsl/src/generated/java/org/apache/camel/builder/component/dsl/KafkaComponentBuilderFactory.java
 
b/dsl/camel-componentdsl/src/generated/java/org/apache/camel/builder/component/dsl/KafkaComponentBuilderFactory.java
index feaa422d09a..d919a48dfea 100644
--- 
a/dsl/camel-componentdsl/src/generated/java/org/apache/camel/builder/component/dsl/KafkaComponentBuilderFactory.java
+++ 
b/dsl/camel-componentdsl/src/generated/java/org/apache/camel/builder/component/dsl/KafkaComponentBuilderFactory.java
@@ -254,7 +254,7 @@ public interface KafkaComponentBuilderFactory {
          * message that caused a failure, and then re-attempt to process this
          * message. However this can lead to endless processing of the same
          * message if its bound to fail every time, eg a poison message.
-         * Therefore its recommended to deal with that for example by using
+         * Therefore it is recommended to deal with that for example by using
          * Camel's error handler.
          * 
          * The option is a: &lt;code&gt;boolean&lt;/code&gt; type.
@@ -501,7 +501,7 @@ public interface KafkaComponentBuilderFactory {
             return this;
         }
         /**
-         * Deserializer class for key that implements the Deserializer
+         * Deserializer class for the key that implements the Deserializer
          * interface.
          * 
          * The option is a: &lt;code&gt;java.lang.String&lt;/code&gt; type.
@@ -654,9 +654,9 @@ public interface KafkaComponentBuilderFactory {
             return this;
         }
         /**
-         * Set if KafkaConsumer will read from beginning or end on startup:
-         * SeekPolicy.BEGINNING: read from beginning. SeekPolicy.END: read from
-         * end.
+         * Set if KafkaConsumer will read from the beginning or the end on
+         * startup: SeekPolicy.BEGINNING: read from the beginning.
+         * SeekPolicy.END: read from the end.
          * 
          * The option is a:
          * &lt;code&gt;org.apache.camel.component.kafka.SeekPolicy&lt;/code&gt;
@@ -765,12 +765,12 @@ public interface KafkaComponentBuilderFactory {
          * possible. However, one part of the validation is DNS resolution of
          * the bootstrap broker hostnames. This may be a temporary networking
          * problem, and could potentially be recoverable. While other errors 
are
-         * fatal such as some invalid kafka configurations. Unfortunately
+         * fatal, such as some invalid kafka configurations. Unfortunately,
          * kafka-client does not separate this kind of errors. Camel will by
          * default retry forever, and therefore never give up. If you want to
          * give up after many attempts then set this option and Camel will then
-         * when giving up terminate the consumer. You can manually restart the
-         * consumer by stopping and starting the route, to try again.
+         * when giving up terminate the consumer. To try again, you can 
manually
+         * restart the consumer by stopping, and starting the route.
          * 
          * The option is a: &lt;code&gt;int&lt;/code&gt; type.
          * 
@@ -874,7 +874,7 @@ public interface KafkaComponentBuilderFactory {
          * subscribing the consumer to the kafka topic could be temporary 
errors
          * due to network issues, and could potentially be recoverable. Camel
          * will by default retry forever, and therefore never give up. If you
-         * want to give up after many attempts then set this option and Camel
+         * want to give up after many attempts, then set this option and Camel
          * will then when giving up terminate the consumer. You can manually
          * restart the consumer by stopping and starting the route, to try
          * again.
@@ -1100,7 +1100,7 @@ public interface KafkaComponentBuilderFactory {
          * accomplishes this by adding a small amount of artificial delay that
          * is, rather than immediately sending out a record the producer will
          * wait for up to the given delay to allow other records to be sent so
-         * that the sends can be batched together. This can be thought of as
+         * that they can be batched together. This can be thought of as
          * analogous to Nagle's algorithm in TCP. This setting gives the upper
          * bound on the delay for batching: once we get batch.size worth of
          * records for a partition it will be sent immediately regardless of
@@ -1132,7 +1132,7 @@ public interface KafkaComponentBuilderFactory {
          * serializers or partitioner is not counted against this timeout). For
          * partitionsFor() this timeout bounds the time spent waiting for
          * metadata if it is unavailable. The transaction-related methods 
always
-         * block, but may timeout if the transaction coordinator could not be
+         * block, but may time out if the transaction coordinator could not be
          * discovered or did not respond within the timeout.
          * 
          * The option is a: &lt;code&gt;java.lang.Integer&lt;/code&gt; type.
@@ -1293,9 +1293,9 @@ public interface KafkaComponentBuilderFactory {
          * requests whenever multiple records are being sent to the same
          * partition. This helps performance on both the client and the server.
          * This configuration controls the default batch size in bytes. No
-         * attempt will be made to batch records larger than this size.Requests
+         * attempt will be made to batch records larger than this size. 
Requests
          * sent to brokers will contain multiple batches, one for each 
partition
-         * with data available to be sent.A small batch size will make batching
+         * with data available to be sent. A small batch size will make 
batching
          * less common and may reduce throughput (a batch size of zero will
          * disable batching entirely). A very large batch size may use memory a
          * bit more wastefully as we will always allocate a buffer of the
@@ -1472,9 +1472,9 @@ public interface KafkaComponentBuilderFactory {
         }
         /**
          * Before each retry, the producer refreshes the metadata of relevant
-         * topics to see if a new leader has been elected. Since leader 
election
-         * takes a bit of time, this property specifies the amount of time that
-         * the producer waits before refreshing the metadata.
+         * topics to see if a new leader has been elected. Since the leader
+         * election takes a bit of time, this property specifies the amount of
+         * time that the producer waits before refreshing the metadata.
          * 
          * The option is a: &lt;code&gt;java.lang.Integer&lt;/code&gt; type.
          * 
@@ -1958,8 +1958,8 @@ public interface KafkaComponentBuilderFactory {
             return this;
         }
         /**
-         * The location of the key store file. This is optional for client and
-         * can be used for two-way authentication for client.
+         * The location of the key store file. This is optional for the client
+         * and can be used for two-way authentication for the client.
          * 
          * The option is a: &lt;code&gt;java.lang.String&lt;/code&gt; type.
          * 
@@ -1974,7 +1974,7 @@ public interface KafkaComponentBuilderFactory {
             return this;
         }
         /**
-         * The store password for the key store file. This is optional for
+         * The store password for the key store file. This is optional for the
          * client and only needed if sslKeystoreLocation' is configured. Key
          * store password is not supported for PEM format.
          * 
diff --git 
a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/KafkaEndpointBuilderFactory.java
 
b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/KafkaEndpointBuilderFactory.java
index 02ec4c69b27..2c777640af8 100644
--- 
a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/KafkaEndpointBuilderFactory.java
+++ 
b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/KafkaEndpointBuilderFactory.java
@@ -376,7 +376,7 @@ public interface KafkaEndpointBuilderFactory {
          * message that caused a failure, and then re-attempt to process this
          * message. However this can lead to endless processing of the same
          * message if its bound to fail every time, eg a poison message.
-         * Therefore its recommended to deal with that for example by using
+         * Therefore it is recommended to deal with that for example by using
          * Camel's error handler.
          * 
          * The option is a: &lt;code&gt;boolean&lt;/code&gt; type.
@@ -400,7 +400,7 @@ public interface KafkaEndpointBuilderFactory {
          * message that caused a failure, and then re-attempt to process this
          * message. However this can lead to endless processing of the same
          * message if its bound to fail every time, eg a poison message.
-         * Therefore its recommended to deal with that for example by using
+         * Therefore it is recommended to deal with that for example by using
          * Camel's error handler.
          * 
          * The option will be converted to a &lt;code&gt;boolean&lt;/code&gt;
@@ -799,7 +799,7 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * Deserializer class for key that implements the Deserializer
+         * Deserializer class for the key that implements the Deserializer
          * interface.
          * 
          * The option is a: &lt;code&gt;java.lang.String&lt;/code&gt; type.
@@ -1073,9 +1073,9 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * Set if KafkaConsumer will read from beginning or end on startup:
-         * SeekPolicy.BEGINNING: read from beginning. SeekPolicy.END: read from
-         * end.
+         * Set if KafkaConsumer will read from the beginning or the end on
+         * startup: SeekPolicy.BEGINNING: read from the beginning.
+         * SeekPolicy.END: read from the end.
          * 
          * The option is a:
          * &lt;code&gt;org.apache.camel.component.kafka.SeekPolicy&lt;/code&gt;
@@ -1092,9 +1092,9 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * Set if KafkaConsumer will read from beginning or end on startup:
-         * SeekPolicy.BEGINNING: read from beginning. SeekPolicy.END: read from
-         * end.
+         * Set if KafkaConsumer will read from the beginning or the end on
+         * startup: SeekPolicy.BEGINNING: read from the beginning.
+         * SeekPolicy.END: read from the end.
          * 
          * The option will be converted to a
          * &lt;code&gt;org.apache.camel.component.kafka.SeekPolicy&lt;/code&gt;
@@ -1633,8 +1633,8 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * The location of the key store file. This is optional for client and
-         * can be used for two-way authentication for client.
+         * The location of the key store file. This is optional for the client
+         * and can be used for two-way authentication for the client.
          * 
          * The option is a: &lt;code&gt;java.lang.String&lt;/code&gt; type.
          * 
@@ -1649,7 +1649,7 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * The store password for the key store file. This is optional for
+         * The store password for the key store file. This is optional for the
          * client and only needed if sslKeystoreLocation' is configured. Key
          * store password is not supported for PEM format.
          * 
@@ -2573,7 +2573,7 @@ public interface KafkaEndpointBuilderFactory {
          * accomplishes this by adding a small amount of artificial delay that
          * is, rather than immediately sending out a record the producer will
          * wait for up to the given delay to allow other records to be sent so
-         * that the sends can be batched together. This can be thought of as
+         * that they can be batched together. This can be thought of as
          * analogous to Nagle's algorithm in TCP. This setting gives the upper
          * bound on the delay for batching: once we get batch.size worth of
          * records for a partition it will be sent immediately regardless of
@@ -2605,7 +2605,7 @@ public interface KafkaEndpointBuilderFactory {
          * accomplishes this by adding a small amount of artificial delay that
          * is, rather than immediately sending out a record the producer will
          * wait for up to the given delay to allow other records to be sent so
-         * that the sends can be batched together. This can be thought of as
+         * that they can be batched together. This can be thought of as
          * analogous to Nagle's algorithm in TCP. This setting gives the upper
          * bound on the delay for batching: once we get batch.size worth of
          * records for a partition it will be sent immediately regardless of
@@ -2638,7 +2638,7 @@ public interface KafkaEndpointBuilderFactory {
          * serializers or partitioner is not counted against this timeout). For
          * partitionsFor() this timeout bounds the time spent waiting for
          * metadata if it is unavailable. The transaction-related methods 
always
-         * block, but may timeout if the transaction coordinator could not be
+         * block, but may time out if the transaction coordinator could not be
          * discovered or did not respond within the timeout.
          * 
          * The option is a: &lt;code&gt;java.lang.Integer&lt;/code&gt; type.
@@ -2662,7 +2662,7 @@ public interface KafkaEndpointBuilderFactory {
          * serializers or partitioner is not counted against this timeout). For
          * partitionsFor() this timeout bounds the time spent waiting for
          * metadata if it is unavailable. The transaction-related methods 
always
-         * block, but may timeout if the transaction coordinator could not be
+         * block, but may time out if the transaction coordinator could not be
          * discovered or did not respond within the timeout.
          * 
          * The option will be converted to a
@@ -2934,9 +2934,9 @@ public interface KafkaEndpointBuilderFactory {
          * requests whenever multiple records are being sent to the same
          * partition. This helps performance on both the client and the server.
          * This configuration controls the default batch size in bytes. No
-         * attempt will be made to batch records larger than this size.Requests
+         * attempt will be made to batch records larger than this size. 
Requests
          * sent to brokers will contain multiple batches, one for each 
partition
-         * with data available to be sent.A small batch size will make batching
+         * with data available to be sent. A small batch size will make 
batching
          * less common and may reduce throughput (a batch size of zero will
          * disable batching entirely). A very large batch size may use memory a
          * bit more wastefully as we will always allocate a buffer of the
@@ -2960,9 +2960,9 @@ public interface KafkaEndpointBuilderFactory {
          * requests whenever multiple records are being sent to the same
          * partition. This helps performance on both the client and the server.
          * This configuration controls the default batch size in bytes. No
-         * attempt will be made to batch records larger than this size.Requests
+         * attempt will be made to batch records larger than this size. 
Requests
          * sent to brokers will contain multiple batches, one for each 
partition
-         * with data available to be sent.A small batch size will make batching
+         * with data available to be sent. A small batch size will make 
batching
          * less common and may reduce throughput (a batch size of zero will
          * disable batching entirely). A very large batch size may use memory a
          * bit more wastefully as we will always allocate a buffer of the
@@ -3267,9 +3267,9 @@ public interface KafkaEndpointBuilderFactory {
         }
         /**
          * Before each retry, the producer refreshes the metadata of relevant
-         * topics to see if a new leader has been elected. Since leader 
election
-         * takes a bit of time, this property specifies the amount of time that
-         * the producer waits before refreshing the metadata.
+         * topics to see if a new leader has been elected. Since the leader
+         * election takes a bit of time, this property specifies the amount of
+         * time that the producer waits before refreshing the metadata.
          * 
          * The option is a: &lt;code&gt;java.lang.Integer&lt;/code&gt; type.
          * 
@@ -3286,9 +3286,9 @@ public interface KafkaEndpointBuilderFactory {
         }
         /**
          * Before each retry, the producer refreshes the metadata of relevant
-         * topics to see if a new leader has been elected. Since leader 
election
-         * takes a bit of time, this property specifies the amount of time that
-         * the producer waits before refreshing the metadata.
+         * topics to see if a new leader has been elected. Since the leader
+         * election takes a bit of time, this property specifies the amount of
+         * time that the producer waits before refreshing the metadata.
          * 
          * The option will be converted to a
          * &lt;code&gt;java.lang.Integer&lt;/code&gt; type.
@@ -3864,8 +3864,8 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * The location of the key store file. This is optional for client and
-         * can be used for two-way authentication for client.
+         * The location of the key store file. This is optional for the client
+         * and can be used for two-way authentication for the client.
          * 
          * The option is a: &lt;code&gt;java.lang.String&lt;/code&gt; type.
          * 
@@ -3880,7 +3880,7 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * The store password for the key store file. This is optional for
+         * The store password for the key store file. This is optional for the
          * client and only needed if sslKeystoreLocation' is configured. Key
          * store password is not supported for PEM format.
          * 
@@ -4746,8 +4746,8 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * The location of the key store file. This is optional for client and
-         * can be used for two-way authentication for client.
+         * The location of the key store file. This is optional for the client
+         * and can be used for two-way authentication for the client.
          * 
          * The option is a: &lt;code&gt;java.lang.String&lt;/code&gt; type.
          * 
@@ -4762,7 +4762,7 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * The store password for the key store file. This is optional for
+         * The store password for the key store file. This is optional for the
          * client and only needed if sslKeystoreLocation' is configured. Key
          * store password is not supported for PEM format.
          * 

Reply via email to