This is an automated email from the ASF dual-hosted git repository. acosentino pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/camel.git
commit a2851212679af28d81f91024cb00b0defe0fe01f Author: Andrea Cosentino <[email protected]> AuthorDate: Tue Jul 11 09:57:16 2023 +0200 Regen Signed-off-by: Andrea Cosentino <[email protected]> --- ...neCacheProducerMultiOperationSameCacheTest.java | 3 - ...adCacheProducerMultiOperationSameCacheTest.java | 3 - .../camel/component/kafka/KafkaComponent.java | 10 ++-- .../camel/component/kafka/KafkaConfiguration.java | 26 ++++----- .../others/examples/json/platform-http-main.json | 1 + docs/components/modules/others/nav.adoc | 1 + .../modules/others/pages/platform-http-main.adoc | 1 + .../dsl/KafkaComponentBuilderFactory.java | 38 ++++++------- .../endpoint/dsl/KafkaEndpointBuilderFactory.java | 64 +++++++++++----------- 9 files changed, 72 insertions(+), 75 deletions(-) diff --git a/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/cache/CaffeineCacheProducerMultiOperationSameCacheTest.java b/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/cache/CaffeineCacheProducerMultiOperationSameCacheTest.java index abcebc693ab..fec5aeefcc1 100644 --- a/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/cache/CaffeineCacheProducerMultiOperationSameCacheTest.java +++ b/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/cache/CaffeineCacheProducerMultiOperationSameCacheTest.java @@ -16,9 +16,6 @@ */ package org.apache.camel.component.caffeine.cache; -import java.util.HashMap; -import java.util.Map; - import org.apache.camel.builder.RouteBuilder; import org.apache.camel.component.caffeine.CaffeineConstants; import org.apache.camel.component.mock.MockEndpoint; diff --git a/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/loadcache/CaffeineLoadCacheProducerMultiOperationSameCacheTest.java b/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/loadcache/CaffeineLoadCacheProducerMultiOperationSameCacheTest.java index 68eec560449..3a955336dc2 100644 --- a/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/loadcache/CaffeineLoadCacheProducerMultiOperationSameCacheTest.java +++ b/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/loadcache/CaffeineLoadCacheProducerMultiOperationSameCacheTest.java @@ -16,9 +16,6 @@ */ package org.apache.camel.component.caffeine.loadcache; -import java.util.HashMap; -import java.util.Map; - import org.apache.camel.builder.RouteBuilder; import org.apache.camel.component.caffeine.CaffeineConstants; import org.apache.camel.component.mock.MockEndpoint; diff --git a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaComponent.java b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaComponent.java index 2f2618939f7..7b8b5322a2e 100644 --- a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaComponent.java +++ b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaComponent.java @@ -175,8 +175,8 @@ public class KafkaComponent extends DefaultComponent implements SSLContextParame * invalid kafka configurations. Unfortunately, kafka-client does not separate this kind of errors. * * Camel will by default retry forever, and therefore never give up. If you want to give up after many attempts then - * set this option and Camel will then when giving up terminate the consumer. To try again, you can manually restart the consumer - * by stopping, and starting the route. + * set this option and Camel will then when giving up terminate the consumer. To try again, you can manually restart + * the consumer by stopping, and starting the route. */ public void setCreateConsumerBackoffMaxAttempts(int createConsumerBackoffMaxAttempts) { this.createConsumerBackoffMaxAttempts = createConsumerBackoffMaxAttempts; @@ -204,9 +204,9 @@ public class KafkaComponent extends DefaultComponent implements SSLContextParame * Error during subscribing the consumer to the kafka topic could be temporary errors due to network issues, and * could potentially be recoverable. * - * Camel will by default retry forever, and therefore never give up. If you want to give up after many attempts, then - * set this option and Camel will then when giving up terminate the consumer. You can manually restart the consumer - * by stopping and starting the route, to try again. + * Camel will by default retry forever, and therefore never give up. If you want to give up after many attempts, + * then set this option and Camel will then when giving up terminate the consumer. You can manually restart the + * consumer by stopping and starting the route, to try again. */ public void setSubscribeConsumerBackoffMaxAttempts(int subscribeConsumerBackoffMaxAttempts) { this.subscribeConsumerBackoffMaxAttempts = subscribeConsumerBackoffMaxAttempts; diff --git a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaConfiguration.java b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaConfiguration.java index a695f2028fb..06aad479337 100755 --- a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaConfiguration.java +++ b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaConfiguration.java @@ -824,8 +824,8 @@ public class KafkaConfiguration implements Cloneable, HeaderFilterStrategyAware * <tt>false</tt> then the consumer continues to the next message and processes it. If the option is <tt>true</tt> * then the consumer breaks out, and will seek back to offset of the message that caused a failure, and then * re-attempt to process this message. However this can lead to endless processing of the same message if its bound - * to fail every time, eg a poison message. Therefore it is recommended to deal with that for example by using Camel's - * error handler. + * to fail every time, eg a poison message. Therefore it is recommended to deal with that for example by using + * Camel's error handler. */ public void setBreakOnFirstError(boolean breakOnFirstError) { this.breakOnFirstError = breakOnFirstError; @@ -889,8 +889,8 @@ public class KafkaConfiguration implements Cloneable, HeaderFilterStrategyAware /** * Before each retry, the producer refreshes the metadata of relevant topics to see if a new leader has been - * elected. Since the leader election takes a bit of time, this property specifies the amount of time that the producer - * waits before refreshing the metadata. + * elected. Since the leader election takes a bit of time, this property specifies the amount of time that the + * producer waits before refreshing the metadata. */ public void setRetryBackoffMs(Integer retryBackoffMs) { this.retryBackoffMs = retryBackoffMs; @@ -1220,8 +1220,8 @@ public class KafkaConfiguration implements Cloneable, HeaderFilterStrategyAware } /** - * The location of the key store file. This is optional for the client and can be used for two-way authentication for - * the client. + * The location of the key store file. This is optional for the client and can be used for two-way authentication + * for the client. */ public void setSslKeystoreLocation(String sslKeystoreLocation) { this.sslKeystoreLocation = sslKeystoreLocation; @@ -1232,8 +1232,8 @@ public class KafkaConfiguration implements Cloneable, HeaderFilterStrategyAware } /** - * The store password for the key store file. This is optional for the client and only needed if sslKeystoreLocation' is - * configured. Key store password is not supported for PEM format. + * The store password for the key store file. This is optional for the client and only needed if + * sslKeystoreLocation' is configured. Key store password is not supported for PEM format. */ public void setSslKeystorePassword(String sslKeystorePassword) { this.sslKeystorePassword = sslKeystorePassword; @@ -1399,9 +1399,9 @@ public class KafkaConfiguration implements Cloneable, HeaderFilterStrategyAware * request. Normally this occurs only under load when records arrive faster than they can be sent out. However in * some circumstances the client may want to reduce the number of requests even under moderate load. This setting * accomplishes this by adding a small amount of artificial delay that is, rather than immediately sending out a - * record the producer will wait for up to the given delay to allow other records to be sent so that they can - * be batched together. This can be thought of as analogous to Nagle's algorithm in TCP. This setting gives the - * upper bound on the delay for batching: once we get batch.size worth of records for a partition it will be sent + * record the producer will wait for up to the given delay to allow other records to be sent so that they can be + * batched together. This can be thought of as analogous to Nagle's algorithm in TCP. This setting gives the upper + * bound on the delay for batching: once we get batch.size worth of records for a partition it will be sent * immediately regardless of this setting, however if we have fewer than this many bytes accumulated for this * partition we will 'linger' for the specified time waiting for more records to show up. This setting defaults to 0 * (i.e. no delay). Setting linger.ms=5, for example, would have the effect of reducing the number of requests sent @@ -1663,8 +1663,8 @@ public class KafkaConfiguration implements Cloneable, HeaderFilterStrategyAware } /** - * Set if KafkaConsumer will read from the beginning or the end on startup: SeekPolicy.BEGINNING: read from the beginning. - * SeekPolicy.END: read from the end. + * Set if KafkaConsumer will read from the beginning or the end on startup: SeekPolicy.BEGINNING: read from the + * beginning. SeekPolicy.END: read from the end. */ public void setSeekTo(SeekPolicy seekTo) { this.seekTo = seekTo; diff --git a/docs/components/modules/others/examples/json/platform-http-main.json b/docs/components/modules/others/examples/json/platform-http-main.json new file mode 120000 index 00000000000..cabf7d49547 --- /dev/null +++ b/docs/components/modules/others/examples/json/platform-http-main.json @@ -0,0 +1 @@ +../../../../../../components/camel-platform-http-main/src/generated/resources/platform-http-main.json \ No newline at end of file diff --git a/docs/components/modules/others/nav.adoc b/docs/components/modules/others/nav.adoc index 6f03eacacad..fd975c765b7 100644 --- a/docs/components/modules/others/nav.adoc +++ b/docs/components/modules/others/nav.adoc @@ -36,6 +36,7 @@ ** xref:microprofile-health.adoc[Microprofile Health] ** xref:openapi-java.adoc[Openapi Java] ** xref:opentelemetry.adoc[OpenTelemetry] +** xref:platform-http-main.adoc[Platform Http Main] ** xref:platform-http-vertx.adoc[Platform Http Vertx] ** xref:reactive-executor-tomcat.adoc[Reactive Executor Tomcat] ** xref:reactive-executor-vertx.adoc[Reactive Executor Vert.x] diff --git a/docs/components/modules/others/pages/platform-http-main.adoc b/docs/components/modules/others/pages/platform-http-main.adoc new file mode 120000 index 00000000000..2ac31070d90 --- /dev/null +++ b/docs/components/modules/others/pages/platform-http-main.adoc @@ -0,0 +1 @@ +../../../../../components/camel-platform-http-main/src/main/docs/platform-http-main.adoc \ No newline at end of file diff --git a/dsl/camel-componentdsl/src/generated/java/org/apache/camel/builder/component/dsl/KafkaComponentBuilderFactory.java b/dsl/camel-componentdsl/src/generated/java/org/apache/camel/builder/component/dsl/KafkaComponentBuilderFactory.java index feaa422d09a..d919a48dfea 100644 --- a/dsl/camel-componentdsl/src/generated/java/org/apache/camel/builder/component/dsl/KafkaComponentBuilderFactory.java +++ b/dsl/camel-componentdsl/src/generated/java/org/apache/camel/builder/component/dsl/KafkaComponentBuilderFactory.java @@ -254,7 +254,7 @@ public interface KafkaComponentBuilderFactory { * message that caused a failure, and then re-attempt to process this * message. However this can lead to endless processing of the same * message if its bound to fail every time, eg a poison message. - * Therefore its recommended to deal with that for example by using + * Therefore it is recommended to deal with that for example by using * Camel's error handler. * * The option is a: <code>boolean</code> type. @@ -501,7 +501,7 @@ public interface KafkaComponentBuilderFactory { return this; } /** - * Deserializer class for key that implements the Deserializer + * Deserializer class for the key that implements the Deserializer * interface. * * The option is a: <code>java.lang.String</code> type. @@ -654,9 +654,9 @@ public interface KafkaComponentBuilderFactory { return this; } /** - * Set if KafkaConsumer will read from beginning or end on startup: - * SeekPolicy.BEGINNING: read from beginning. SeekPolicy.END: read from - * end. + * Set if KafkaConsumer will read from the beginning or the end on + * startup: SeekPolicy.BEGINNING: read from the beginning. + * SeekPolicy.END: read from the end. * * The option is a: * <code>org.apache.camel.component.kafka.SeekPolicy</code> @@ -765,12 +765,12 @@ public interface KafkaComponentBuilderFactory { * possible. However, one part of the validation is DNS resolution of * the bootstrap broker hostnames. This may be a temporary networking * problem, and could potentially be recoverable. While other errors are - * fatal such as some invalid kafka configurations. Unfortunately + * fatal, such as some invalid kafka configurations. Unfortunately, * kafka-client does not separate this kind of errors. Camel will by * default retry forever, and therefore never give up. If you want to * give up after many attempts then set this option and Camel will then - * when giving up terminate the consumer. You can manually restart the - * consumer by stopping and starting the route, to try again. + * when giving up terminate the consumer. To try again, you can manually + * restart the consumer by stopping, and starting the route. * * The option is a: <code>int</code> type. * @@ -874,7 +874,7 @@ public interface KafkaComponentBuilderFactory { * subscribing the consumer to the kafka topic could be temporary errors * due to network issues, and could potentially be recoverable. Camel * will by default retry forever, and therefore never give up. If you - * want to give up after many attempts then set this option and Camel + * want to give up after many attempts, then set this option and Camel * will then when giving up terminate the consumer. You can manually * restart the consumer by stopping and starting the route, to try * again. @@ -1100,7 +1100,7 @@ public interface KafkaComponentBuilderFactory { * accomplishes this by adding a small amount of artificial delay that * is, rather than immediately sending out a record the producer will * wait for up to the given delay to allow other records to be sent so - * that the sends can be batched together. This can be thought of as + * that they can be batched together. This can be thought of as * analogous to Nagle's algorithm in TCP. This setting gives the upper * bound on the delay for batching: once we get batch.size worth of * records for a partition it will be sent immediately regardless of @@ -1132,7 +1132,7 @@ public interface KafkaComponentBuilderFactory { * serializers or partitioner is not counted against this timeout). For * partitionsFor() this timeout bounds the time spent waiting for * metadata if it is unavailable. The transaction-related methods always - * block, but may timeout if the transaction coordinator could not be + * block, but may time out if the transaction coordinator could not be * discovered or did not respond within the timeout. * * The option is a: <code>java.lang.Integer</code> type. @@ -1293,9 +1293,9 @@ public interface KafkaComponentBuilderFactory { * requests whenever multiple records are being sent to the same * partition. This helps performance on both the client and the server. * This configuration controls the default batch size in bytes. No - * attempt will be made to batch records larger than this size.Requests + * attempt will be made to batch records larger than this size. Requests * sent to brokers will contain multiple batches, one for each partition - * with data available to be sent.A small batch size will make batching + * with data available to be sent. A small batch size will make batching * less common and may reduce throughput (a batch size of zero will * disable batching entirely). A very large batch size may use memory a * bit more wastefully as we will always allocate a buffer of the @@ -1472,9 +1472,9 @@ public interface KafkaComponentBuilderFactory { } /** * Before each retry, the producer refreshes the metadata of relevant - * topics to see if a new leader has been elected. Since leader election - * takes a bit of time, this property specifies the amount of time that - * the producer waits before refreshing the metadata. + * topics to see if a new leader has been elected. Since the leader + * election takes a bit of time, this property specifies the amount of + * time that the producer waits before refreshing the metadata. * * The option is a: <code>java.lang.Integer</code> type. * @@ -1958,8 +1958,8 @@ public interface KafkaComponentBuilderFactory { return this; } /** - * The location of the key store file. This is optional for client and - * can be used for two-way authentication for client. + * The location of the key store file. This is optional for the client + * and can be used for two-way authentication for the client. * * The option is a: <code>java.lang.String</code> type. * @@ -1974,7 +1974,7 @@ public interface KafkaComponentBuilderFactory { return this; } /** - * The store password for the key store file. This is optional for + * The store password for the key store file. This is optional for the * client and only needed if sslKeystoreLocation' is configured. Key * store password is not supported for PEM format. * diff --git a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/KafkaEndpointBuilderFactory.java b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/KafkaEndpointBuilderFactory.java index 02ec4c69b27..2c777640af8 100644 --- a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/KafkaEndpointBuilderFactory.java +++ b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/KafkaEndpointBuilderFactory.java @@ -376,7 +376,7 @@ public interface KafkaEndpointBuilderFactory { * message that caused a failure, and then re-attempt to process this * message. However this can lead to endless processing of the same * message if its bound to fail every time, eg a poison message. - * Therefore its recommended to deal with that for example by using + * Therefore it is recommended to deal with that for example by using * Camel's error handler. * * The option is a: <code>boolean</code> type. @@ -400,7 +400,7 @@ public interface KafkaEndpointBuilderFactory { * message that caused a failure, and then re-attempt to process this * message. However this can lead to endless processing of the same * message if its bound to fail every time, eg a poison message. - * Therefore its recommended to deal with that for example by using + * Therefore it is recommended to deal with that for example by using * Camel's error handler. * * The option will be converted to a <code>boolean</code> @@ -799,7 +799,7 @@ public interface KafkaEndpointBuilderFactory { return this; } /** - * Deserializer class for key that implements the Deserializer + * Deserializer class for the key that implements the Deserializer * interface. * * The option is a: <code>java.lang.String</code> type. @@ -1073,9 +1073,9 @@ public interface KafkaEndpointBuilderFactory { return this; } /** - * Set if KafkaConsumer will read from beginning or end on startup: - * SeekPolicy.BEGINNING: read from beginning. SeekPolicy.END: read from - * end. + * Set if KafkaConsumer will read from the beginning or the end on + * startup: SeekPolicy.BEGINNING: read from the beginning. + * SeekPolicy.END: read from the end. * * The option is a: * <code>org.apache.camel.component.kafka.SeekPolicy</code> @@ -1092,9 +1092,9 @@ public interface KafkaEndpointBuilderFactory { return this; } /** - * Set if KafkaConsumer will read from beginning or end on startup: - * SeekPolicy.BEGINNING: read from beginning. SeekPolicy.END: read from - * end. + * Set if KafkaConsumer will read from the beginning or the end on + * startup: SeekPolicy.BEGINNING: read from the beginning. + * SeekPolicy.END: read from the end. * * The option will be converted to a * <code>org.apache.camel.component.kafka.SeekPolicy</code> @@ -1633,8 +1633,8 @@ public interface KafkaEndpointBuilderFactory { return this; } /** - * The location of the key store file. This is optional for client and - * can be used for two-way authentication for client. + * The location of the key store file. This is optional for the client + * and can be used for two-way authentication for the client. * * The option is a: <code>java.lang.String</code> type. * @@ -1649,7 +1649,7 @@ public interface KafkaEndpointBuilderFactory { return this; } /** - * The store password for the key store file. This is optional for + * The store password for the key store file. This is optional for the * client and only needed if sslKeystoreLocation' is configured. Key * store password is not supported for PEM format. * @@ -2573,7 +2573,7 @@ public interface KafkaEndpointBuilderFactory { * accomplishes this by adding a small amount of artificial delay that * is, rather than immediately sending out a record the producer will * wait for up to the given delay to allow other records to be sent so - * that the sends can be batched together. This can be thought of as + * that they can be batched together. This can be thought of as * analogous to Nagle's algorithm in TCP. This setting gives the upper * bound on the delay for batching: once we get batch.size worth of * records for a partition it will be sent immediately regardless of @@ -2605,7 +2605,7 @@ public interface KafkaEndpointBuilderFactory { * accomplishes this by adding a small amount of artificial delay that * is, rather than immediately sending out a record the producer will * wait for up to the given delay to allow other records to be sent so - * that the sends can be batched together. This can be thought of as + * that they can be batched together. This can be thought of as * analogous to Nagle's algorithm in TCP. This setting gives the upper * bound on the delay for batching: once we get batch.size worth of * records for a partition it will be sent immediately regardless of @@ -2638,7 +2638,7 @@ public interface KafkaEndpointBuilderFactory { * serializers or partitioner is not counted against this timeout). For * partitionsFor() this timeout bounds the time spent waiting for * metadata if it is unavailable. The transaction-related methods always - * block, but may timeout if the transaction coordinator could not be + * block, but may time out if the transaction coordinator could not be * discovered or did not respond within the timeout. * * The option is a: <code>java.lang.Integer</code> type. @@ -2662,7 +2662,7 @@ public interface KafkaEndpointBuilderFactory { * serializers or partitioner is not counted against this timeout). For * partitionsFor() this timeout bounds the time spent waiting for * metadata if it is unavailable. The transaction-related methods always - * block, but may timeout if the transaction coordinator could not be + * block, but may time out if the transaction coordinator could not be * discovered or did not respond within the timeout. * * The option will be converted to a @@ -2934,9 +2934,9 @@ public interface KafkaEndpointBuilderFactory { * requests whenever multiple records are being sent to the same * partition. This helps performance on both the client and the server. * This configuration controls the default batch size in bytes. No - * attempt will be made to batch records larger than this size.Requests + * attempt will be made to batch records larger than this size. Requests * sent to brokers will contain multiple batches, one for each partition - * with data available to be sent.A small batch size will make batching + * with data available to be sent. A small batch size will make batching * less common and may reduce throughput (a batch size of zero will * disable batching entirely). A very large batch size may use memory a * bit more wastefully as we will always allocate a buffer of the @@ -2960,9 +2960,9 @@ public interface KafkaEndpointBuilderFactory { * requests whenever multiple records are being sent to the same * partition. This helps performance on both the client and the server. * This configuration controls the default batch size in bytes. No - * attempt will be made to batch records larger than this size.Requests + * attempt will be made to batch records larger than this size. Requests * sent to brokers will contain multiple batches, one for each partition - * with data available to be sent.A small batch size will make batching + * with data available to be sent. A small batch size will make batching * less common and may reduce throughput (a batch size of zero will * disable batching entirely). A very large batch size may use memory a * bit more wastefully as we will always allocate a buffer of the @@ -3267,9 +3267,9 @@ public interface KafkaEndpointBuilderFactory { } /** * Before each retry, the producer refreshes the metadata of relevant - * topics to see if a new leader has been elected. Since leader election - * takes a bit of time, this property specifies the amount of time that - * the producer waits before refreshing the metadata. + * topics to see if a new leader has been elected. Since the leader + * election takes a bit of time, this property specifies the amount of + * time that the producer waits before refreshing the metadata. * * The option is a: <code>java.lang.Integer</code> type. * @@ -3286,9 +3286,9 @@ public interface KafkaEndpointBuilderFactory { } /** * Before each retry, the producer refreshes the metadata of relevant - * topics to see if a new leader has been elected. Since leader election - * takes a bit of time, this property specifies the amount of time that - * the producer waits before refreshing the metadata. + * topics to see if a new leader has been elected. Since the leader + * election takes a bit of time, this property specifies the amount of + * time that the producer waits before refreshing the metadata. * * The option will be converted to a * <code>java.lang.Integer</code> type. @@ -3864,8 +3864,8 @@ public interface KafkaEndpointBuilderFactory { return this; } /** - * The location of the key store file. This is optional for client and - * can be used for two-way authentication for client. + * The location of the key store file. This is optional for the client + * and can be used for two-way authentication for the client. * * The option is a: <code>java.lang.String</code> type. * @@ -3880,7 +3880,7 @@ public interface KafkaEndpointBuilderFactory { return this; } /** - * The store password for the key store file. This is optional for + * The store password for the key store file. This is optional for the * client and only needed if sslKeystoreLocation' is configured. Key * store password is not supported for PEM format. * @@ -4746,8 +4746,8 @@ public interface KafkaEndpointBuilderFactory { return this; } /** - * The location of the key store file. This is optional for client and - * can be used for two-way authentication for client. + * The location of the key store file. This is optional for the client + * and can be used for two-way authentication for the client. * * The option is a: <code>java.lang.String</code> type. * @@ -4762,7 +4762,7 @@ public interface KafkaEndpointBuilderFactory { return this; } /** - * The store password for the key store file. This is optional for + * The store password for the key store file. This is optional for the * client and only needed if sslKeystoreLocation' is configured. Key * store password is not supported for PEM format. *
