This is an automated email from the ASF dual-hosted git repository.

gaojun2048 pushed a commit to branch main
in repository 
https://gitbox.apache.org/repos/asf/incubator-seatunnel-website.git


The following commit(s) were added to refs/heads/main by this push:
     new dcb776c0ff [HotFix] [Doc] Fix Kafka Sink Doc (#220)
dcb776c0ff is described below

commit dcb776c0ff47b5dd03a1c52b6e4716da6fde62fc
Author: Hisoka <[email protected]>
AuthorDate: Mon Mar 27 19:22:20 2023 +0800

    [HotFix] [Doc] Fix Kafka Sink Doc (#220)
---
 .../version-2.3.1/connector-v2/sink/Kafka.md       | 22 ++++++++++++----------
 1 file changed, 12 insertions(+), 10 deletions(-)

diff --git a/versioned_docs/version-2.3.1/connector-v2/sink/Kafka.md 
b/versioned_docs/version-2.3.1/connector-v2/sink/Kafka.md
index 58160d1f2d..38a0e488b0 100644
--- a/versioned_docs/version-2.3.1/connector-v2/sink/Kafka.md
+++ b/versioned_docs/version-2.3.1/connector-v2/sink/Kafka.md
@@ -51,12 +51,10 @@ Currently two formats are supported:
 
 Kafka Brokers List.
 
-### kafka.* [kafka producer config]
+### kafka.config [kafka producer config]
 
 In addition to the above parameters that must be specified by the `Kafka 
producer` client, the user can also specify multiple non-mandatory parameters 
for the `producer` client, covering [all the producer parameters specified in 
the official Kafka 
document](https://kafka.apache.org/documentation.html#producerconfigs).
 
-The way to specify the parameter is to add the prefix `kafka.` to the original 
parameter name. For example, the way to specify `request.timeout.ms` is: 
`kafka.request.timeout.ms = 60000` . If these non-essential parameters are not 
specified, they will use the default values given in the official Kafka 
documentation.
-
 ### semantics [string]
 
 Semantics that can be chosen EXACTLY_ONCE/AT_LEAST_ONCE/NON, default NON.
@@ -156,9 +154,11 @@ sink {
       format = json
       kafka.request.timeout.ms = 60000
       semantics = EXACTLY_ONCE
-      kafka.security.protocol=SASL_SSL
-      kafka.sasl.mechanism=SCRAM-SHA-512
-      
kafka.sasl.jaas.config="org.apache.kafka.common.security.scram.ScramLoginModule 
required \nusername=${username}\npassword=${password};"
+      kafka.config = {
+         security.protocol=SASL_SSL
+         sasl.mechanism=SCRAM-SHA-512
+         
sasl.jaas.config="org.apache.kafka.common.security.scram.ScramLoginModule 
required \nusername=${username}\npassword=${password};"
+      }
   }
   
 }
@@ -190,10 +190,12 @@ sink {
       format = json
       kafka.request.timeout.ms = 60000
       semantics = EXACTLY_ONCE
-      kafka.security.protocol=SASL_SSL
-      kafka.sasl.mechanism=AWS_MSK_IAM
-      kafka.sasl.jaas.config="software.amazon.msk.auth.iam.IAMLoginModule 
required;"
-      
kafka.sasl.client.callback.handler.class="software.amazon.msk.auth.iam.IAMClientCallbackHandler"
+      kafka.config = {
+         security.protocol=SASL_SSL
+         sasl.mechanism=AWS_MSK_IAM
+         sasl.jaas.config="software.amazon.msk.auth.iam.IAMLoginModule 
required;"
+         
sasl.client.callback.handler.class="software.amazon.msk.auth.iam.IAMClientCallbackHandler"
+      }
   }
   
 }

Reply via email to