Carl-Zhou-CN commented on code in PR #3297:
URL: 
https://github.com/apache/incubator-seatunnel/pull/3297#discussion_r1013667649


##########
docs/en/connector/sink/Kafka.mdx:
##########
@@ -17,48 +20,114 @@ Engine Supported and plugin name
 
 ## Options
 
+<Tabs
+    groupId="engine-type"
+    defaultValue="spark"
+    values={[
+        {label: 'Spark', value: 'spark'},
+        {label: 'Flink', value: 'flink'},
+    ]}>
+<TabItem value="spark">
+
 | name                       | type   | required | default value |
 | -------------------------- | ------ | -------- | ------------- |
 | producer.bootstrap.servers | string | yes      | -             |
 | topic                      | string | yes      | -             |
 | producer.*                 | string | no       | -             |
+| format                     | string | no       | json          |
+| common-options             | string | no       | -             |
+
+</TabItem>
+<TabItem value="flink">
+
+| name                       | type   | required | default value |
+| -------------------------- | ------ | -------- | ------------- |
+| producer.bootstrap.servers | string | yes      | -             |
+| topics                     | string | yes      | -             |
+| producer.*                 | string | no       | -             |
 | semantic                   | string | no       | -             |
 | common-options             | string | no       | -             |
 
-### producer.bootstrap.servers [string]
+</TabItem>
+</Tabs>
+
+## producer.bootstrap.servers [string]
 
 Kafka Brokers List
 
+### producer [string]
+
+In addition to the above parameters that must be specified by the `Kafka 
producer` client, the user can also specify multiple non-mandatory parameters 
for the `producer` client, covering [all the producer parameters specified in 
the official Kafka 
document](https://kafka.apache.org/documentation.html#producerconfigs).
+
+The way to specify the parameter is to add the prefix `producer.` to the 
original parameter name. For example, the way to specify `request.timeout.ms` 
is: `producer.request.timeout.ms = 60000` . If these non-essential parameters 
are not specified, they will use the default values given in the official Kafka 
documentation.
+
+<Tabs
+    groupId="engine-type"
+    defaultValue="spark"
+    values={[
+        {label: 'Spark', value: 'spark'},
+        {label: 'Flink', value: 'flink'},
+    ]}>
+<TabItem value="spark">
+
 ### topic [string]
 
 Kafka Topic
 
-### producer [string]
+### format [string]
 
-In addition to the above parameters that must be specified by the `Kafka 
producer` client, the user can also specify multiple non-mandatory parameters 
for the `producer` client, covering [all the producer parameters specified in 
the official Kafka 
document](https://kafka.apache.org/documentation.html#producerconfigs).
+Write to kafka's data format:[text],[json]
 
-The way to specify the parameter is to add the prefix `producer.` to the 
original parameter name. For example, the way to specify `request.timeout.ms` 
is: `producer.request.timeout.ms = 60000` . If these non-essential parameters 
are not specified, they will use the default values given in the official Kafka 
documentation.
+</TabItem>
+<TabItem value="flink">
+
+### topics [string]
+
+Kafka Topic todo:Only one topic is supported for now
 
 ### semantic [string]
-Semantics that can be chosen. exactly_once/at_least_once/none, default is 
at_least_once
 
+Semantics that can be chosen. exactly_once/at_least_once/none, default is 
at_least_once
 In exactly_once, flink producer will write all messages in a Kafka transaction 
that will be committed to Kafka on a checkpoint.
-
 In at_least_once, flink producer will wait for all outstanding messages in the 
Kafka buffers to be acknowledged by the Kafka producer on a checkpoint.
-
 NONE does not provide any guarantees: messages may be lost in case of issues 
on the Kafka broker and messages may be duplicated in case of a Flink failure.
-
 please refer to [Flink Kafka Fault 
Tolerance](https://nightlies.apache.org/flink/flink-docs-release-1.14/docs/connectors/datastream/kafka/#fault-tolerance)
 
+</TabItem>
+</Tabs>
+
+
 ### common options [string]
 
-Sink plugin common parameters, please refer to [Sink 
Plugin](common-options.md) for details
+Source plugin common parameters, please refer to [Sink 
Plugin](common-options.mdx) for details
 
 ## Examples
 
+<Tabs
+    groupId="engine-type"
+    defaultValue="spark"
+    values={[
+        {label: 'Spark', value: 'spark'},
+        {label: 'Flink', value: 'flink'},
+    ]}>
+<TabItem value="spark">
+
 ```bash
-kafka {
+    kafka {

Review Comment:
   OK



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to