This is an automated email from the ASF dual-hosted git repository.

corgy pushed a commit to branch dev
in repository https://gitbox.apache.org/repos/asf/seatunnel.git


The following commit(s) were added to refs/heads/dev by this push:
     new 42d516edc8 [Docs][Connector-V2][S3File] Fix connector s3file option 
name (access_secret -> secret_key) (#9892)
42d516edc8 is described below

commit 42d516edc86be9b49f974a96e166d705292bfad4
Author: loupipalien <[email protected]>
AuthorDate: Wed Sep 24 22:12:12 2025 +0800

    [Docs][Connector-V2][S3File] Fix connector s3file option name 
(access_secret -> secret_key) (#9892)
---
 docs/en/connector-v2/source/S3File.md | 4 ++--
 docs/zh/connector-v2/sink/S3File.md   | 2 +-
 docs/zh/connector-v2/source/S3File.md | 4 ++--
 3 files changed, 5 insertions(+), 5 deletions(-)

diff --git a/docs/en/connector-v2/source/S3File.md 
b/docs/en/connector-v2/source/S3File.md
index 7f61c416a2..5ef70e3329 100644
--- a/docs/en/connector-v2/source/S3File.md
+++ b/docs/en/connector-v2/source/S3File.md
@@ -192,13 +192,13 @@ If you assign file type to `parquet` `orc`, schema option 
not required, connecto
 | name                            | type    | required | default value         
                                | Description                                   
                                                                                
                                                                                
                                                                                
                                                                                
              [...]
 
|---------------------------------|---------|----------|-------------------------------------------------------|--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
 [...]
 | path                            | string  | yes      | -                     
                                | The s3 path that needs to be read can have 
sub paths, but the sub paths need to meet certain format requirements. Specific 
requirements can be referred to "parse_partition_from_path" option              
                                                                                
                                                                                
                 [...]
-| file_format_type                | string  | yes      | -                     
                                | File type, supported as the following file 
types: `text` `csv` `parquet` `orc` `json` `excel` `xml` `binary` `markdown`    
                                                                                
                                                                                
                                                                                
                 [...]
+| file_format_type                | string  | yes      | -                     
                                | File type, supported as the following file 
types: `text` `csv` `parquet` `orc` `json` `excel` `xml` `binary` `markdown`    
                                                                                
                                                                                
                                                                                
                 [...]
 | bucket                          | string  | yes      | -                     
                                | The bucket address of s3 file system, for 
example: `s3n://seatunnel-test`, if you use `s3a` protocol, this parameter 
should be `s3a://seatunnel-test`.                                               
                                                                                
                                                                                
                       [...]
 | fs.s3a.endpoint                 | string  | yes      | -                     
                                | fs s3a endpoint                               
                                                                                
                                                                                
                                                                                
                                                                                
              [...]
 | fs.s3a.aws.credentials.provider | string  | yes      | 
com.amazonaws.auth.InstanceProfileCredentialsProvider | The way to authenticate 
s3a. We only support `org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider` 
and `com.amazonaws.auth.InstanceProfileCredentialsProvider` now. More 
information about the credential provider you can see [Hadoop AWS 
Document](https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html#Simple_name.2Fsecret_credentials_with_SimpleAWSCredenti
 [...]
 | read_columns                    | list    | no       | -                     
                                | The read column list of the data source, user 
can use it to implement field projection. The file type supported column 
projection as the following shown: `text` `csv` `parquet` `orc` `json` `excel` 
`xml` . If the user wants to use this feature when reading `text` `json` `csv` 
files, the "schema" option must be configured.                                  
                       [...]
 | access_key                      | string  | no       | -                     
                                | Only used when 
`fs.s3a.aws.credentials.provider = 
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider `                         
                                                                                
                                                                                
                                                                                
          [...]
-| access_secret                   | string  | no       | -                     
                                | Only used when 
`fs.s3a.aws.credentials.provider = 
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider `                         
                                                                                
                                                                                
                                                                                
          [...]
+| secret_key                      | string  | no       | -                     
                                | Only used when 
`fs.s3a.aws.credentials.provider = 
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider `                         
                                                                                
                                                                                
                                                                                
          [...]
 | hadoop_s3_properties            | map     | no       | -                     
                                | If you need to add other option, you could 
add it here and refer to this 
[link](https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html)
                                                                                
                                                                                
                                                               [...]
 | delimiter/field_delimiter       | string  | no       | \001 for text and , 
for csv                           | Field delimiter, used to tell connector how 
to slice and dice fields when reading text files. Default `\001`, the same as 
hive's default delimiter.                                                       
                                                                                
                                                                                
                  [...]
 | row_delimiter                   | string  | no       | \n                    
                                | Row delimiter, used to tell connector how to 
slice and dice rows when reading text files. Default `\n`.                      
                                                                                
                                                                                
                                                                                
               [...]
diff --git a/docs/zh/connector-v2/sink/S3File.md 
b/docs/zh/connector-v2/sink/S3File.md
index f663a1e2cf..a869054248 100644
--- a/docs/zh/connector-v2/sink/S3File.md
+++ b/docs/zh/connector-v2/sink/S3File.md
@@ -110,7 +110,7 @@ import ChangeLog from '../changelog/connector-file-s3.md';
 | fs.s3a.endpoint                       | string  | 是    | -                   
                                  |                                             
                                                                                
        |
 | fs.s3a.aws.credentials.provider       | string  | 是    | 
com.amazonaws.auth.InstanceProfileCredentialsProvider | 认证 s3a 的方式。目前仅支持 
`org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider` 和 
`com.amazonaws.auth.InstanceProfileCredentialsProvider`。 |
 | access_key                            | string  | 否    | -                   
                                  | 仅当 fs.s3a.aws.credentials.provider = 
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider 时使用                       
               |
-| access_secret                         | string  | 否    | -                   
                                  | 仅当 fs.s3a.aws.credentials.provider = 
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider 时使用                       
               |
+| secret_key                            | string  | 否    | -                   
                                  | 仅当 fs.s3a.aws.credentials.provider = 
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider 时使用                       
               |
 | custom_filename                       | boolean | 否    | false               
                                  | 是否需要自定义文件名                                  
                                                                                
        |
 | file_name_expression                  | string  | 否    | "${transactionId}"  
                                  | 仅当 custom_filename 为 true 时使用               
                                                                                
        |
 | filename_time_format                  | string  | 否    | "yyyy.MM.dd"        
                                  | 仅当 custom_filename 为 true 时使用               
                                                                                
        |
diff --git a/docs/zh/connector-v2/source/S3File.md 
b/docs/zh/connector-v2/source/S3File.md
index fdd6027882..1d7b90ed1c 100644
--- a/docs/zh/connector-v2/source/S3File.md
+++ b/docs/zh/connector-v2/source/S3File.md
@@ -192,13 +192,13 @@ schema {
 | 名称                              | 类型      | 是否必需 | 默认值                       
                            | 描述                                                
                                                                                
                                                                                
                                                                                
                    |
 
|---------------------------------|---------|------|-------------------------------------------------------|-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
 | path                            | string  | 是    | -                         
                            | 
需要读取的s3路径,可以有子路径,但子路径需要满足一定的格式要求。具体要求可以参考"parse_partition_from_path"选项          
                                                                                
                                                                                
                                                                      |
-| file_format_type                | string  | 是    | -                         
                            | 文件类型,支持以下文件类型:`text` `csv` `parquet` `orc` `json` 
`excel` `xml` `binary` `markdown`                                               
                                                                                
                                                                                
                               |
+| file_format_type                | string  | 是    | -                         
                            | 文件类型,支持以下文件类型:`text` `csv` `parquet` `orc` `json` 
`excel` `xml` `binary` `markdown`                                               
                                                                                
                                                                                
                    |
 | bucket                          | string  | 是    | -                         
                            | 
s3文件系统的bucket地址,例如:`s3n://seatunnel-test`,如果您使用`s3a`协议,此参数应为`s3a://seatunnel-test`。
                                                                                
                                                                                
                                                                   |
 | fs.s3a.endpoint                 | string  | 是    | -                         
                            | fs s3a端点                                          
                                                                                
                                                                                
                                                                                
                    |
 | fs.s3a.aws.credentials.provider | string  | 是    | 
com.amazonaws.auth.InstanceProfileCredentialsProvider | 
s3a的认证方式。我们目前只支持`org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider`和`com.amazonaws.auth.InstanceProfileCredentialsProvider`。有关凭据提供程序的更多信息,您可以查看[Hadoop
 
AWS文档](https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html#Simple_name.2Fsecret_credentials_with_SimpleAWSCredentialsProvider.2A)
 |
 | read_columns                    | list    | 否    | -                         
                            | 数据源的读取列列表,用户可以使用它来实现字段投影。支持列投影的文件类型如下所示:`text` 
`csv` `parquet` `orc` `json` `excel` `xml`。如果用户想在读取`text` `json` 
`csv`文件时使用此功能,必须配置"schema"选项。                                                   
                                                                                
                                      |
 | access_key                      | string  | 否    | -                         
                            | 仅在`fs.s3a.aws.credentials.provider = 
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider`时使用                       
                                                                                
                                                                                
                                 |
-| access_secret                   | string  | 否    | -                         
                            | 仅在`fs.s3a.aws.credentials.provider = 
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider`时使用                       
                                                                                
                                                                                
                                 |
+| secret_key                      | string  | 否    | -                         
                            | 仅在`fs.s3a.aws.credentials.provider = 
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider`时使用                       
                                                                                
                                                                                
                                 |
 | hadoop_s3_properties            | map     | 否    | -                         
                            | 
如果您需要添加其他选项,可以在此处添加并参考此[链接](https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html)
                                                                                
                                                                                
                                             |
 | delimiter/field_delimiter       | string  | 否    | \001                      
                            | 
字段分隔符,用于告诉连接器在读取文本文件时如何切分字段。默认`\001`,与hive的默认分隔符相同。                             
                                                                                
                                                                                
                                                                      |
 | row_delimiter                   | string  | 否    | \n                        
                            | 行分隔符,用于告诉连接器在读取文本文件时如何切分行。默认`\n`。                 
                                                                                
                                                                                
                                                                                
                    |                                                           
              [...]

Reply via email to