This is an automated email from the ASF dual-hosted git repository. corgy pushed a commit to branch dev in repository https://gitbox.apache.org/repos/asf/seatunnel.git
The following commit(s) were added to refs/heads/dev by this push:
new 42d516edc8 [Docs][Connector-V2][S3File] Fix connector s3file option
name (access_secret -> secret_key) (#9892)
42d516edc8 is described below
commit 42d516edc86be9b49f974a96e166d705292bfad4
Author: loupipalien <[email protected]>
AuthorDate: Wed Sep 24 22:12:12 2025 +0800
[Docs][Connector-V2][S3File] Fix connector s3file option name
(access_secret -> secret_key) (#9892)
---
docs/en/connector-v2/source/S3File.md | 4 ++--
docs/zh/connector-v2/sink/S3File.md | 2 +-
docs/zh/connector-v2/source/S3File.md | 4 ++--
3 files changed, 5 insertions(+), 5 deletions(-)
diff --git a/docs/en/connector-v2/source/S3File.md
b/docs/en/connector-v2/source/S3File.md
index 7f61c416a2..5ef70e3329 100644
--- a/docs/en/connector-v2/source/S3File.md
+++ b/docs/en/connector-v2/source/S3File.md
@@ -192,13 +192,13 @@ If you assign file type to `parquet` `orc`, schema option
not required, connecto
| name | type | required | default value
| Description
[...]
|---------------------------------|---------|----------|-------------------------------------------------------|--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
[...]
| path | string | yes | -
| The s3 path that needs to be read can have
sub paths, but the sub paths need to meet certain format requirements. Specific
requirements can be referred to "parse_partition_from_path" option
[...]
-| file_format_type | string | yes | -
| File type, supported as the following file
types: `text` `csv` `parquet` `orc` `json` `excel` `xml` `binary` `markdown`
[...]
+| file_format_type | string | yes | -
| File type, supported as the following file
types: `text` `csv` `parquet` `orc` `json` `excel` `xml` `binary` `markdown`
[...]
| bucket | string | yes | -
| The bucket address of s3 file system, for
example: `s3n://seatunnel-test`, if you use `s3a` protocol, this parameter
should be `s3a://seatunnel-test`.
[...]
| fs.s3a.endpoint | string | yes | -
| fs s3a endpoint
[...]
| fs.s3a.aws.credentials.provider | string | yes |
com.amazonaws.auth.InstanceProfileCredentialsProvider | The way to authenticate
s3a. We only support `org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider`
and `com.amazonaws.auth.InstanceProfileCredentialsProvider` now. More
information about the credential provider you can see [Hadoop AWS
Document](https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html#Simple_name.2Fsecret_credentials_with_SimpleAWSCredenti
[...]
| read_columns | list | no | -
| The read column list of the data source, user
can use it to implement field projection. The file type supported column
projection as the following shown: `text` `csv` `parquet` `orc` `json` `excel`
`xml` . If the user wants to use this feature when reading `text` `json` `csv`
files, the "schema" option must be configured.
[...]
| access_key | string | no | -
| Only used when
`fs.s3a.aws.credentials.provider =
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider `
[...]
-| access_secret | string | no | -
| Only used when
`fs.s3a.aws.credentials.provider =
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider `
[...]
+| secret_key | string | no | -
| Only used when
`fs.s3a.aws.credentials.provider =
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider `
[...]
| hadoop_s3_properties | map | no | -
| If you need to add other option, you could
add it here and refer to this
[link](https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html)
[...]
| delimiter/field_delimiter | string | no | \001 for text and ,
for csv | Field delimiter, used to tell connector how
to slice and dice fields when reading text files. Default `\001`, the same as
hive's default delimiter.
[...]
| row_delimiter | string | no | \n
| Row delimiter, used to tell connector how to
slice and dice rows when reading text files. Default `\n`.
[...]
diff --git a/docs/zh/connector-v2/sink/S3File.md
b/docs/zh/connector-v2/sink/S3File.md
index f663a1e2cf..a869054248 100644
--- a/docs/zh/connector-v2/sink/S3File.md
+++ b/docs/zh/connector-v2/sink/S3File.md
@@ -110,7 +110,7 @@ import ChangeLog from '../changelog/connector-file-s3.md';
| fs.s3a.endpoint | string | 是 | -
|
|
| fs.s3a.aws.credentials.provider | string | 是 |
com.amazonaws.auth.InstanceProfileCredentialsProvider | 认证 s3a 的方式。目前仅支持
`org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider` 和
`com.amazonaws.auth.InstanceProfileCredentialsProvider`。 |
| access_key | string | 否 | -
| 仅当 fs.s3a.aws.credentials.provider =
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider 时使用
|
-| access_secret | string | 否 | -
| 仅当 fs.s3a.aws.credentials.provider =
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider 时使用
|
+| secret_key | string | 否 | -
| 仅当 fs.s3a.aws.credentials.provider =
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider 时使用
|
| custom_filename | boolean | 否 | false
| 是否需要自定义文件名
|
| file_name_expression | string | 否 | "${transactionId}"
| 仅当 custom_filename 为 true 时使用
|
| filename_time_format | string | 否 | "yyyy.MM.dd"
| 仅当 custom_filename 为 true 时使用
|
diff --git a/docs/zh/connector-v2/source/S3File.md
b/docs/zh/connector-v2/source/S3File.md
index fdd6027882..1d7b90ed1c 100644
--- a/docs/zh/connector-v2/source/S3File.md
+++ b/docs/zh/connector-v2/source/S3File.md
@@ -192,13 +192,13 @@ schema {
| 名称 | 类型 | 是否必需 | 默认值
| 描述
|
|---------------------------------|---------|------|-------------------------------------------------------|-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| path | string | 是 | -
|
需要读取的s3路径,可以有子路径,但子路径需要满足一定的格式要求。具体要求可以参考"parse_partition_from_path"选项
|
-| file_format_type | string | 是 | -
| 文件类型,支持以下文件类型:`text` `csv` `parquet` `orc` `json`
`excel` `xml` `binary` `markdown`
|
+| file_format_type | string | 是 | -
| 文件类型,支持以下文件类型:`text` `csv` `parquet` `orc` `json`
`excel` `xml` `binary` `markdown`
|
| bucket | string | 是 | -
|
s3文件系统的bucket地址,例如:`s3n://seatunnel-test`,如果您使用`s3a`协议,此参数应为`s3a://seatunnel-test`。
|
| fs.s3a.endpoint | string | 是 | -
| fs s3a端点
|
| fs.s3a.aws.credentials.provider | string | 是 |
com.amazonaws.auth.InstanceProfileCredentialsProvider |
s3a的认证方式。我们目前只支持`org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider`和`com.amazonaws.auth.InstanceProfileCredentialsProvider`。有关凭据提供程序的更多信息,您可以查看[Hadoop
AWS文档](https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html#Simple_name.2Fsecret_credentials_with_SimpleAWSCredentialsProvider.2A)
|
| read_columns | list | 否 | -
| 数据源的读取列列表,用户可以使用它来实现字段投影。支持列投影的文件类型如下所示:`text`
`csv` `parquet` `orc` `json` `excel` `xml`。如果用户想在读取`text` `json`
`csv`文件时使用此功能,必须配置"schema"选项。
|
| access_key | string | 否 | -
| 仅在`fs.s3a.aws.credentials.provider =
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider`时使用
|
-| access_secret | string | 否 | -
| 仅在`fs.s3a.aws.credentials.provider =
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider`时使用
|
+| secret_key | string | 否 | -
| 仅在`fs.s3a.aws.credentials.provider =
org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider`时使用
|
| hadoop_s3_properties | map | 否 | -
|
如果您需要添加其他选项,可以在此处添加并参考此[链接](https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html)
|
| delimiter/field_delimiter | string | 否 | \001
|
字段分隔符,用于告诉连接器在读取文本文件时如何切分字段。默认`\001`,与hive的默认分隔符相同。
|
| row_delimiter | string | 否 | \n
| 行分隔符,用于告诉连接器在读取文本文件时如何切分行。默认`\n`。
|
[...]
