This is an automated email from the ASF dual-hosted git repository.

morningman pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/doris-website.git


The following commit(s) were added to refs/heads/master by this push:
     new e6f0323922a [2.1.5] add some docs for 2.1.5 (#882)
e6f0323922a is described below

commit e6f0323922a14a91cb87ead0d07c042890802f60
Author: Mingyu Chen <[email protected]>
AuthorDate: Fri Jul 19 13:51:35 2024 +0800

    [2.1.5] add some docs for 2.1.5 (#882)
---
 docs/data-operate/import/insert-into-manual.md            |  6 ++++++
 docs/lakehouse/datalake-analytics/hive.md                 | 15 +++++++++++++++
 .../Data-Manipulation-Statements/Manipulation/EXPORT.md   |  2 +-
 .../current/data-operate/import/insert-into-manual.md     |  8 +++++++-
 .../current/lakehouse/datalake-analytics/hive.md          | 15 +++++++++++++++
 .../Data-Manipulation-Statements/Manipulation/EXPORT.md   |  2 +-
 .../version-2.1/data-operate/import/insert-into-manual.md |  8 +++++++-
 .../version-2.1/lakehouse/datalake-analytics/hive.md      | 15 +++++++++++++++
 .../Data-Manipulation-Statements/Manipulation/EXPORT.md   |  2 ++
 .../version-2.1/data-operate/import/insert-into-manual.md |  6 ++++++
 .../version-2.1/lakehouse/datalake-analytics/hive.md      | 15 +++++++++++++++
 .../Data-Manipulation-Statements/Manipulation/EXPORT.md   |  2 ++
 12 files changed, 92 insertions(+), 4 deletions(-)

diff --git a/docs/data-operate/import/insert-into-manual.md 
b/docs/data-operate/import/insert-into-manual.md
index 61116978795..d9d7984e5b4 100644
--- a/docs/data-operate/import/insert-into-manual.md
+++ b/docs/data-operate/import/insert-into-manual.md
@@ -192,6 +192,12 @@ enable_insert_strict
 - Description: If this is set to true, INSERT INTO will fail when the task 
involves invalid data. If set to false, INSERT INTO will ignore invalid rows, 
and the import will be considered successful as long as at least one row is 
imported successfully.
 - Explanation: INSERT INTO cannot control the error rate, so this parameter is 
used to either strictly check data quality or completely ignore invalid data. 
Common reasons for data invalidity include: source data column length exceeding 
destination column length, column type mismatch, partition mismatch, and column 
order mismatch.
 
+insert_max_filter_ratio
+
+- Default value: 1.0
+
+- Description: Since version 2.1.5. Only effective when `enable_insert_strict` 
is false. Used to control the error tolerance when using `INSERT INTO FROM 
S3/HDFS/LOCAL()`. The default value is 1.0, which means all errors are 
tolerated. It can be a decimal between 0 and 1. It means that when the number 
of error rows exceeds this ratio, the INSERT task will fail.
+
 ### Return values
 
 INSERT INTO an SQL statement, and it returns different results based on 
different query outcomes:
diff --git a/docs/lakehouse/datalake-analytics/hive.md 
b/docs/lakehouse/datalake-analytics/hive.md
index 6aaac07b1c2..741c9bd373d 100644
--- a/docs/lakehouse/datalake-analytics/hive.md
+++ b/docs/lakehouse/datalake-analytics/hive.md
@@ -509,6 +509,11 @@ This section mainly introduces how to connect to a Hive + 
HDFS cluster with Kerb
 
  Note that in some cases the file location of `krb5.conf` may depend on the 
environment variable `KRB5_CONFIG` or the `-Djava.security.krb5.conf` in the 
JVM parameters. Please check these properties to determine the exact location 
of `krb5.conf`.
 
+ To customize the location of `krb5.conf`:
+ 
+ - FE: Configure the JVM parameter `-Djava.security.krb5.conf` in `fe.conf`.
+ - BE: Use the `kerberos_krb5_conf_path` configuration item in `be.conf`, the 
default value is `/etc/krb5.conf`.
+
 * JVM parameters
 
  Please add the following options to the JVM of FE and BE (located in 
`fe.conf` and `be.conf`):
@@ -656,3 +661,13 @@ Hive transactional tables are tables in Hive that support 
ACID (Atomicity, Consi
 
 Currently, it does not support scenarios involving Original Files.
 When a table is transformed into a transactional table, subsequent newly 
written data files will use the schema of the Hive transactional table. 
However, existing data files will not be converted to the schema of the 
transactional table. These existing files are referred to as Original Files.
+
+## Best Practices
+
+- Handling of Empty Lines in Hive Text Format Tables
+
+    By default, Doris ignores empty lines in Text format tables. Starting from 
version 2.1.5, you can control this behavior by setting the session variable 
`read_csv_empty_line_as_null`.
+
+    `set read_csv_empty_line_as_null = true;`
+
+    The variable defaults to false, indicating that empty lines are ignored. 
If set to true, the empty line will be read as a line with "all columns are 
null" and returned, which is consistent with the behavior of some query engines 
in the Hadoop ecosystem.
diff --git 
a/docs/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
 
b/docs/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
index d913c48c6e6..301b97f180c 100644
--- 
a/docs/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
+++ 
b/docs/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
@@ -98,7 +98,7 @@ The bottom layer of the `Export` statement actually executes 
the `select...outfi
 
   - `timeout`: This is the timeout parameter of the export job, the default 
timeout is 2 hours, and the unit is seconds.
 
-  - `compress_type`: When specifying the export file format as Parquet or ORC, 
you can choose the compression method for the Parquet or ORC files. For Parquet 
file format, you can specify the compression method as SNAPPY, GZIP, BROTLI, 
ZSTD, LZ4, or PLAIN, with the default being SNAPPY. For ORC file format, you 
can specify the compression method as PLAIN, SNAPPY, ZLIB, or ZSTD, with the 
default being ZLIB. This parameter is supported starting from version 2.1.5. 
(PLAIN means no compressi [...]
+  - `compress_type`: (since 2.1.5) When specifying the export file format as 
Parquet or ORC, you can choose the compression method for the Parquet or ORC 
files. For Parquet file format, you can specify the compression method as 
SNAPPY, GZIP, BROTLI, ZSTD, LZ4, or PLAIN, with the default being SNAPPY. For 
ORC file format, you can specify the compression method as PLAIN, SNAPPY, ZLIB, 
or ZSTD, with the default being ZLIB. This parameter is supported starting from 
version 2.1.5. (PLAIN mean [...]
 
   > Note that to use the `delete_existing_files` parameter, you also need to 
add the configuration `enable_delete_existing_files = true` to the fe.conf file 
and restart the FE. Only then will the `delete_existing_files` parameter take 
effect. Setting `delete_existing_files = true` is a dangerous operation and it 
is recommended to only use it in a testing environment.
 
diff --git 
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/insert-into-manual.md
 
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/insert-into-manual.md
index 245af10c9e2..198c0be9549 100644
--- 
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/insert-into-manual.md
+++ 
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/insert-into-manual.md
@@ -201,6 +201,12 @@ VALUES (val1, val2, ...), (val3, val4, ...), ...;
 
 - 解释:INSERT INTO 
无法控制错误率,只能通过该参数设置为严格检查数据质量或完全忽略错误数据。常见的数据不合格的原因有:源数据列长度超过目的数据列长度、列类型不匹配、分区不匹配、列顺序不匹配等。
 
+**insert_max_filter_ratio**
+
+- 默认值:1.0
+
+- 参数描述:自 2.1.5 版本。仅当 `enable_insert_strict` 值为 false 时生效。用于控制当使用 `INSERT INTO 
FROM S3/HDFS/LOCAL()` 时,设定错误容忍率的。默认为 1.0 表示容忍所有错误。可以取值 0 ~ 1 
之间的小数。表示当错误行数超过该比例后,INSERT 任务会失败。
+
 **enable_nereids_dml_with_pipeline**
 
   设置为 `true` 后,`insert into` 语句将尝试通过 Pipeline 引擎执行。详见[导入](./load-manual)文档。
@@ -449,4 +455,4 @@ FROM s3(
 
 ## 更多帮助
 
-关于 Insert Into 使用的更多详细语法,请参阅 [INSERT 
INTO](../../sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/INSERT)
 命令手册,也可以在 MySQL 客户端命令行下输入 `HELP INSERT` 获取更多帮助信息。
\ No newline at end of file
+关于 Insert Into 使用的更多详细语法,请参阅 [INSERT 
INTO](../../sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/INSERT)
 命令手册,也可以在 MySQL 客户端命令行下输入 `HELP INSERT` 获取更多帮助信息。
diff --git 
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/lakehouse/datalake-analytics/hive.md
 
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/lakehouse/datalake-analytics/hive.md
index ec522ba8e66..9b7b9ebedde 100644
--- 
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/lakehouse/datalake-analytics/hive.md
+++ 
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/lakehouse/datalake-analytics/hive.md
@@ -491,6 +491,11 @@ Doris 支持为指定的 External Hive Catalog 使用 Apache Ranger 进行鉴权
 
  注意,某些情况下,`krb5.conf` 的文件位置可能取决于环境变量 `KRB5_CONFIG` 或 JVM 参数中的 
`-Djava.security.krb5.conf` 参数。请检查这些属性以确定 `krb5.conf` 的确切位置。
 
+ 如需自定义`krb5.conf`的位置:
+
+ - FE:在 `fe.conf` 配置 JVM 参数 `-Djava.security.krb5.conf`。
+ - BE:在 `be.conf` 使用 `kerberos_krb5_conf_path` 配置项,默认值为`/etc/krb5.conf`。
+
 * JVM 参数
 
  请在 FE 和 BE 的 JVM 参数中添加如下配置(位于 `fe.conf` 和 `be.conf` 中):
@@ -638,3 +643,13 @@ Hive Transactional 表是 Hive 中支持 ACID 语义的表。详情可见 [Hive
 
 目前不支持 Original Files 的场景。
 当一个表转换成 Transactional 表之后,后续新写的数据文件会使用 Hive Transactional 表的 
Schema,但是已经存在的数据文件是不会转化成 Transactional 表的 Schema,这样的文件称为 Original Files。
+
+## 最佳实践
+
+- Hive Text 格式表的中空行行为的处理
+
+    默认情况下,Doris 会忽略 Text 格式表中的空行。从 2.1.5 版本开始,可以通过设置会话变量 
`read_csv_empty_line_as_null` 来控制该行为。
+
+    `set read_csv_empty_line_as_null = true;`
+
+    该变量默认为 false,表示忽略空行。如果设置为 true,这空行会读取为“所有列都是 null” 的行并返回,这种行为和部分 Hadoop 
生态中查询引擎的行为一致。
diff --git 
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
 
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
index 7d2abcedbf2..9e391fab581 100644
--- 
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
+++ 
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
@@ -96,7 +96,7 @@ EXPORT
 
   - `timeout`:导出作业的超时时间,默认为2小时,单位是秒。
 
-  - `compress_type`:当指定导出的文件格式为 Parquet / ORC 文件时,可以指定 Parquet / ORC 
文件使用的压缩方式。Parquet 文件格式可指定压缩方式为 SNAPPY,GZIP,BROTLI,ZSTD,LZ4 及 PLAIN,默认值为 
SNAPPY。ORC 文件格式可指定压缩方式为 PLAIN,SNAPPY,ZLIB 以及 ZSTD,默认值为 
ZLIB。该参数自2.1.5版本开始支持。(PLAIN 就是不采用压缩)
+  - `compress_type`:(自 2.1.5 支持) 当指定导出的文件格式为 Parquet / ORC 文件时,可以指定 Parquet / 
ORC 文件使用的压缩方式。Parquet 文件格式可指定压缩方式为 SNAPPY,GZIP,BROTLI,ZSTD,LZ4 及 PLAIN,默认值为 
SNAPPY。ORC 文件格式可指定压缩方式为 PLAIN,SNAPPY,ZLIB 以及 ZSTD,默认值为 
ZLIB。该参数自2.1.5版本开始支持。(PLAIN 就是不采用压缩)
 
   :::caution 注意
   要使用delete_existing_files参数,还需要在fe.conf中添加配置`enable_delete_existing_files = 
true`并重启fe,此时delete_existing_files才会生效。delete_existing_files = true 
是一个危险的操作,建议只在测试环境中使用。
diff --git 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/insert-into-manual.md
 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/insert-into-manual.md
index 245af10c9e2..198c0be9549 100644
--- 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/insert-into-manual.md
+++ 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/insert-into-manual.md
@@ -201,6 +201,12 @@ VALUES (val1, val2, ...), (val3, val4, ...), ...;
 
 - 解释:INSERT INTO 
无法控制错误率,只能通过该参数设置为严格检查数据质量或完全忽略错误数据。常见的数据不合格的原因有:源数据列长度超过目的数据列长度、列类型不匹配、分区不匹配、列顺序不匹配等。
 
+**insert_max_filter_ratio**
+
+- 默认值:1.0
+
+- 参数描述:自 2.1.5 版本。仅当 `enable_insert_strict` 值为 false 时生效。用于控制当使用 `INSERT INTO 
FROM S3/HDFS/LOCAL()` 时,设定错误容忍率的。默认为 1.0 表示容忍所有错误。可以取值 0 ~ 1 
之间的小数。表示当错误行数超过该比例后,INSERT 任务会失败。
+
 **enable_nereids_dml_with_pipeline**
 
   设置为 `true` 后,`insert into` 语句将尝试通过 Pipeline 引擎执行。详见[导入](./load-manual)文档。
@@ -449,4 +455,4 @@ FROM s3(
 
 ## 更多帮助
 
-关于 Insert Into 使用的更多详细语法,请参阅 [INSERT 
INTO](../../sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/INSERT)
 命令手册,也可以在 MySQL 客户端命令行下输入 `HELP INSERT` 获取更多帮助信息。
\ No newline at end of file
+关于 Insert Into 使用的更多详细语法,请参阅 [INSERT 
INTO](../../sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/INSERT)
 命令手册,也可以在 MySQL 客户端命令行下输入 `HELP INSERT` 获取更多帮助信息。
diff --git 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/lakehouse/datalake-analytics/hive.md
 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/lakehouse/datalake-analytics/hive.md
index ec522ba8e66..9b7b9ebedde 100644
--- 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/lakehouse/datalake-analytics/hive.md
+++ 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/lakehouse/datalake-analytics/hive.md
@@ -491,6 +491,11 @@ Doris 支持为指定的 External Hive Catalog 使用 Apache Ranger 进行鉴权
 
  注意,某些情况下,`krb5.conf` 的文件位置可能取决于环境变量 `KRB5_CONFIG` 或 JVM 参数中的 
`-Djava.security.krb5.conf` 参数。请检查这些属性以确定 `krb5.conf` 的确切位置。
 
+ 如需自定义`krb5.conf`的位置:
+
+ - FE:在 `fe.conf` 配置 JVM 参数 `-Djava.security.krb5.conf`。
+ - BE:在 `be.conf` 使用 `kerberos_krb5_conf_path` 配置项,默认值为`/etc/krb5.conf`。
+
 * JVM 参数
 
  请在 FE 和 BE 的 JVM 参数中添加如下配置(位于 `fe.conf` 和 `be.conf` 中):
@@ -638,3 +643,13 @@ Hive Transactional 表是 Hive 中支持 ACID 语义的表。详情可见 [Hive
 
 目前不支持 Original Files 的场景。
 当一个表转换成 Transactional 表之后,后续新写的数据文件会使用 Hive Transactional 表的 
Schema,但是已经存在的数据文件是不会转化成 Transactional 表的 Schema,这样的文件称为 Original Files。
+
+## 最佳实践
+
+- Hive Text 格式表的中空行行为的处理
+
+    默认情况下,Doris 会忽略 Text 格式表中的空行。从 2.1.5 版本开始,可以通过设置会话变量 
`read_csv_empty_line_as_null` 来控制该行为。
+
+    `set read_csv_empty_line_as_null = true;`
+
+    该变量默认为 false,表示忽略空行。如果设置为 true,这空行会读取为“所有列都是 null” 的行并返回,这种行为和部分 Hadoop 
生态中查询引擎的行为一致。
diff --git 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
index 07d71dc8677..54c3036c51d 100644
--- 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
+++ 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
@@ -96,6 +96,8 @@ EXPORT
 
   - `timeout`:导出作业的超时时间,默认为2小时,单位是秒。
 
+  - `compress_type`:(自 2.1.5 支持) 当指定导出的文件格式为 Parquet / ORC 文件时,可以指定 Parquet / 
ORC 文件使用的压缩方式。Parquet 文件格式可指定压缩方式为 SNAPPY,GZIP,BROTLI,ZSTD,LZ4 及 PLAIN,默认值为 
SNAPPY。ORC 文件格式可指定压缩方式为 PLAIN,SNAPPY,ZLIB 以及 ZSTD,默认值为 
ZLIB。该参数自2.1.5版本开始支持。(PLAIN 就是不采用压缩)
+
   > 
注意:要使用delete_existing_files参数,还需要在fe.conf中添加配置`enable_delete_existing_files = 
true`并重启fe,此时delete_existing_files才会生效。delete_existing_files = true 
是一个危险的操作,建议只在测试环境中使用。
 
 
diff --git 
a/versioned_docs/version-2.1/data-operate/import/insert-into-manual.md 
b/versioned_docs/version-2.1/data-operate/import/insert-into-manual.md
index 61116978795..d9d7984e5b4 100644
--- a/versioned_docs/version-2.1/data-operate/import/insert-into-manual.md
+++ b/versioned_docs/version-2.1/data-operate/import/insert-into-manual.md
@@ -192,6 +192,12 @@ enable_insert_strict
 - Description: If this is set to true, INSERT INTO will fail when the task 
involves invalid data. If set to false, INSERT INTO will ignore invalid rows, 
and the import will be considered successful as long as at least one row is 
imported successfully.
 - Explanation: INSERT INTO cannot control the error rate, so this parameter is 
used to either strictly check data quality or completely ignore invalid data. 
Common reasons for data invalidity include: source data column length exceeding 
destination column length, column type mismatch, partition mismatch, and column 
order mismatch.
 
+insert_max_filter_ratio
+
+- Default value: 1.0
+
+- Description: Since version 2.1.5. Only effective when `enable_insert_strict` 
is false. Used to control the error tolerance when using `INSERT INTO FROM 
S3/HDFS/LOCAL()`. The default value is 1.0, which means all errors are 
tolerated. It can be a decimal between 0 and 1. It means that when the number 
of error rows exceeds this ratio, the INSERT task will fail.
+
 ### Return values
 
 INSERT INTO an SQL statement, and it returns different results based on 
different query outcomes:
diff --git a/versioned_docs/version-2.1/lakehouse/datalake-analytics/hive.md 
b/versioned_docs/version-2.1/lakehouse/datalake-analytics/hive.md
index 6aaac07b1c2..741c9bd373d 100644
--- a/versioned_docs/version-2.1/lakehouse/datalake-analytics/hive.md
+++ b/versioned_docs/version-2.1/lakehouse/datalake-analytics/hive.md
@@ -509,6 +509,11 @@ This section mainly introduces how to connect to a Hive + 
HDFS cluster with Kerb
 
  Note that in some cases the file location of `krb5.conf` may depend on the 
environment variable `KRB5_CONFIG` or the `-Djava.security.krb5.conf` in the 
JVM parameters. Please check these properties to determine the exact location 
of `krb5.conf`.
 
+ To customize the location of `krb5.conf`:
+ 
+ - FE: Configure the JVM parameter `-Djava.security.krb5.conf` in `fe.conf`.
+ - BE: Use the `kerberos_krb5_conf_path` configuration item in `be.conf`, the 
default value is `/etc/krb5.conf`.
+
 * JVM parameters
 
  Please add the following options to the JVM of FE and BE (located in 
`fe.conf` and `be.conf`):
@@ -656,3 +661,13 @@ Hive transactional tables are tables in Hive that support 
ACID (Atomicity, Consi
 
 Currently, it does not support scenarios involving Original Files.
 When a table is transformed into a transactional table, subsequent newly 
written data files will use the schema of the Hive transactional table. 
However, existing data files will not be converted to the schema of the 
transactional table. These existing files are referred to as Original Files.
+
+## Best Practices
+
+- Handling of Empty Lines in Hive Text Format Tables
+
+    By default, Doris ignores empty lines in Text format tables. Starting from 
version 2.1.5, you can control this behavior by setting the session variable 
`read_csv_empty_line_as_null`.
+
+    `set read_csv_empty_line_as_null = true;`
+
+    The variable defaults to false, indicating that empty lines are ignored. 
If set to true, the empty line will be read as a line with "all columns are 
null" and returned, which is consistent with the behavior of some query engines 
in the Hadoop ecosystem.
diff --git 
a/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
 
b/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
index c6479d0e824..301b97f180c 100644
--- 
a/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
+++ 
b/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/Manipulation/EXPORT.md
@@ -98,6 +98,8 @@ The bottom layer of the `Export` statement actually executes 
the `select...outfi
 
   - `timeout`: This is the timeout parameter of the export job, the default 
timeout is 2 hours, and the unit is seconds.
 
+  - `compress_type`: (since 2.1.5) When specifying the export file format as 
Parquet or ORC, you can choose the compression method for the Parquet or ORC 
files. For Parquet file format, you can specify the compression method as 
SNAPPY, GZIP, BROTLI, ZSTD, LZ4, or PLAIN, with the default being SNAPPY. For 
ORC file format, you can specify the compression method as PLAIN, SNAPPY, ZLIB, 
or ZSTD, with the default being ZLIB. This parameter is supported starting from 
version 2.1.5. (PLAIN mean [...]
+
   > Note that to use the `delete_existing_files` parameter, you also need to 
add the configuration `enable_delete_existing_files = true` to the fe.conf file 
and restart the FE. Only then will the `delete_existing_files` parameter take 
effect. Setting `delete_existing_files = true` is a dangerous operation and it 
is recommended to only use it in a testing environment.
 
 - `WITH BROKER`


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to