This is an automated email from the ASF dual-hosted git repository.
morningman pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/doris-website.git
The following commit(s) were added to refs/heads/master by this push:
new 30dc8aacacb [fix](HDFS-HA) Modify HDFS HA properties (#867)
30dc8aacacb is described below
commit 30dc8aacacb0da11c8187bbffdf0e9dd577b9832
Author: Tiewei Fang <[email protected]>
AuthorDate: Fri Jul 19 13:12:34 2024 +0800
[fix](HDFS-HA) Modify HDFS HA properties (#867)
---
docs/data-operate/import/broker-load-manual.md | 6 +++---
.../Data-Definition-Statements/Create/CREATE-RESOURCE.md | 2 +-
.../sql-statements/Data-Manipulation-Statements/Load/BROKER-LOAD.md | 3 +--
.../sql-statements/Data-Manipulation-Statements/OUTFILE.md | 2 +-
docs/table-design/cold-hot-separation.md | 2 +-
.../current/data-operate/import/broker-load-manual.md | 6 +++---
.../Data-Definition-Statements/Create/CREATE-RESOURCE.md | 2 +-
.../current/table-design/cold-hot-separation.md | 2 +-
.../docusaurus-plugin-content-docs/version-1.2/advanced/broker.md | 6 +++---
.../version-1.2/data-operate/import/import-way/spark-load-manual.md | 2 +-
.../Data-Definition-Statements/Create/CREATE-RESOURCE.md | 2 +-
.../sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md | 2 +-
.../sql-reference/Data-Manipulation-Statements/OUTFILE.md | 2 +-
.../version-2.0/data-operate/import/broker-load-manual.md | 6 +++---
.../Data-Definition-Statements/Create/CREATE-RESOURCE.md | 2 +-
.../sql-reference/Data-Manipulation-Statements/OUTFILE.md | 2 +-
.../version-2.0/table-design/cold-hot-separation.md | 2 +-
.../version-2.1/data-operate/import/broker-load-manual.md | 6 +++---
.../Data-Definition-Statements/Create/CREATE-RESOURCE.md | 2 +-
.../version-2.1/table-design/cold-hot-separation.md | 2 +-
versioned_docs/version-1.2/advanced/broker.md | 6 +++---
.../version-1.2/data-operate/import/import-way/spark-load-manual.md | 2 +-
.../Data-Definition-Statements/Create/CREATE-RESOURCE.md | 2 +-
.../sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md | 2 +-
.../version-2.0/data-operate/import/broker-load-manual.md | 6 +++---
.../Data-Definition-Statements/Create/CREATE-RESOURCE.md | 2 +-
.../sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md | 2 +-
.../sql-reference/Data-Manipulation-Statements/OUTFILE.md | 2 +-
versioned_docs/version-2.0/table-design/cold-hot-separation.md | 2 +-
.../version-2.1/data-operate/import/broker-load-manual.md | 6 +++---
.../Data-Definition-Statements/Create/CREATE-RESOURCE.md | 2 +-
.../sql-statements/Data-Manipulation-Statements/Load/BROKER-LOAD.md | 2 +-
.../sql-statements/Data-Manipulation-Statements/OUTFILE.md | 2 +-
versioned_docs/version-2.1/table-design/cold-hot-separation.md | 2 +-
34 files changed, 50 insertions(+), 51 deletions(-)
diff --git a/docs/data-operate/import/broker-load-manual.md
b/docs/data-operate/import/broker-load-manual.md
index 553b8365584..a1d328e938b 100644
--- a/docs/data-operate/import/broker-load-manual.md
+++ b/docs/data-operate/import/broker-load-manual.md
@@ -163,7 +163,7 @@ This configuration is used to access HDFS clusters deployed
in HA (High Availabi
- **dfs.namenode.rpc-address.xxx.nn:** Specifies the RPC address information
for the namenode. In this context, nn represents the namenode name configured
in dfs.ha.namenodes.xxx. For example: "dfs.namenode.rpc-address.my_ha.my_nn" =
"host:port".
-- **dfs.client.failover.proxy.provider:** Specifies the provider for client
connections to the namenode. The default is
org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider.
+- **dfs.client.failover.proxy.provider.[nameservice ID]:** Specifies the
provider for client connections to the namenode. The default is
org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider.
An example configuration is as follows:
@@ -174,7 +174,7 @@ An example configuration is as follows:
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
@@ -189,7 +189,7 @@ HA mode can be combined with the previous two
authentication methods for cluster
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
diff --git
a/docs/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
b/docs/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
index 0dff3597175..ff57aa909c0 100644
---
a/docs/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
+++
b/docs/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
@@ -189,7 +189,7 @@ illustrate:
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/docs/sql-manual/sql-statements/Data-Manipulation-Statements/Load/BROKER-LOAD.md
b/docs/sql-manual/sql-statements/Data-Manipulation-Statements/Load/BROKER-LOAD.md
index ecd8dc647c8..daadf3a4b83 100644
---
a/docs/sql-manual/sql-statements/Data-Manipulation-Statements/Load/BROKER-LOAD.md
+++
b/docs/sql-manual/sql-statements/Data-Manipulation-Statements/Load/BROKER-LOAD.md
@@ -277,12 +277,11 @@ WITH BROKER broker_name
(
"username" = "",
"password" = "",
- "fs.defaultFS" = "hdfs://my_ha",
"dfs.nameservices" = "my_ha",
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
````
diff --git
a/docs/sql-manual/sql-statements/Data-Manipulation-Statements/OUTFILE.md
b/docs/sql-manual/sql-statements/Data-Manipulation-Statements/OUTFILE.md
index ff35235aaab..5799c27a53c 100644
--- a/docs/sql-manual/sql-statements/Data-Manipulation-Statements/OUTFILE.md
+++ b/docs/sql-manual/sql-statements/Data-Manipulation-Statements/OUTFILE.md
@@ -238,7 +238,7 @@ Parquet and ORC file formats have their own data types. The
export function of D
"broker.dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"broker.dfs.namenode.rpc-address.my_ha.my_namenode1" =
"nn1_host:rpc_port",
"broker.dfs.namenode.rpc-address.my_ha.my_namenode2" =
"nn2_host:rpc_port",
- "broker.dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "broker.dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
````
diff --git a/docs/table-design/cold-hot-separation.md
b/docs/table-design/cold-hot-separation.md
index 736d2a70619..875f0645c22 100644
--- a/docs/table-design/cold-hot-separation.md
+++ b/docs/table-design/cold-hot-separation.md
@@ -113,7 +113,7 @@ CREATE RESOURCE "remote_hdfs" PROPERTIES (
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
CREATE STORAGE POLICY test_policy PROPERTIES (
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/broker-load-manual.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/broker-load-manual.md
index d922f2827fd..8606ec9d02e 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/broker-load-manual.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/broker-load-manual.md
@@ -163,7 +163,7 @@ username 配置为要访问的用户,密码置空即可。
- `dfs.namenode.rpc-address.xxx.nn`:指定 namenode 的 rpc 地址信息。其中 nn 表示
`dfs.ha.namenodes.xxx` 中配置的 namenode
的名字,如:"dfs.namenode.rpc-address.my_ha.my_nn" = "host:port"。
-- `dfs.client.failover.proxy.provider`:指定 client 连接 namenode 的
provider,默认为:org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider。
+- `dfs.client.failover.proxy.provider.[nameservice ID]`:指定 client 连接 namenode
的
provider,默认为:org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider。
示例如下:
@@ -174,7 +174,7 @@ username 配置为要访问的用户,密码置空即可。
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
@@ -189,7 +189,7 @@ HA 模式可以和前面两种认证方式组合,进行集群访问。如通
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
index 0572ae508a4..d8428345593 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
@@ -189,7 +189,7 @@ Spark 用于 ETL 时需要指定 working_dir 和 broker。说明如下:
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/table-design/cold-hot-separation.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/table-design/cold-hot-separation.md
index cd1b28b7de7..fee3852aba7 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/table-design/cold-hot-separation.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/table-design/cold-hot-separation.md
@@ -111,7 +111,7 @@ CREATE RESOURCE "remote_hdfs" PROPERTIES (
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
CREATE STORAGE POLICY test_policy PROPERTIES (
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/advanced/broker.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/advanced/broker.md
index e82123e7264..8b59f474050 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/advanced/broker.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/advanced/broker.md
@@ -166,7 +166,7 @@ WITH BROKER "broker_name"
- `dfs.nameservices`:指定 hdfs 服务的名字,自定义,如:"dfs.nameservices" = "my_ha"。
- `dfs.ha.namenodes.xxx`:自定义 namenode 的名字,多个名字以逗号分隔。其中 xxx 为
`dfs.nameservices` 中自定义的名字,如: "dfs.ha.namenodes.my_ha" = "my_nn"。
- `dfs.namenode.rpc-address.xxx.nn`:指定 namenode 的rpc地址信息。其中 nn 表示
`dfs.ha.namenodes.xxx` 中配置的 namenode
的名字,如:"dfs.namenode.rpc-address.my_ha.my_nn" = "host:port"。
- - `dfs.client.failover.proxy.provider`:指定 client 连接 namenode 的
provider,默认为:org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider。
+ - `dfs.client.failover.proxy.provider.[nameservice ID]`:指定 client 连接
namenode 的
provider,默认为:org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider。
示例如下:
@@ -176,7 +176,7 @@ WITH BROKER "broker_name"
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
@@ -190,7 +190,7 @@ WITH BROKER "broker_name"
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/data-operate/import/import-way/spark-load-manual.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/data-operate/import/import-way/spark-load-manual.md
index 07968d3c296..717f6c7c364 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/data-operate/import/import-way/spark-load-manual.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/data-operate/import/import-way/spark-load-manual.md
@@ -246,7 +246,7 @@ PROPERTIES
"broker.dfs.ha.namenodes.HDFS4001273" = "mynamenode1, mynamenode2",
"broker.dfs.namenode.rpc-address.nameservices01.mynamenode1" = "xxxx:8020",
"broker.dfs.namenode.rpc-address.nameservices01.mynamenode2" = "xxxx:8020",
- "broker.dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "broker.dfs.client.failover.proxy.provider.nameservices01" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
index 10606d7df43..cf2312175d4 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
@@ -189,7 +189,7 @@ PROPERTIES ("key"="value", ...);
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
index 71835f18d38..988a8a3201b 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
@@ -253,7 +253,7 @@ WITH BROKER broker_name
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/sql-manual/sql-reference/Data-Manipulation-Statements/OUTFILE.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/sql-manual/sql-reference/Data-Manipulation-Statements/OUTFILE.md
index c85dea91036..44a455de5bc 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/sql-manual/sql-reference/Data-Manipulation-Statements/OUTFILE.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2/sql-manual/sql-reference/Data-Manipulation-Statements/OUTFILE.md
@@ -166,7 +166,7 @@ INTO OUTFILE "file_path"
"broker.dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"broker.dfs.namenode.rpc-address.my_ha.my_namenode1" =
"nn1_host:rpc_port",
"broker.dfs.namenode.rpc-address.my_ha.my_namenode2" =
"nn2_host:rpc_port",
- "broker.dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "broker.dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/data-operate/import/broker-load-manual.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/data-operate/import/broker-load-manual.md
index 0d103df81e6..2e5f2894cac 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/data-operate/import/broker-load-manual.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/data-operate/import/broker-load-manual.md
@@ -163,7 +163,7 @@ username 配置为要访问的用户,密码置空即可。
- `dfs.namenode.rpc-address.xxx.nn`:指定 namenode 的 rpc 地址信息。其中 nn 表示
`dfs.ha.namenodes.xxx` 中配置的 namenode
的名字,如:"dfs.namenode.rpc-address.my_ha.my_nn" = "host:port"。
-- `dfs.client.failover.proxy.provider`:指定 client 连接 namenode 的
provider,默认为:org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider。
+- `dfs.client.failover.proxy.provider.[nameservice ID]`:指定 client 连接 namenode
的
provider,默认为:org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider。
示例如下:
@@ -174,7 +174,7 @@ username 配置为要访问的用户,密码置空即可。
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
@@ -189,7 +189,7 @@ HA 模式可以和前面两种认证方式组合,进行集群访问。如通
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
index da162d031ae..6db7db4c20d 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
@@ -189,7 +189,7 @@ Spark 用于 ETL 时需要指定 working_dir 和 broker。说明如下:
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/sql-manual/sql-reference/Data-Manipulation-Statements/OUTFILE.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/sql-manual/sql-reference/Data-Manipulation-Statements/OUTFILE.md
index 6dae4f8f51b..3cedfa869e1 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/sql-manual/sql-reference/Data-Manipulation-Statements/OUTFILE.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/sql-manual/sql-reference/Data-Manipulation-Statements/OUTFILE.md
@@ -239,7 +239,7 @@ parquet、orc文件格式拥有自己的数据类型,Doris的导出功能能
"broker.dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"broker.dfs.namenode.rpc-address.my_ha.my_namenode1" =
"nn1_host:rpc_port",
"broker.dfs.namenode.rpc-address.my_ha.my_namenode2" =
"nn2_host:rpc_port",
- "broker.dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "broker.dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/table-design/cold-hot-separation.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/table-design/cold-hot-separation.md
index c859e106728..671f6e4623c 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/table-design/cold-hot-separation.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/table-design/cold-hot-separation.md
@@ -111,7 +111,7 @@ CREATE RESOURCE "remote_hdfs" PROPERTIES (
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
CREATE STORAGE POLICY test_policy PROPERTIES (
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/broker-load-manual.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/broker-load-manual.md
index d922f2827fd..8606ec9d02e 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/broker-load-manual.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/broker-load-manual.md
@@ -163,7 +163,7 @@ username 配置为要访问的用户,密码置空即可。
- `dfs.namenode.rpc-address.xxx.nn`:指定 namenode 的 rpc 地址信息。其中 nn 表示
`dfs.ha.namenodes.xxx` 中配置的 namenode
的名字,如:"dfs.namenode.rpc-address.my_ha.my_nn" = "host:port"。
-- `dfs.client.failover.proxy.provider`:指定 client 连接 namenode 的
provider,默认为:org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider。
+- `dfs.client.failover.proxy.provider.[nameservice ID]`:指定 client 连接 namenode
的
provider,默认为:org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider。
示例如下:
@@ -174,7 +174,7 @@ username 配置为要访问的用户,密码置空即可。
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
@@ -189,7 +189,7 @@ HA 模式可以和前面两种认证方式组合,进行集群访问。如通
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
index 0572ae508a4..d8428345593 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
@@ -189,7 +189,7 @@ Spark 用于 ETL 时需要指定 working_dir 和 broker。说明如下:
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/table-design/cold-hot-separation.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/table-design/cold-hot-separation.md
index cd1b28b7de7..fee3852aba7 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/table-design/cold-hot-separation.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/table-design/cold-hot-separation.md
@@ -111,7 +111,7 @@ CREATE RESOURCE "remote_hdfs" PROPERTIES (
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
CREATE STORAGE POLICY test_policy PROPERTIES (
diff --git a/versioned_docs/version-1.2/advanced/broker.md
b/versioned_docs/version-1.2/advanced/broker.md
index aa4ce1fe529..a98c67852c9 100644
--- a/versioned_docs/version-1.2/advanced/broker.md
+++ b/versioned_docs/version-1.2/advanced/broker.md
@@ -174,7 +174,7 @@ Authentication information is usually provided as a
Key-Value in the Property Ma
* `dfs.nameservices`: Specify the name of the hdfs service, custom, such
as "dfs.nameservices" = "my_ha".
* `dfs.ha.namenodes.xxx`: Custom namenode names. Multiple names are
separated by commas, where xxx is the custom name in `dfs.nameservices`, such
as" dfs.ha.namenodes.my_ha "=" my_nn ".
* `dfs.namenode.rpc-address.xxx.nn`: Specify the rpc address information
of namenode, Where nn represents the name of the namenode configured in
`dfs.ha.namenodes.xxx`, such as: "dfs.namenode.rpc-address.my_ha.my_nn" =
"host:port".
- * `dfs.client.failover.proxy.provider`: Specify the provider for the
client to connect to the namenode. The default is:
org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider.
+ * `dfs.client.failover.proxy.provider.[nameservice ID]`: Specify the
provider for the client to connect to the namenode. The default is:
org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider.
Examples are as follows:
@@ -184,7 +184,7 @@ Authentication information is usually provided as a
Key-Value in the Property Ma
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
@@ -198,7 +198,7 @@ Authentication information is usually provided as a
Key-Value in the Property Ma
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
The configuration for accessing the HDFS cluster can be written to the
hdfs-site.xml file. When users use the Broker process to read data from the
HDFS cluster, they only need to fill in the cluster file path and
authentication information.
diff --git
a/versioned_docs/version-1.2/data-operate/import/import-way/spark-load-manual.md
b/versioned_docs/version-1.2/data-operate/import/import-way/spark-load-manual.md
index fd78ac819fb..13edf563de0 100644
---
a/versioned_docs/version-1.2/data-operate/import/import-way/spark-load-manual.md
+++
b/versioned_docs/version-1.2/data-operate/import/import-way/spark-load-manual.md
@@ -272,7 +272,7 @@ PROPERTIES
"broker.dfs.ha.namenodes.HDFS4001273" = "mynamenode1, mynamenode2",
"broker.dfs.namenode.rpc-address.nameservices01.mynamenode1" = "xxxx:8020",
"broker.dfs.namenode.rpc-address.nameservices01.mynamenode2" = "xxxx:8020",
- "broker.dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "broker.dfs.client.failover.proxy.provider.nameservices01" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/versioned_docs/version-1.2/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
b/versioned_docs/version-1.2/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
index 484b5d5a78b..4d2d8f04d94 100644
---
a/versioned_docs/version-1.2/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
+++
b/versioned_docs/version-1.2/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
@@ -189,7 +189,7 @@ illustrate:
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/versioned_docs/version-1.2/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
b/versioned_docs/version-1.2/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
index 9c6f41f8e45..d2fb9a3d144 100644
---
a/versioned_docs/version-1.2/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
+++
b/versioned_docs/version-1.2/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
@@ -253,7 +253,7 @@ WITH BROKER broker_name
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
````
diff --git
a/versioned_docs/version-2.0/data-operate/import/broker-load-manual.md
b/versioned_docs/version-2.0/data-operate/import/broker-load-manual.md
index b11f5ae044a..4ae5847da93 100644
--- a/versioned_docs/version-2.0/data-operate/import/broker-load-manual.md
+++ b/versioned_docs/version-2.0/data-operate/import/broker-load-manual.md
@@ -163,7 +163,7 @@ This configuration is used to access HDFS clusters deployed
in HA (High Availabi
- **dfs.namenode.rpc-address.xxx.nn:** Specifies the RPC address information
for the namenode. In this context, nn represents the namenode name configured
in dfs.ha.namenodes.xxx. For example: "dfs.namenode.rpc-address.my_ha.my_nn" =
"host:port".
-- **dfs.client.failover.proxy.provider:** Specifies the provider for client
connections to the namenode. The default is
org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider.
+- **dfs.client.failover.proxy.provider.[nameservice ID]:** Specifies the
provider for client connections to the namenode. The default is
org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider.
An example configuration is as follows:
@@ -174,7 +174,7 @@ An example configuration is as follows:
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
@@ -189,7 +189,7 @@ HA mode can be combined with the previous two
authentication methods for cluster
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
diff --git
a/versioned_docs/version-2.0/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
b/versioned_docs/version-2.0/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
index 04f70bd0056..85421952c97 100644
---
a/versioned_docs/version-2.0/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
+++
b/versioned_docs/version-2.0/sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-RESOURCE.md
@@ -189,7 +189,7 @@ illustrate:
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/versioned_docs/version-2.0/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
b/versioned_docs/version-2.0/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
index 2eb22d00c87..0b813215102 100644
---
a/versioned_docs/version-2.0/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
+++
b/versioned_docs/version-2.0/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
@@ -282,7 +282,7 @@ WITH BROKER broker_name
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
````
diff --git
a/versioned_docs/version-2.0/sql-manual/sql-reference/Data-Manipulation-Statements/OUTFILE.md
b/versioned_docs/version-2.0/sql-manual/sql-reference/Data-Manipulation-Statements/OUTFILE.md
index 9bf2608ce95..b5db23c9cdf 100644
---
a/versioned_docs/version-2.0/sql-manual/sql-reference/Data-Manipulation-Statements/OUTFILE.md
+++
b/versioned_docs/version-2.0/sql-manual/sql-reference/Data-Manipulation-Statements/OUTFILE.md
@@ -237,7 +237,7 @@ Parquet and ORC file formats have their own data types. The
export function of D
"broker.dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"broker.dfs.namenode.rpc-address.my_ha.my_namenode1" =
"nn1_host:rpc_port",
"broker.dfs.namenode.rpc-address.my_ha.my_namenode2" =
"nn2_host:rpc_port",
- "broker.dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "broker.dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
````
diff --git a/versioned_docs/version-2.0/table-design/cold-hot-separation.md
b/versioned_docs/version-2.0/table-design/cold-hot-separation.md
index 0a2b92cf971..43348379611 100644
--- a/versioned_docs/version-2.0/table-design/cold-hot-separation.md
+++ b/versioned_docs/version-2.0/table-design/cold-hot-separation.md
@@ -113,7 +113,7 @@ CREATE RESOURCE "remote_hdfs" PROPERTIES (
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
CREATE STORAGE POLICY test_policy PROPERTIES (
diff --git
a/versioned_docs/version-2.1/data-operate/import/broker-load-manual.md
b/versioned_docs/version-2.1/data-operate/import/broker-load-manual.md
index 553b8365584..a1d328e938b 100644
--- a/versioned_docs/version-2.1/data-operate/import/broker-load-manual.md
+++ b/versioned_docs/version-2.1/data-operate/import/broker-load-manual.md
@@ -163,7 +163,7 @@ This configuration is used to access HDFS clusters deployed
in HA (High Availabi
- **dfs.namenode.rpc-address.xxx.nn:** Specifies the RPC address information
for the namenode. In this context, nn represents the namenode name configured
in dfs.ha.namenodes.xxx. For example: "dfs.namenode.rpc-address.my_ha.my_nn" =
"host:port".
-- **dfs.client.failover.proxy.provider:** Specifies the provider for client
connections to the namenode. The default is
org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider.
+- **dfs.client.failover.proxy.provider.[nameservice ID]:** Specifies the
provider for client connections to the namenode. The default is
org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider.
An example configuration is as follows:
@@ -174,7 +174,7 @@ An example configuration is as follows:
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
@@ -189,7 +189,7 @@ HA mode can be combined with the previous two
authentication methods for cluster
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
)
```
diff --git
a/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
b/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
index 0dff3597175..ff57aa909c0 100644
---
a/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
+++
b/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Definition-Statements/Create/CREATE-RESOURCE.md
@@ -189,7 +189,7 @@ illustrate:
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
```
diff --git
a/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/Load/BROKER-LOAD.md
b/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/Load/BROKER-LOAD.md
index 77045b05ef9..19c8d75287d 100644
---
a/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/Load/BROKER-LOAD.md
+++
b/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/Load/BROKER-LOAD.md
@@ -282,7 +282,7 @@ WITH BROKER broker_name
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
````
diff --git
a/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/OUTFILE.md
b/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/OUTFILE.md
index e0ab12ec935..eb37d5d9036 100644
---
a/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/OUTFILE.md
+++
b/versioned_docs/version-2.1/sql-manual/sql-statements/Data-Manipulation-Statements/OUTFILE.md
@@ -237,7 +237,7 @@ Parquet and ORC file formats have their own data types. The
export function of D
"broker.dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"broker.dfs.namenode.rpc-address.my_ha.my_namenode1" =
"nn1_host:rpc_port",
"broker.dfs.namenode.rpc-address.my_ha.my_namenode2" =
"nn2_host:rpc_port",
- "broker.dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "broker.dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
````
diff --git a/versioned_docs/version-2.1/table-design/cold-hot-separation.md
b/versioned_docs/version-2.1/table-design/cold-hot-separation.md
index 736d2a70619..875f0645c22 100644
--- a/versioned_docs/version-2.1/table-design/cold-hot-separation.md
+++ b/versioned_docs/version-2.1/table-design/cold-hot-separation.md
@@ -113,7 +113,7 @@ CREATE RESOURCE "remote_hdfs" PROPERTIES (
"dfs.ha.namenodes.my_ha" = "my_namenode1, my_namenode2",
"dfs.namenode.rpc-address.my_ha.my_namenode1" = "nn1_host:rpc_port",
"dfs.namenode.rpc-address.my_ha.my_namenode2" = "nn2_host:rpc_port",
- "dfs.client.failover.proxy.provider" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
+ "dfs.client.failover.proxy.provider.my_ha" =
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
);
CREATE STORAGE POLICY test_policy PROPERTIES (
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]