This is an automated email from the ASF dual-hosted git repository.
adoroszlai pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/ozone.git
The following commit(s) were added to refs/heads/master by this push:
new d34aee40c5 HDDS-12172. Rename Java constants of DFSConfigKeysLegacy
keys (#7922)
d34aee40c5 is described below
commit d34aee40c5799518ae0b93c7d01bf4c8b78a4054
Author: Sarveksha Yeshavantha Raju
<[email protected]>
AuthorDate: Thu Feb 27 21:32:25 2025 +0530
HDDS-12172. Rename Java constants of DFSConfigKeysLegacy keys (#7922)
---
.../apache/hadoop/hdds/DFSConfigKeysLegacy.java | 73 ----------------------
.../org/apache/hadoop/hdds/HddsConfigKeys.java | 28 +++++++++
.../java/org/apache/hadoop/hdds/HddsUtils.java | 14 ++---
.../hadoop/hdds/conf/OzoneConfiguration.java | 26 ++++----
.../org/apache/hadoop/hdds/ratis/RatisHelper.java | 6 +-
.../org/apache/hadoop/hdds/scm/ScmConfigKeys.java | 4 ++
.../org/apache/hadoop/ozone/OzoneConfigKeys.java | 1 -
.../apache/hadoop/ozone/HddsDatanodeService.java | 10 +--
.../container/common/helpers/ContainerMetrics.java | 4 +-
.../ozone/protocolPB/ReconDatanodeProtocolPB.java | 4 +-
.../StorageContainerDatanodeProtocolPB.java | 4 +-
.../hadoop/ozone/TestHddsDatanodeService.java | 3 +-
.../hadoop/ozone/TestHddsSecureDatanodeInit.java | 3 +-
.../container/common/volume/TestVolumeSet.java | 3 +-
.../common/volume/TestVolumeSetDiskChecks.java | 7 +--
.../hdds/protocol/SecretKeyProtocolDatanode.java | 4 +-
.../protocolPB/ReconfigureProtocolDatanodePB.java | 4 +-
.../protocolPB/SecretKeyProtocolDatanodePB.java | 4 +-
.../hadoop/hdds/server/http/BaseHttpServer.java | 9 ++-
.../apache/hadoop/hdds/utils/HddsServerUtil.java | 10 +--
.../hdds/scm/node/NodeDecommissionManager.java | 6 +-
.../hadoop/hdds/scm/node/SCMNodeManager.java | 6 +-
.../hdds/scm/server/StorageContainerManager.java | 3 +-
.../hadoop/hdds/scm/node/TestSCMNodeManager.java | 5 +-
.../scm/server/TestSCMBlockProtocolServer.java | 6 +-
.../http/server/metrics/HttpFSServerMetrics.java | 5 +-
.../hadoop/hdds/scm/TestSecretKeySnapshot.java | 8 +--
.../apache/hadoop/hdds/scm/TestSecretKeysApi.java | 8 +--
.../org/apache/hadoop/ozone/TestBlockTokens.java | 8 +--
.../apache/hadoop/ozone/TestBlockTokensCLI.java | 8 +--
.../apache/hadoop/ozone/TestOMSortDatanodes.java | 5 +-
.../hadoop/ozone/TestOzoneConfigurationFields.java | 14 ++++-
.../hadoop/ozone/UniformDatanodesFactory.java | 4 +-
.../container/metrics/TestContainerMetrics.java | 4 +-
.../org/apache/hadoop/ozone/om/KeyManagerImpl.java | 9 +--
35 files changed, 141 insertions(+), 179 deletions(-)
diff --git
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/DFSConfigKeysLegacy.java
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/DFSConfigKeysLegacy.java
deleted file mode 100644
index b712c18072..0000000000
---
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/DFSConfigKeysLegacy.java
+++ /dev/null
@@ -1,73 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements. See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.hadoop.hdds;
-
-/**
- * Legacy HDFS keys used by ozone.
- *
- * THey are the HDFS specific config keys. It would be better to use ozone
- * specific explicit configuration keys with ozone to make it more visible.
- */
-@Deprecated
-public final class DFSConfigKeysLegacy {
-
- private DFSConfigKeysLegacy() {
- }
-
- public static final String DFS_DATANODE_DNS_INTERFACE_KEY =
- "hdds.datanode.dns.interface";
- public static final String DFS_DATANODE_DNS_NAMESERVER_KEY =
- "hdds.datanode.dns.nameserver";
-
- public static final String DFS_DATANODE_HOST_NAME_KEY =
- "hdds.datanode.hostname";
-
- public static final String DFS_DATANODE_DATA_DIR_KEY =
- "hdds.datanode.data.dir";
-
- public static final String DFS_DATANODE_USE_DN_HOSTNAME =
- "hdds.datanode.use.datanode.hostname";
-
- public static final boolean DFS_DATANODE_USE_DN_HOSTNAME_DEFAULT = false;
-
- public static final String DFS_XFRAME_OPTION_ENABLED = "hdds.xframe.enabled";
-
- public static final boolean DFS_XFRAME_OPTION_ENABLED_DEFAULT = true;
-
- public static final String DFS_XFRAME_OPTION_VALUE = "hdds.xframe.value";
-
- public static final String DFS_XFRAME_OPTION_VALUE_DEFAULT = "SAMEORIGIN";
-
- public static final String DFS_METRICS_SESSION_ID_KEY =
- "hdds.metrics.session-id";
-
- public static final String NET_TOPOLOGY_NODE_SWITCH_MAPPING_IMPL_KEY =
- "net.topology.node.switch.mapping.impl";
-
- public static final String DFS_DATANODE_KERBEROS_PRINCIPAL_KEY =
- "hdds.datanode.kerberos.principal";
-
- public static final String DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY =
- "hdds.datanode.kerberos.keytab.file";
-
- public static final String DFS_METRICS_PERCENTILES_INTERVALS_KEY =
- "hdds.metrics.percentiles.intervals";
-
-}
-
-
diff --git
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/HddsConfigKeys.java
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/HddsConfigKeys.java
index a2aae1936c..d516cd5f27 100644
---
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/HddsConfigKeys.java
+++
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/HddsConfigKeys.java
@@ -405,4 +405,32 @@ private HddsConfigKeys() {
public static final String
OZONE_DATANODE_IO_METRICS_PERCENTILES_INTERVALS_SECONDS_KEY =
"ozone.volume.io.percentiles.intervals.seconds";
+
+ public static final String HDDS_DATANODE_DNS_INTERFACE_KEY =
+ "hdds.datanode.dns.interface";
+ public static final String HDDS_DATANODE_DNS_NAMESERVER_KEY =
+ "hdds.datanode.dns.nameserver";
+ public static final String HDDS_DATANODE_HOST_NAME_KEY =
+ "hdds.datanode.hostname";
+ public static final String HDDS_DATANODE_DATA_DIR_KEY =
+ "hdds.datanode.data.dir";
+ public static final String HDDS_DATANODE_USE_DN_HOSTNAME =
+ "hdds.datanode.use.datanode.hostname";
+ public static final boolean HDDS_DATANODE_USE_DN_HOSTNAME_DEFAULT = false;
+
+ public static final String HDDS_XFRAME_OPTION_ENABLED =
"hdds.xframe.enabled";
+ public static final boolean HDDS_XFRAME_OPTION_ENABLED_DEFAULT = true;
+ public static final String HDDS_XFRAME_OPTION_VALUE = "hdds.xframe.value";
+ public static final String HDDS_XFRAME_OPTION_VALUE_DEFAULT = "SAMEORIGIN";
+
+ public static final String HDDS_METRICS_SESSION_ID_KEY =
+ "hdds.metrics.session-id";
+
+ public static final String HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY =
+ "hdds.datanode.kerberos.principal";
+ public static final String HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY =
+ "hdds.datanode.kerberos.keytab.file";
+ public static final String HDDS_METRICS_PERCENTILES_INTERVALS_KEY =
+ "hdds.metrics.percentiles.intervals";
+
}
diff --git
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/HddsUtils.java
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/HddsUtils.java
index 2ffa73d18c..90c365cbbb 100644
--- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/HddsUtils.java
+++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/HddsUtils.java
@@ -17,14 +17,14 @@
package org.apache.hadoop.hdds;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_DNS_INTERFACE_KEY;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_DNS_NAMESERVER_KEY;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_HOST_NAME_KEY;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_CLIENT_ADDRESS_KEY;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_CLIENT_BIND_HOST_DEFAULT;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_CLIENT_BIND_HOST_KEY;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_CLIENT_PORT_DEFAULT;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_CLIENT_PORT_KEY;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_DNS_INTERFACE_KEY;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_DNS_NAMESERVER_KEY;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_HOST_NAME_KEY;
import static
org.apache.hadoop.hdds.recon.ReconConfigKeys.OZONE_RECON_ADDRESS_KEY;
import static
org.apache.hadoop.hdds.recon.ReconConfigKeys.OZONE_RECON_DATANODE_PORT_DEFAULT;
import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_ADDRESS_KEY;
@@ -363,7 +363,7 @@ public static InetSocketAddress getReconAddresses(
*/
public static String getHostName(ConfigurationSource conf)
throws UnknownHostException {
- String name = conf.get(DFS_DATANODE_HOST_NAME_KEY);
+ String name = conf.get(HDDS_DATANODE_HOST_NAME_KEY);
if (name == null) {
String dnsInterface = conf.get(
CommonConfigurationKeysPublic.HADOOP_SECURITY_DNS_INTERFACE_KEY);
@@ -373,9 +373,9 @@ public static String getHostName(ConfigurationSource conf)
if (dnsInterface == null) {
// Try the legacy configuration keys.
- dnsInterface = conf.get(DFS_DATANODE_DNS_INTERFACE_KEY);
- dnsInterface = conf.get(DFS_DATANODE_DNS_INTERFACE_KEY);
- nameServer = conf.get(DFS_DATANODE_DNS_NAMESERVER_KEY);
+ dnsInterface = conf.get(HDDS_DATANODE_DNS_INTERFACE_KEY);
+ dnsInterface = conf.get(HDDS_DATANODE_DNS_INTERFACE_KEY);
+ nameServer = conf.get(HDDS_DATANODE_DNS_NAMESERVER_KEY);
} else {
// If HADOOP_SECURITY_DNS_* is set then also attempt hosts file
// resolution if DNS fails. We will not use hosts file resolution
diff --git
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/conf/OzoneConfiguration.java
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/conf/OzoneConfiguration.java
index be689a7196..58f33d445a 100644
---
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/conf/OzoneConfiguration.java
+++
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/conf/OzoneConfiguration.java
@@ -47,7 +47,7 @@
import javax.xml.bind.annotation.XmlElement;
import javax.xml.bind.annotation.XmlRootElement;
import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.annotation.InterfaceAudience;
import org.apache.hadoop.hdds.scm.ScmConfigKeys;
import org.apache.hadoop.hdds.utils.LegacyHadoopConfigurationSource;
@@ -315,7 +315,7 @@ private static void addDeprecatedKeys() {
HDDS_DATANODE_RATIS_PREFIX_KEY + "."
+ RaftServerConfigKeys.PREFIX + "." + "rpc.slowness.timeout"),
new DeprecationDelta("dfs.datanode.keytab.file",
- DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY),
+ HddsConfigKeys.HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY),
new DeprecationDelta("ozone.scm.chunk.layout",
ScmConfigKeys.OZONE_SCM_CONTAINER_LAYOUT_KEY),
new DeprecationDelta("hdds.datanode.replication.work.dir",
@@ -381,21 +381,21 @@ private static void addDeprecatedKeys() {
new DeprecationDelta("dfs.ratis.snapshot.threshold",
ScmConfigKeys.HDDS_RATIS_SNAPSHOT_THRESHOLD_KEY),
new DeprecationDelta("dfs.datanode.dns.interface",
- DFSConfigKeysLegacy.DFS_DATANODE_DNS_INTERFACE_KEY),
+ HddsConfigKeys.HDDS_DATANODE_DNS_INTERFACE_KEY),
new DeprecationDelta("dfs.datanode.dns.nameserver",
- DFSConfigKeysLegacy.DFS_DATANODE_DNS_NAMESERVER_KEY),
+ HddsConfigKeys.HDDS_DATANODE_DNS_NAMESERVER_KEY),
new DeprecationDelta("dfs.datanode.hostname",
- DFSConfigKeysLegacy.DFS_DATANODE_HOST_NAME_KEY),
+ HddsConfigKeys.HDDS_DATANODE_HOST_NAME_KEY),
new DeprecationDelta("dfs.datanode.data.dir",
- DFSConfigKeysLegacy.DFS_DATANODE_DATA_DIR_KEY),
+ HddsConfigKeys.HDDS_DATANODE_DATA_DIR_KEY),
new DeprecationDelta("dfs.datanode.use.datanode.hostname",
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME),
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME),
new DeprecationDelta("dfs.xframe.enabled",
- DFSConfigKeysLegacy.DFS_XFRAME_OPTION_ENABLED),
+ HddsConfigKeys.HDDS_XFRAME_OPTION_ENABLED),
new DeprecationDelta("dfs.xframe.value",
- DFSConfigKeysLegacy.DFS_XFRAME_OPTION_VALUE),
+ HddsConfigKeys.HDDS_XFRAME_OPTION_VALUE),
new DeprecationDelta("dfs.metrics.session-id",
- DFSConfigKeysLegacy.DFS_METRICS_SESSION_ID_KEY),
+ HddsConfigKeys.HDDS_METRICS_SESSION_ID_KEY),
new DeprecationDelta("dfs.client.https.keystore.resource",
OzoneConfigKeys.OZONE_CLIENT_HTTPS_KEYSTORE_RESOURCE_KEY),
new DeprecationDelta("dfs.https.server.keystore.resource",
@@ -403,11 +403,11 @@ private static void addDeprecatedKeys() {
new DeprecationDelta("dfs.http.policy",
OzoneConfigKeys.OZONE_HTTP_POLICY_KEY),
new DeprecationDelta("dfs.datanode.kerberos.principal",
- DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY),
+ HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY),
new DeprecationDelta("dfs.datanode.kerberos.keytab.file",
- DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY),
+ HddsConfigKeys.HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY),
new DeprecationDelta("dfs.metrics.percentiles.intervals",
- DFSConfigKeysLegacy.DFS_METRICS_PERCENTILES_INTERVALS_KEY),
+ HddsConfigKeys.HDDS_METRICS_PERCENTILES_INTERVALS_KEY),
});
}
diff --git
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/ratis/RatisHelper.java
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/ratis/RatisHelper.java
index 2a061a628c..bcff6e6071 100644
---
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/ratis/RatisHelper.java
+++
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/ratis/RatisHelper.java
@@ -34,7 +34,7 @@
import java.util.function.BooleanSupplier;
import java.util.stream.Collectors;
import javax.net.ssl.TrustManager;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.StringUtils;
import org.apache.hadoop.hdds.conf.ConfigurationSource;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
@@ -450,8 +450,8 @@ public static Long getMinReplicatedIndex(
private static boolean datanodeUseHostName() {
return CONF.getBoolean(
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME,
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME_DEFAULT);
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME,
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME_DEFAULT);
}
private static <U> Class<? extends U> getClass(String name,
diff --git
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfigKeys.java
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfigKeys.java
index cda2133269..941df45c2d 100644
---
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfigKeys.java
+++
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfigKeys.java
@@ -629,6 +629,10 @@ public final class ScmConfigKeys {
"ozone.scm.ha.dbtransactionbuffer.flush.interval";
public static final long
OZONE_SCM_HA_DBTRANSACTIONBUFFER_FLUSH_INTERVAL_DEFAULT = 600 * 1000L;
+
+ public static final String NET_TOPOLOGY_NODE_SWITCH_MAPPING_IMPL_KEY =
+ "net.topology.node.switch.mapping.impl";
+
/**
* Never constructed.
*/
diff --git
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/OzoneConfigKeys.java
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/OzoneConfigKeys.java
index dfa5dc41c2..5dc4e15fc0 100644
---
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/OzoneConfigKeys.java
+++
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/OzoneConfigKeys.java
@@ -595,7 +595,6 @@ public final class OzoneConfigKeys {
OZONE_CLIENT_BUCKET_REPLICATION_CONFIG_REFRESH_PERIOD_DEFAULT_MS =
300 * 1000;
-
// Values for bucket layout configurations.
public static final String OZONE_BUCKET_LAYOUT_LEGACY =
"LEGACY";
diff --git
a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/HddsDatanodeService.java
b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/HddsDatanodeService.java
index 50b2f9844b..5df8a23505 100644
---
a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/HddsDatanodeService.java
+++
b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/HddsDatanodeService.java
@@ -44,8 +44,8 @@
import java.util.concurrent.atomic.AtomicBoolean;
import javax.management.ObjectName;
import org.apache.hadoop.conf.Configurable;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
import org.apache.hadoop.hdds.DatanodeVersion;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.HddsUtils;
import org.apache.hadoop.hdds.cli.GenericCli;
import org.apache.hadoop.hdds.cli.HddsVersionProvider;
@@ -248,16 +248,16 @@ public String getNamespace() {
UserGroupInformation.AuthenticationMethod.KERBEROS)) {
LOG.info("Ozone security is enabled. Attempting login for Hdds " +
"Datanode user. Principal: {},keytab: {}", conf.get(
- DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY),
+ HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY),
conf.get(
- DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY));
+ HddsConfigKeys.HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY));
UserGroupInformation.setConfiguration(conf);
SecurityUtil
.login(conf,
- DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY,
- DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY,
+ HddsConfigKeys.HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY,
+ HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY,
hostname);
} else {
throw new AuthenticationException(SecurityUtil.
diff --git
a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/helpers/ContainerMetrics.java
b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/helpers/ContainerMetrics.java
index 8c620f45e7..d37a8626cf 100644
---
a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/helpers/ContainerMetrics.java
+++
b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/helpers/ContainerMetrics.java
@@ -19,7 +19,7 @@
import java.io.Closeable;
import java.util.EnumMap;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.annotation.InterfaceAudience;
import org.apache.hadoop.hdds.conf.ConfigurationSource;
import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos;
@@ -94,7 +94,7 @@ public static ContainerMetrics create(ConfigurationSource
conf) {
MetricsSystem ms = DefaultMetricsSystem.instance();
// Percentile measurement is off by default, by watching no intervals
int[] intervals =
-
conf.getInts(DFSConfigKeysLegacy.DFS_METRICS_PERCENTILES_INTERVALS_KEY);
+ conf.getInts(HddsConfigKeys.HDDS_METRICS_PERCENTILES_INTERVALS_KEY);
return ms.register(STORAGE_CONTAINER_METRICS,
"Storage Container Node Metrics",
new ContainerMetrics(intervals));
diff --git
a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocolPB/ReconDatanodeProtocolPB.java
b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocolPB/ReconDatanodeProtocolPB.java
index 859e595d51..d55587c457 100644
---
a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocolPB/ReconDatanodeProtocolPB.java
+++
b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocolPB/ReconDatanodeProtocolPB.java
@@ -19,7 +19,7 @@
import static
org.apache.hadoop.hdds.recon.ReconConfig.ConfigStrings.OZONE_RECON_KERBEROS_PRINCIPAL_KEY;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.ipc.ProtocolInfo;
import org.apache.hadoop.security.KerberosInfo;
@@ -32,7 +32,7 @@
protocolVersion = 1)
@KerberosInfo(
serverPrincipal = OZONE_RECON_KERBEROS_PRINCIPAL_KEY,
- clientPrincipal = DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY)
+ clientPrincipal = HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY)
public interface ReconDatanodeProtocolPB extends
StorageContainerDatanodeProtocolPB {
}
diff --git
a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocolPB/StorageContainerDatanodeProtocolPB.java
b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocolPB/StorageContainerDatanodeProtocolPB.java
index a6263ea706..c20a90fe3a 100644
---
a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocolPB/StorageContainerDatanodeProtocolPB.java
+++
b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocolPB/StorageContainerDatanodeProtocolPB.java
@@ -17,7 +17,7 @@
package org.apache.hadoop.ozone.protocolPB;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import
org.apache.hadoop.hdds.protocol.proto.StorageContainerDatanodeProtocolProtos.StorageContainerDatanodeProtocolService;
import org.apache.hadoop.hdds.scm.ScmConfig;
import org.apache.hadoop.ipc.ProtocolInfo;
@@ -33,7 +33,7 @@
protocolVersion = 1)
@KerberosInfo(
serverPrincipal = ScmConfig.ConfigStrings.HDDS_SCM_KERBEROS_PRINCIPAL_KEY,
- clientPrincipal = DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY)
+ clientPrincipal = HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY)
public interface StorageContainerDatanodeProtocolPB extends
StorageContainerDatanodeProtocolService.BlockingInterface {
}
diff --git
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/TestHddsDatanodeService.java
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/TestHddsDatanodeService.java
index b9141a68b5..7547036a59 100644
---
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/TestHddsDatanodeService.java
+++
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/TestHddsDatanodeService.java
@@ -34,7 +34,6 @@
import java.util.ArrayList;
import java.util.List;
import java.util.UUID;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.hadoop.hdds.scm.ScmConfigKeys;
@@ -99,7 +98,7 @@ public void setUp() throws IOException {
conf.setBoolean(HDDS_CONTAINER_TOKEN_ENABLED, true);
String volumeDir = testDir + OZONE_URI_DELIMITER + "disk1";
- conf.set(DFSConfigKeysLegacy.DFS_DATANODE_DATA_DIR_KEY, volumeDir);
+ conf.set(HddsConfigKeys.HDDS_DATANODE_DATA_DIR_KEY, volumeDir);
}
@ParameterizedTest
diff --git
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/TestHddsSecureDatanodeInit.java
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/TestHddsSecureDatanodeInit.java
index addab7902f..ec64da7637 100644
---
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/TestHddsSecureDatanodeInit.java
+++
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/TestHddsSecureDatanodeInit.java
@@ -47,7 +47,6 @@
import java.util.List;
import java.util.concurrent.Callable;
import org.apache.commons.io.FileUtils;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.hadoop.hdds.protocol.DatanodeDetails;
@@ -98,7 +97,7 @@ public static void setUp() throws Exception {
conf.set(HddsConfigKeys.OZONE_METADATA_DIRS, testDir.getPath());
//conf.set(ScmConfigKeys.OZONE_SCM_NAMES, "localhost");
String volumeDir = testDir + "/disk1";
- conf.set(DFSConfigKeysLegacy.DFS_DATANODE_DATA_DIR_KEY, volumeDir);
+ conf.set(HddsConfigKeys.HDDS_DATANODE_DATA_DIR_KEY, volumeDir);
conf.setBoolean(OZONE_SECURITY_ENABLED_KEY, true);
conf.setClass(OzoneConfigKeys.HDDS_DATANODE_PLUGINS_KEY,
diff --git
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestVolumeSet.java
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestVolumeSet.java
index 4614ee952b..ab99d5f883 100644
---
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestVolumeSet.java
+++
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestVolumeSet.java
@@ -34,7 +34,6 @@
import java.util.List;
import java.util.UUID;
import org.apache.commons.io.FileUtils;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.hadoop.ozone.OzoneConfigKeys;
@@ -77,7 +76,7 @@ public void setup() throws Exception {
String dataDirKey = volume1 + "," + volume2;
volumes.add(volume1);
volumes.add(volume2);
- conf.set(DFSConfigKeysLegacy.DFS_DATANODE_DATA_DIR_KEY, dataDirKey);
+ conf.set(HddsConfigKeys.HDDS_DATANODE_DATA_DIR_KEY, dataDirKey);
conf.set(OzoneConfigKeys.HDDS_CONTAINER_RATIS_DATANODE_STORAGE_DIR,
dataDirKey);
initializeVolumeSet();
diff --git
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestVolumeSetDiskChecks.java
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestVolumeSetDiskChecks.java
index 6d073fdd55..4f1838ce9f 100644
---
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestVolumeSetDiskChecks.java
+++
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestVolumeSetDiskChecks.java
@@ -41,7 +41,6 @@
import java.util.Set;
import java.util.UUID;
import org.apache.commons.io.FileUtils;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.conf.ConfigurationSource;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
@@ -91,7 +90,7 @@ public class TestVolumeSetDiskChecks {
@AfterEach
public void cleanup() {
final Collection<String> dirs = conf.getTrimmedStringCollection(
- DFSConfigKeysLegacy.DFS_DATANODE_DATA_DIR_KEY);
+ HddsConfigKeys.HDDS_DATANODE_DATA_DIR_KEY);
for (String d: dirs) {
FileUtils.deleteQuietly(new File(d));
@@ -116,7 +115,7 @@ public void testOzoneDirsAreCreated() throws IOException {
// Verify that the Ozone dirs were created during initialization.
Collection<String> dirs = conf.getTrimmedStringCollection(
- DFSConfigKeysLegacy.DFS_DATANODE_DATA_DIR_KEY);
+ HddsConfigKeys.HDDS_DATANODE_DATA_DIR_KEY);
for (String d : dirs) {
assertTrue(new File(d).isDirectory());
}
@@ -223,7 +222,7 @@ private OzoneConfiguration getConfWithDataNodeDirs(int
numDirs) {
for (int i = 0; i < numDirs; ++i) {
dirs.add(new File(dir, randomAlphanumeric(10)).toString());
}
- ozoneConf.set(DFSConfigKeysLegacy.DFS_DATANODE_DATA_DIR_KEY,
+ ozoneConf.set(HddsConfigKeys.HDDS_DATANODE_DATA_DIR_KEY,
String.join(",", dirs));
final List<String> metaDirs = new ArrayList<>();
diff --git
a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocol/SecretKeyProtocolDatanode.java
b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocol/SecretKeyProtocolDatanode.java
index 8857da07f2..66af07484a 100644
---
a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocol/SecretKeyProtocolDatanode.java
+++
b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocol/SecretKeyProtocolDatanode.java
@@ -17,7 +17,7 @@
package org.apache.hadoop.hdds.protocol;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY;
import static
org.apache.hadoop.hdds.scm.ScmConfig.ConfigStrings.HDDS_SCM_KERBEROS_PRINCIPAL_KEY;
import org.apache.hadoop.hdds.annotation.InterfaceAudience;
@@ -28,7 +28,7 @@
*/
@KerberosInfo(
serverPrincipal = HDDS_SCM_KERBEROS_PRINCIPAL_KEY,
- clientPrincipal = DFS_DATANODE_KERBEROS_PRINCIPAL_KEY
+ clientPrincipal = HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY
)
@InterfaceAudience.Private
public interface SecretKeyProtocolDatanode extends SecretKeyProtocol {
diff --git
a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocolPB/ReconfigureProtocolDatanodePB.java
b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocolPB/ReconfigureProtocolDatanodePB.java
index a7f3992da1..78b2c221e4 100644
---
a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocolPB/ReconfigureProtocolDatanodePB.java
+++
b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocolPB/ReconfigureProtocolDatanodePB.java
@@ -17,7 +17,7 @@
package org.apache.hadoop.hdds.protocolPB;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.ipc.ProtocolInfo;
import org.apache.hadoop.security.KerberosInfo;
@@ -28,6 +28,6 @@
@ProtocolInfo(
protocolName = "org.apache.hadoop.hdds.protocol.ReconfigureProtocol",
protocolVersion = 1)
-@KerberosInfo(serverPrincipal =
DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY)
+@KerberosInfo(serverPrincipal =
HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY)
public interface ReconfigureProtocolDatanodePB extends ReconfigureProtocolPB {
}
diff --git
a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocolPB/SecretKeyProtocolDatanodePB.java
b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocolPB/SecretKeyProtocolDatanodePB.java
index d391127eb1..fc07daaf43 100644
---
a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocolPB/SecretKeyProtocolDatanodePB.java
+++
b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocolPB/SecretKeyProtocolDatanodePB.java
@@ -17,7 +17,7 @@
package org.apache.hadoop.hdds.protocolPB;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY;
import static
org.apache.hadoop.hdds.scm.ScmConfig.ConfigStrings.HDDS_SCM_KERBEROS_PRINCIPAL_KEY;
import
org.apache.hadoop.hdds.protocol.proto.SCMSecretKeyProtocolProtos.SCMSecretKeyProtocolService;
@@ -33,7 +33,7 @@
protocolVersion = 1)
@KerberosInfo(
serverPrincipal = HDDS_SCM_KERBEROS_PRINCIPAL_KEY,
- clientPrincipal = DFS_DATANODE_KERBEROS_PRINCIPAL_KEY
+ clientPrincipal = HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY
)
public interface SecretKeyProtocolDatanodePB extends
SCMSecretKeyProtocolService.BlockingInterface {
diff --git
a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/http/BaseHttpServer.java
b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/http/BaseHttpServer.java
index 1ff3e53452..3144fd63ac 100644
---
a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/http/BaseHttpServer.java
+++
b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/http/BaseHttpServer.java
@@ -44,7 +44,6 @@
import javax.servlet.http.HttpServlet;
import org.apache.commons.lang3.StringUtils;
import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.conf.ConfigurationSource;
import org.apache.hadoop.hdds.conf.HddsConfServlet;
@@ -128,12 +127,12 @@ public BaseHttpServer(MutableConfigurationSource conf,
String name)
}
final boolean xFrameEnabled = conf.getBoolean(
- DFSConfigKeysLegacy.DFS_XFRAME_OPTION_ENABLED,
- DFSConfigKeysLegacy.DFS_XFRAME_OPTION_ENABLED_DEFAULT);
+ HddsConfigKeys.HDDS_XFRAME_OPTION_ENABLED,
+ HddsConfigKeys.HDDS_XFRAME_OPTION_ENABLED_DEFAULT);
final String xFrameOptionValue = conf.getTrimmed(
- DFSConfigKeysLegacy.DFS_XFRAME_OPTION_VALUE,
- DFSConfigKeysLegacy.DFS_XFRAME_OPTION_VALUE_DEFAULT);
+ HddsConfigKeys.HDDS_XFRAME_OPTION_VALUE,
+ HddsConfigKeys.HDDS_XFRAME_OPTION_VALUE_DEFAULT);
builder.configureXFrame(xFrameEnabled).setXFrameOption(xFrameOptionValue);
diff --git
a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/HddsServerUtil.java
b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/HddsServerUtil.java
index 629dbbcf54..f58887ad61 100644
---
a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/HddsServerUtil.java
+++
b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/HddsServerUtil.java
@@ -17,7 +17,7 @@
package org.apache.hadoop.hdds.utils;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_DATA_DIR_KEY;
+import static org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_DATA_DIR_KEY;
import static org.apache.hadoop.hdds.HddsConfigKeys.HDDS_HEARTBEAT_INTERVAL;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_HEARTBEAT_INTERVAL_DEFAULT;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_RECON_HEARTBEAT_INTERVAL;
@@ -67,7 +67,7 @@
import org.apache.commons.compress.utils.IOUtils;
import org.apache.commons.lang3.SystemUtils;
import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.HddsUtils;
import org.apache.hadoop.hdds.conf.ConfigurationSource;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
@@ -409,11 +409,11 @@ public static Collection<String> getDatanodeStorageDirs(
Collection<String> rawLocations = conf.getTrimmedStringCollection(
HDDS_DATANODE_DIR_KEY);
if (rawLocations.isEmpty()) {
- rawLocations =
conf.getTrimmedStringCollection(DFS_DATANODE_DATA_DIR_KEY);
+ rawLocations =
conf.getTrimmedStringCollection(HDDS_DATANODE_DATA_DIR_KEY);
}
if (rawLocations.isEmpty()) {
throw new IllegalArgumentException("No location configured in either "
- + HDDS_DATANODE_DIR_KEY + " or " + DFS_DATANODE_DATA_DIR_KEY);
+ + HDDS_DATANODE_DIR_KEY + " or " + HDDS_DATANODE_DATA_DIR_KEY);
}
return rawLocations;
}
@@ -580,7 +580,7 @@ public static MetricsSystem initializeMetrics(
MetricsSystem metricsSystem = DefaultMetricsSystem.initialize(serverName);
try {
JvmMetrics.create(serverName,
- configuration.get(DFSConfigKeysLegacy.DFS_METRICS_SESSION_ID_KEY),
+ configuration.get(HddsConfigKeys.HDDS_METRICS_SESSION_ID_KEY),
DefaultMetricsSystem.instance());
CpuMetrics.create();
} catch (MetricsException e) {
diff --git
a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeDecommissionManager.java
b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeDecommissionManager.java
index 3ba7ff2fef..a4deae441f 100644
---
a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeDecommissionManager.java
+++
b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeDecommissionManager.java
@@ -35,7 +35,7 @@
import java.util.concurrent.TimeUnit;
import java.util.stream.Collectors;
import org.apache.commons.lang3.tuple.Pair;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.client.ECReplicationConfig;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.hadoop.hdds.protocol.DatanodeDetails;
@@ -273,8 +273,8 @@ public NodeDecommissionManager(OzoneConfiguration config,
NodeManager nm, Contai
);
useHostnames = config.getBoolean(
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME,
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME_DEFAULT);
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME,
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME_DEFAULT);
long monitorInterval = config.getTimeDuration(
ScmConfigKeys.OZONE_SCM_DATANODE_ADMIN_MONITOR_INTERVAL,
diff --git
a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/SCMNodeManager.java
b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/SCMNodeManager.java
index cbeafa4dbd..3efaf4bba4 100644
---
a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/SCMNodeManager.java
+++
b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/SCMNodeManager.java
@@ -50,7 +50,7 @@
import java.util.function.Function;
import java.util.stream.Collectors;
import javax.management.ObjectName;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.hadoop.hdds.protocol.DatanodeDetails;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos.NodeOperationalState;
@@ -182,8 +182,8 @@ public SCMNodeManager(
this.clusterMap = networkTopology;
this.nodeResolver = nodeResolver;
this.useHostname = conf.getBoolean(
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME,
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME_DEFAULT);
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME,
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME_DEFAULT);
this.numPipelinesPerMetadataVolume =
conf.getInt(ScmConfigKeys.OZONE_SCM_PIPELINE_PER_METADATA_VOLUME,
ScmConfigKeys.OZONE_SCM_PIPELINE_PER_METADATA_VOLUME_DEFAULT);
diff --git
a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/StorageContainerManager.java
b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/StorageContainerManager.java
index 7e898acc0e..71a08b3240 100644
---
a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/StorageContainerManager.java
+++
b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/StorageContainerManager.java
@@ -57,7 +57,6 @@
import javax.management.ObjectName;
import org.apache.commons.lang3.tuple.Pair;
import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.HddsUtils;
import org.apache.hadoop.hdds.annotation.InterfaceAudience;
@@ -709,7 +708,7 @@ private void initializeSystemManagers(OzoneConfiguration
conf,
Class<? extends DNSToSwitchMapping> dnsToSwitchMappingClass =
conf.getClass(
- DFSConfigKeysLegacy.NET_TOPOLOGY_NODE_SWITCH_MAPPING_IMPL_KEY,
+ ScmConfigKeys.NET_TOPOLOGY_NODE_SWITCH_MAPPING_IMPL_KEY,
TableMapping.class, DNSToSwitchMapping.class);
DNSToSwitchMapping newInstance = ReflectionUtils.newInstance(
dnsToSwitchMappingClass, conf);
diff --git
a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestSCMNodeManager.java
b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestSCMNodeManager.java
index 318e68983c..25802ddb81 100644
---
a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestSCMNodeManager.java
+++
b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestSCMNodeManager.java
@@ -69,7 +69,6 @@
import java.util.stream.Collectors;
import java.util.stream.Stream;
import org.apache.hadoop.fs.FileUtil;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.client.RatisReplicationConfig;
import org.apache.hadoop.hdds.client.ReplicationConfig;
@@ -1828,7 +1827,7 @@ void testScmRegisterNodeWithNetworkTopology(boolean
useHostname)
OzoneConfiguration conf = getConf();
conf.setTimeDuration(OZONE_SCM_HEARTBEAT_PROCESS_INTERVAL, 1000,
MILLISECONDS);
- conf.setBoolean(DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME,
+ conf.setBoolean(HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME,
useHostname);
// create table mapping file
@@ -1942,7 +1941,7 @@ void testGetNodesByAddress(boolean useHostname)
OzoneConfiguration conf = getConf();
conf.setTimeDuration(OZONE_SCM_HEARTBEAT_PROCESS_INTERVAL, 1000,
MILLISECONDS);
- conf.setBoolean(DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME,
+ conf.setBoolean(HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME,
useHostname);
// create a set of hosts - note two hosts on "host1"
diff --git
a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/server/TestSCMBlockProtocolServer.java
b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/server/TestSCMBlockProtocolServer.java
index d94c7e5a11..895baef27d 100644
---
a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/server/TestSCMBlockProtocolServer.java
+++
b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/server/TestSCMBlockProtocolServer.java
@@ -38,7 +38,7 @@
import java.util.concurrent.ThreadLocalRandom;
import java.util.concurrent.TimeoutException;
import java.util.stream.Collectors;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.client.ContainerBlockID;
import org.apache.hadoop.hdds.client.RatisReplicationConfig;
import org.apache.hadoop.hdds.client.ReplicationConfig;
@@ -346,8 +346,8 @@ private List<String> getNetworkNames() {
private String nodeAddress(DatanodeDetails dn) {
boolean useHostname = config.getBoolean(
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME,
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME_DEFAULT);
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME,
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME_DEFAULT);
return useHostname ? dn.getHostName() : dn.getIpAddress();
}
}
diff --git
a/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/metrics/HttpFSServerMetrics.java
b/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/metrics/HttpFSServerMetrics.java
index ea3dd0c5f3..5b2d28b02c 100644
---
a/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/metrics/HttpFSServerMetrics.java
+++
b/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/metrics/HttpFSServerMetrics.java
@@ -17,6 +17,7 @@
package org.apache.ozone.fs.http.server.metrics;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_METRICS_SESSION_ID_KEY;
import static org.apache.hadoop.metrics2.impl.MsInfo.SessionId;
import java.util.concurrent.ThreadLocalRandom;
@@ -46,8 +47,6 @@
@Metrics(about = "HttpFSServer metrics", context = "httpfs")
public class HttpFSServerMetrics {
- private static final String DFS_METRICS_SESSION_ID_KEY
- = "hdds.metrics.session-id";
private @Metric MutableCounterLong bytesWritten;
private @Metric MutableCounterLong bytesRead;
@@ -78,7 +77,7 @@ public HttpFSServerMetrics(String name, String sessionId,
public static HttpFSServerMetrics create(Configuration conf,
String serverName) {
- String sessionId = conf.get(DFS_METRICS_SESSION_ID_KEY);
+ String sessionId = conf.get(HDDS_METRICS_SESSION_ID_KEY);
MetricsSystem ms = DefaultMetricsSystem.instance();
JvmMetrics jm = JvmMetrics.create("HttpFSServer", sessionId, ms);
String name = "ServerActivity-" + (serverName.isEmpty()
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSecretKeySnapshot.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSecretKeySnapshot.java
index 17dc48403d..384c4e3319 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSecretKeySnapshot.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSecretKeySnapshot.java
@@ -18,9 +18,9 @@
package org.apache.hadoop.hdds.scm;
import static
org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHENTICATION;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY;
import static org.apache.hadoop.hdds.HddsConfigKeys.HDDS_BLOCK_TOKEN_ENABLED;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_SECRET_KEY_EXPIRY_DURATION;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_SECRET_KEY_ROTATE_CHECK_DURATION;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_SECRET_KEY_ROTATE_DURATION;
@@ -172,7 +172,7 @@ private void setSecureConfig() throws IOException {
conf.set(HDDS_SCM_HTTP_KERBEROS_PRINCIPAL_KEY, "HTTP_SCM/" + hostAndRealm);
conf.set(OZONE_OM_KERBEROS_PRINCIPAL_KEY, "scm/" + hostAndRealm);
conf.set(OZONE_OM_HTTP_KERBEROS_PRINCIPAL_KEY, "HTTP_OM/" + hostAndRealm);
- conf.set(DFS_DATANODE_KERBEROS_PRINCIPAL_KEY, "scm/" + hostAndRealm);
+ conf.set(HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY, "scm/" + hostAndRealm);
ozoneKeytab = new File(workDir, "scm.keytab");
spnegoKeytab = new File(workDir, "http.keytab");
@@ -185,7 +185,7 @@ private void setSecureConfig() throws IOException {
ozoneKeytab.getAbsolutePath());
conf.set(OZONE_OM_HTTP_KERBEROS_KEYTAB_FILE,
spnegoKeytab.getAbsolutePath());
- conf.set(DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY,
+ conf.set(HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY,
ozoneKeytab.getAbsolutePath());
conf.setBoolean(HDDS_BLOCK_TOKEN_ENABLED, true);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSecretKeysApi.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSecretKeysApi.java
index 3550ff1304..0e069e1088 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSecretKeysApi.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSecretKeysApi.java
@@ -19,9 +19,9 @@
import static
org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHENTICATION;
import static
org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHORIZATION;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY;
import static org.apache.hadoop.hdds.HddsConfigKeys.HDDS_BLOCK_TOKEN_ENABLED;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_SECRET_KEY_EXPIRY_DURATION;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_SECRET_KEY_ROTATE_CHECK_DURATION;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_SECRET_KEY_ROTATE_DURATION;
@@ -155,7 +155,7 @@ private void setSecureConfig() throws IOException {
conf.set(HDDS_SCM_HTTP_KERBEROS_PRINCIPAL_KEY, "HTTP_SCM/" + hostAndRealm);
conf.set(OZONE_OM_KERBEROS_PRINCIPAL_KEY, ozonePrincipal);
conf.set(OZONE_OM_HTTP_KERBEROS_PRINCIPAL_KEY, "HTTP_OM/" + hostAndRealm);
- conf.set(DFS_DATANODE_KERBEROS_PRINCIPAL_KEY, ozonePrincipal);
+ conf.set(HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY, ozonePrincipal);
ozoneKeytab = new File(workDir, "scm.keytab");
spnegoKeytab = new File(workDir, "http.keytab");
@@ -170,7 +170,7 @@ private void setSecureConfig() throws IOException {
ozoneKeytab.getAbsolutePath());
conf.set(OZONE_OM_HTTP_KERBEROS_KEYTAB_FILE,
spnegoKeytab.getAbsolutePath());
- conf.set(DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY,
+ conf.set(HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY,
ozoneKeytab.getAbsolutePath());
conf.setBoolean(HADOOP_SECURITY_AUTHORIZATION, true);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokens.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokens.java
index 50c99acb33..752d2894da 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokens.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokens.java
@@ -19,10 +19,10 @@
import static java.util.Objects.requireNonNull;
import static
org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHENTICATION;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY;
import static org.apache.hadoop.hdds.HddsConfigKeys.HDDS_BLOCK_TOKEN_ENABLED;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_CONTAINER_TOKEN_ENABLED;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_SECRET_KEY_EXPIRY_DURATION;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_SECRET_KEY_ROTATE_CHECK_DURATION;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_SECRET_KEY_ROTATE_DURATION;
@@ -354,7 +354,7 @@ private static void setSecureConfig() throws IOException {
conf.set(HDDS_SCM_HTTP_KERBEROS_PRINCIPAL_KEY, "HTTP_SCM/" + hostAndRealm);
conf.set(OZONE_OM_KERBEROS_PRINCIPAL_KEY, "scm/" + hostAndRealm);
conf.set(OZONE_OM_HTTP_KERBEROS_PRINCIPAL_KEY, "HTTP_OM/" + hostAndRealm);
- conf.set(DFS_DATANODE_KERBEROS_PRINCIPAL_KEY, "scm/" + hostAndRealm);
+ conf.set(HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY, "scm/" + hostAndRealm);
ozoneKeytab = new File(workDir, "scm.keytab");
spnegoKeytab = new File(workDir, "http.keytab");
@@ -369,7 +369,7 @@ private static void setSecureConfig() throws IOException {
ozoneKeytab.getAbsolutePath());
conf.set(OZONE_OM_HTTP_KERBEROS_KEYTAB_FILE,
spnegoKeytab.getAbsolutePath());
- conf.set(DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY,
+ conf.set(HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY,
ozoneKeytab.getAbsolutePath());
}
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokensCLI.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokensCLI.java
index 01ab556c28..393009d18e 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokensCLI.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokensCLI.java
@@ -19,10 +19,10 @@
import static java.time.Duration.between;
import static
org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHENTICATION;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_KERBEROS_PRINCIPAL_KEY;
import static org.apache.hadoop.hdds.HddsConfigKeys.HDDS_BLOCK_TOKEN_ENABLED;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_CONTAINER_TOKEN_ENABLED;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY;
+import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY;
import static
org.apache.hadoop.hdds.scm.ScmConfig.ConfigStrings.HDDS_SCM_KERBEROS_KEYTAB_FILE_KEY;
import static
org.apache.hadoop.hdds.scm.ScmConfig.ConfigStrings.HDDS_SCM_KERBEROS_PRINCIPAL_KEY;
import static
org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_CLIENT_ADDRESS_KEY;
@@ -171,7 +171,7 @@ private static void setSecureConfig() throws IOException {
conf.set(HDDS_SCM_HTTP_KERBEROS_PRINCIPAL_KEY, "HTTP_SCM/" + hostAndRealm);
conf.set(OZONE_OM_KERBEROS_PRINCIPAL_KEY, "scm/" + hostAndRealm);
conf.set(OZONE_OM_HTTP_KERBEROS_PRINCIPAL_KEY, "HTTP_OM/" + hostAndRealm);
- conf.set(DFS_DATANODE_KERBEROS_PRINCIPAL_KEY, "scm/" + hostAndRealm);
+ conf.set(HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY, "scm/" + hostAndRealm);
ozoneKeytab = new File(workDir, "scm.keytab");
spnegoKeytab = new File(workDir, "http.keytab");
@@ -184,7 +184,7 @@ private static void setSecureConfig() throws IOException {
ozoneKeytab.getAbsolutePath());
conf.set(OZONE_OM_HTTP_KERBEROS_KEYTAB_FILE,
spnegoKeytab.getAbsolutePath());
- conf.set(DFS_DATANODE_KERBEROS_KEYTAB_FILE_KEY,
+ conf.set(HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY,
ozoneKeytab.getAbsolutePath());
}
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestOMSortDatanodes.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestOMSortDatanodes.java
index a7126bcb20..c7f7a843fe 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestOMSortDatanodes.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestOMSortDatanodes.java
@@ -29,7 +29,6 @@
import java.util.ArrayList;
import java.util.List;
import java.util.Map;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.hadoop.hdds.protocol.DatanodeDetails;
@@ -183,8 +182,8 @@ private static void assertRackOrder(String rack,
List<DatanodeDetails> list) {
private String nodeAddress(DatanodeDetails dn) {
boolean useHostname = config.getBoolean(
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME,
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME_DEFAULT);
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME,
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME_DEFAULT);
return useHostname ? dn.getHostName() : dn.getIpAddress();
}
}
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestOzoneConfigurationFields.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestOzoneConfigurationFields.java
index 9c040f9aa0..63afb9aed6 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestOzoneConfigurationFields.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestOzoneConfigurationFields.java
@@ -132,7 +132,19 @@ private void addPropertiesNotInXml() {
HddsConfigKeys.HDDS_DATANODE_VOLUME_MIN_FREE_SPACE_PERCENT,
OzoneConfigKeys.HDDS_SCM_CLIENT_RPC_TIME_OUT,
OzoneConfigKeys.HDDS_SCM_CLIENT_MAX_RETRY_TIMEOUT,
- OzoneConfigKeys.HDDS_SCM_CLIENT_FAILOVER_MAX_RETRY
+ OzoneConfigKeys.HDDS_SCM_CLIENT_FAILOVER_MAX_RETRY,
+ HddsConfigKeys.HDDS_DATANODE_DNS_INTERFACE_KEY,
+ HddsConfigKeys.HDDS_DATANODE_DNS_NAMESERVER_KEY,
+ HddsConfigKeys.HDDS_DATANODE_HOST_NAME_KEY,
+ HddsConfigKeys.HDDS_DATANODE_DATA_DIR_KEY,
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME,
+ HddsConfigKeys.HDDS_XFRAME_OPTION_ENABLED,
+ HddsConfigKeys.HDDS_XFRAME_OPTION_VALUE,
+ HddsConfigKeys.HDDS_METRICS_SESSION_ID_KEY,
+ ScmConfigKeys.NET_TOPOLOGY_NODE_SWITCH_MAPPING_IMPL_KEY,
+ HddsConfigKeys.HDDS_DATANODE_KERBEROS_PRINCIPAL_KEY,
+ HddsConfigKeys.HDDS_DATANODE_KERBEROS_KEYTAB_FILE_KEY,
+ HddsConfigKeys.HDDS_METRICS_PERCENTILES_INTERVALS_KEY
));
}
}
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/UniformDatanodesFactory.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/UniformDatanodesFactory.java
index d143baa0fc..e9672bc601 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/UniformDatanodesFactory.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/UniformDatanodesFactory.java
@@ -17,8 +17,8 @@
package org.apache.hadoop.ozone;
-import static
org.apache.hadoop.hdds.DFSConfigKeysLegacy.DFS_DATANODE_DATA_DIR_KEY;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_CLIENT_ADDRESS_KEY;
+import static org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_DATA_DIR_KEY;
import static
org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DATANODE_HTTP_ADDRESS_KEY;
import static org.apache.hadoop.hdds.HddsConfigKeys.OZONE_METADATA_DIRS;
import static
org.apache.hadoop.hdds.scm.ScmConfigKeys.HDDS_DATANODE_DIR_DU_RESERVED;
@@ -96,7 +96,7 @@ public OzoneConfiguration apply(OzoneConfiguration conf)
throws IOException {
}
String reservedSpaceString = String.join(",", reservedSpaceList);
String listOfDirs = String.join(",", dataDirs);
- dnConf.set(DFS_DATANODE_DATA_DIR_KEY, listOfDirs);
+ dnConf.set(HDDS_DATANODE_DATA_DIR_KEY, listOfDirs);
dnConf.set(HDDS_DATANODE_DIR_KEY, listOfDirs);
dnConf.set(HDDS_DATANODE_DIR_DU_RESERVED, reservedSpaceString);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/metrics/TestContainerMetrics.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/metrics/TestContainerMetrics.java
index 34d836a1c9..ebb4cca97e 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/metrics/TestContainerMetrics.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/metrics/TestContainerMetrics.java
@@ -31,7 +31,7 @@
import java.util.Map;
import java.util.UUID;
import org.apache.commons.io.FileUtils;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.client.BlockID;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.hadoop.hdds.protocol.DatanodeDetails;
@@ -89,7 +89,7 @@ public class TestContainerMetrics {
@BeforeAll
public static void setup() {
DefaultMetricsSystem.setMiniClusterMode(true);
- CONF.setInt(DFSConfigKeysLegacy.DFS_METRICS_PERCENTILES_INTERVALS_KEY,
+ CONF.setInt(HddsConfigKeys.HDDS_METRICS_PERCENTILES_INTERVALS_KEY,
DFS_METRICS_PERCENTILES_INTERVALS);
CONF.setBoolean(OzoneConfigKeys.HDDS_CONTAINER_RATIS_DATASTREAM_ENABLED,
false);
CONF.set(OzoneConfigKeys.OZONE_METADATA_DIRS, testDir.toString());
diff --git
a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java
b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java
index 8a735ec20e..fb9e1f6f92 100644
---
a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java
+++
b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java
@@ -103,11 +103,12 @@
import org.apache.hadoop.crypto.key.KeyProviderCryptoExtension;
import
org.apache.hadoop.crypto.key.KeyProviderCryptoExtension.EncryptedKeyVersion;
import org.apache.hadoop.fs.FileEncryptionInfo;
-import org.apache.hadoop.hdds.DFSConfigKeysLegacy;
+import org.apache.hadoop.hdds.HddsConfigKeys;
import org.apache.hadoop.hdds.client.ReplicationConfig;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.hadoop.hdds.protocol.DatanodeDetails;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
+import org.apache.hadoop.hdds.scm.ScmConfigKeys;
import
org.apache.hadoop.hdds.scm.container.common.helpers.ContainerWithPipeline;
import org.apache.hadoop.hdds.scm.net.InnerNode;
import org.apache.hadoop.hdds.scm.net.Node;
@@ -353,7 +354,7 @@ public void start(OzoneConfiguration configuration) {
Class<? extends DNSToSwitchMapping> dnsToSwitchMappingClass =
configuration.getClass(
- DFSConfigKeysLegacy.NET_TOPOLOGY_NODE_SWITCH_MAPPING_IMPL_KEY,
+ ScmConfigKeys.NET_TOPOLOGY_NODE_SWITCH_MAPPING_IMPL_KEY,
TableMapping.class, DNSToSwitchMapping.class);
DNSToSwitchMapping newInstance = ReflectionUtils.newInstance(
dnsToSwitchMappingClass, configuration);
@@ -1960,8 +1961,8 @@ private Node getClientNode(String clientMachine,
List<DatanodeDetails> nodes) {
List<DatanodeDetails> matchingNodes = new ArrayList<>();
boolean useHostname = ozoneManager.getConfiguration().getBoolean(
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME,
- DFSConfigKeysLegacy.DFS_DATANODE_USE_DN_HOSTNAME_DEFAULT);
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME,
+ HddsConfigKeys.HDDS_DATANODE_USE_DN_HOSTNAME_DEFAULT);
for (DatanodeDetails node : nodes) {
if ((useHostname ? node.getHostName() : node.getIpAddress()).equals(
clientMachine)) {
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]