This is an automated email from the ASF dual-hosted git repository.

exceptionfactory pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/nifi.git


The following commit(s) were added to refs/heads/main by this push:
     new 2e90fe1b15 NIFI-15020 Standardized property names in Hadoop, 
Hazelcast, HL7, and HubSpot bundles (#10398)
2e90fe1b15 is described below

commit 2e90fe1b15e49e805cd66f413982b91c0c16cbfa
Author: dan-s1 <[email protected]>
AuthorDate: Thu Oct 16 23:39:25 2025 -0400

    NIFI-15020 Standardized property names in Hadoop, Hazelcast, HL7, and 
HubSpot bundles (#10398)
    
    Signed-off-by: David Handermann <[email protected]>
---
 .../hadoop/CreateHadoopSequenceFile.java           | 10 ++++-
 .../apache/nifi/processors/hadoop/DeleteHDFS.java  | 14 +++++--
 .../nifi/processors/hadoop/GetHDFSFileInfo.java    | 45 +++++++++++----------
 .../apache/nifi/processors/hadoop/ListHDFS.java    | 22 +++++++----
 .../org/apache/nifi/processors/hadoop/PutHDFS.java | 10 ++++-
 .../hashicorp/HashiCorpVaultClientService.java     | 30 ++++++++------
 .../cacheclient/HazelcastMapCacheClient.java       | 17 +++++---
 .../EmbeddedHazelcastCacheManager.java             | 31 ++++++++-------
 .../ExternalHazelcastCacheManager.java             | 46 +++++++++++-----------
 .../IMapBasedHazelcastCacheManager.java            |  9 ++++-
 .../cachemanager/TestHazelcastProcessor.java       |  3 +-
 .../nifi/processors/hl7/ExtractHL7Attributes.java  | 21 ++++++----
 .../apache/nifi/processors/hubspot/GetHubSpot.java | 33 +++++++++-------
 13 files changed, 173 insertions(+), 118 deletions(-)

diff --git 
a/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/CreateHadoopSequenceFile.java
 
b/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/CreateHadoopSequenceFile.java
index b25c3d3291..047c6577f7 100644
--- 
a/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/CreateHadoopSequenceFile.java
+++ 
b/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/CreateHadoopSequenceFile.java
@@ -30,6 +30,7 @@ import org.apache.nifi.components.PropertyDescriptor;
 import org.apache.nifi.flowfile.FlowFile;
 import org.apache.nifi.flowfile.attributes.CoreAttributes;
 import org.apache.nifi.flowfile.attributes.StandardFlowFileMediaType;
+import org.apache.nifi.migration.PropertyConfiguration;
 import org.apache.nifi.processor.ProcessContext;
 import org.apache.nifi.processor.ProcessSession;
 import org.apache.nifi.processor.Relationship;
@@ -86,8 +87,7 @@ public class CreateHadoopSequenceFile extends 
AbstractHadoopProcessor {
 
     // Optional Properties.
     static final PropertyDescriptor COMPRESSION_TYPE = new 
PropertyDescriptor.Builder()
-            .displayName("Compression type")
-            .name("compression type")
+            .name("Compression Type")
             .description("Type of compression to use when creating Sequence 
File")
             .allowableValues(SequenceFile.CompressionType.values())
             .build();
@@ -180,4 +180,10 @@ public class CreateHadoopSequenceFile extends 
AbstractHadoopProcessor {
         }
 
     }
+
+    @Override
+    public void migrateProperties(PropertyConfiguration config) {
+        super.migrateProperties(config);
+        config.renameProperty("compression type", COMPRESSION_TYPE.getName());
+    }
 }
diff --git 
a/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/DeleteHDFS.java
 
b/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/DeleteHDFS.java
index f19faab995..fc09cdf10b 100644
--- 
a/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/DeleteHDFS.java
+++ 
b/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/DeleteHDFS.java
@@ -32,6 +32,7 @@ import org.apache.nifi.components.PropertyDescriptor;
 import org.apache.nifi.components.RequiredPermission;
 import org.apache.nifi.expression.ExpressionLanguageScope;
 import org.apache.nifi.flowfile.FlowFile;
+import org.apache.nifi.migration.PropertyConfiguration;
 import org.apache.nifi.processor.ProcessContext;
 import org.apache.nifi.processor.ProcessSession;
 import org.apache.nifi.processor.Relationship;
@@ -86,8 +87,7 @@ public class DeleteHDFS extends AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor FILE_OR_DIRECTORY = new 
PropertyDescriptor.Builder()
-            .name("file_or_directory")
-            .displayName("Path")
+            .name("Path")
             .description("The HDFS file or directory to delete. A wildcard 
expression may be used to only delete certain files")
             .required(true)
             .addValidator(StandardValidators.NON_EMPTY_VALIDATOR)
@@ -95,8 +95,7 @@ public class DeleteHDFS extends AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor RECURSIVE = new 
PropertyDescriptor.Builder()
-            .name("recursive")
-            .displayName("Recursive")
+            .name("Recursive")
             .description("Remove contents of a non-empty directory 
recursively")
             .allowableValues("true", "false")
             .required(true)
@@ -223,6 +222,13 @@ public class DeleteHDFS extends AbstractHadoopProcessor {
 
     }
 
+    @Override
+    public void migrateProperties(PropertyConfiguration config) {
+        super.migrateProperties(config);
+        config.renameProperty("file_or_directory", 
FILE_OR_DIRECTORY.getName());
+        config.renameProperty("recursive", RECURSIVE.getName());
+    }
+
     protected Relationship getSuccessRelationship() {
         return REL_SUCCESS;
     }
diff --git 
a/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/GetHDFSFileInfo.java
 
b/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/GetHDFSFileInfo.java
index 134062c4dd..3cb3aa8bfe 100644
--- 
a/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/GetHDFSFileInfo.java
+++ 
b/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/GetHDFSFileInfo.java
@@ -49,6 +49,7 @@ import org.apache.nifi.components.ValidationResult;
 import org.apache.nifi.expression.ExpressionLanguageScope;
 import org.apache.nifi.flowfile.FlowFile;
 import org.apache.nifi.flowfile.attributes.CoreAttributes;
+import org.apache.nifi.migration.PropertyConfiguration;
 import org.apache.nifi.processor.ProcessContext;
 import org.apache.nifi.processor.ProcessSession;
 import org.apache.nifi.processor.ProcessorInitializationContext;
@@ -97,8 +98,7 @@ import static 
org.apache.nifi.processors.hadoop.GetHDFSFileInfo.HDFSFileInfoRequ
 public class GetHDFSFileInfo extends AbstractHadoopProcessor {
     public static final String APPLICATION_JSON = "application/json";
     public static final PropertyDescriptor FULL_PATH = new 
PropertyDescriptor.Builder()
-            .displayName("Full path")
-            .name("gethdfsfileinfo-full-path")
+            .name("Full Path")
             .description("A directory to start listing from, or a file's full 
path.")
             .required(true)
             
.expressionLanguageSupported(ExpressionLanguageScope.FLOWFILE_ATTRIBUTES)
@@ -107,8 +107,7 @@ public class GetHDFSFileInfo extends 
AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor RECURSE_SUBDIRS = new 
PropertyDescriptor.Builder()
-            .displayName("Recurse Subdirectories")
-            .name("gethdfsfileinfo-recurse-subdirs")
+            .name("Recurse Subdirectories")
             .description("Indicates whether to list files from subdirectories 
of the HDFS directory")
             .required(true)
             .allowableValues("true", "false")
@@ -117,8 +116,7 @@ public class GetHDFSFileInfo extends 
AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor DIR_FILTER = new 
PropertyDescriptor.Builder()
-            .displayName("Directory Filter")
-            .name("gethdfsfileinfo-dir-filter")
+            .name("Directory Filter")
             .description("Regex. Only directories whose names match the given 
regular expression will be picked up. If not provided, any filter would be 
apply (performance considerations).")
             
.expressionLanguageSupported(ExpressionLanguageScope.FLOWFILE_ATTRIBUTES)
             .required(false)
@@ -126,8 +124,7 @@ public class GetHDFSFileInfo extends 
AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor FILE_FILTER = new 
PropertyDescriptor.Builder()
-            .displayName("File Filter")
-            .name("gethdfsfileinfo-file-filter")
+            .name("File Filter")
             .description("Regex. Only files whose names match the given 
regular expression will be picked up. If not provided, any filter would be 
apply (performance considerations).")
             
.expressionLanguageSupported(ExpressionLanguageScope.FLOWFILE_ATTRIBUTES)
             .required(false)
@@ -135,8 +132,7 @@ public class GetHDFSFileInfo extends 
AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor FILE_EXCLUDE_FILTER = new 
PropertyDescriptor.Builder()
-            .displayName("Exclude Files")
-            .name("gethdfsfileinfo-file-exclude-filter")
+            .name("Exclude Files")
             .description("Regex. Files whose names match the given regular 
expression will not be picked up. If not provided, any filter won't be apply 
(performance considerations).")
             
.expressionLanguageSupported(ExpressionLanguageScope.FLOWFILE_ATTRIBUTES)
             .required(false)
@@ -144,8 +140,7 @@ public class GetHDFSFileInfo extends 
AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor IGNORE_DOTTED_DIRS = new 
PropertyDescriptor.Builder()
-            .displayName("Ignore Dotted Directories")
-            .name("gethdfsfileinfo-ignore-dotted-dirs")
+            .name("Ignore Dotted Directories")
             .description("If true, directories whose names begin with a dot 
(\".\") will be ignored")
             .required(true)
             .addValidator(StandardValidators.BOOLEAN_VALIDATOR)
@@ -154,8 +149,7 @@ public class GetHDFSFileInfo extends 
AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor IGNORE_DOTTED_FILES = new 
PropertyDescriptor.Builder()
-            .displayName("Ignore Dotted Files")
-            .name("gethdfsfileinfo-ignore-dotted-files")
+            .name("Ignore Dotted Files")
             .description("If true, files whose names begin with a dot (\".\") 
will be ignored")
             .required(true)
             .addValidator(StandardValidators.BOOLEAN_VALIDATOR)
@@ -174,8 +168,7 @@ public class GetHDFSFileInfo extends 
AbstractHadoopProcessor {
             "Don't group results. Generate flowfile per each HDFS object.");
 
     public static final PropertyDescriptor GROUPING = new 
PropertyDescriptor.Builder()
-            .displayName("Group Results")
-            .name("gethdfsfileinfo-group")
+            .name("Group Results")
             .description("Groups HDFS objects")
             .required(true)
             .addValidator(StandardValidators.NON_EMPTY_VALIDATOR)
@@ -184,8 +177,7 @@ public class GetHDFSFileInfo extends 
AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor BATCH_SIZE = new 
PropertyDescriptor.Builder()
-            .displayName("Batch Size")
-            .name("gethdfsfileinfo-batch-size")
+            .name("Batch Size")
             .description("Number of records to put into an output flowfile 
when 'Destination' is set to 'Content'"
                     + " and 'Group Results' is set to 'None'")
             .required(false)
@@ -201,8 +193,7 @@ public class GetHDFSFileInfo extends 
AbstractHadoopProcessor {
             "Details of given HDFS object will be stored in a content in JSON 
format");
 
     public static final PropertyDescriptor DESTINATION = new 
PropertyDescriptor.Builder()
-            .displayName("Destination")
-            .name("gethdfsfileinfo-destination")
+            .name("Destination")
             .description("Sets the destination for the resutls. When set to 
'Content', attributes of flowfile won't be used for storing results. ")
             .required(true)
             .addValidator(StandardValidators.NON_EMPTY_VALIDATOR)
@@ -344,6 +335,20 @@ public class GetHDFSFileInfo extends 
AbstractHadoopProcessor {
         }
     }
 
+    @Override
+    public void migrateProperties(PropertyConfiguration config) {
+        super.migrateProperties(config);
+        config.renameProperty("gethdfsfileinfo-full-path", 
FULL_PATH.getName());
+        config.renameProperty("gethdfsfileinfo-recurse-subdirs", 
RECURSE_SUBDIRS.getName());
+        config.renameProperty("gethdfsfileinfo-dir-filter", 
DIR_FILTER.getName());
+        config.renameProperty("gethdfsfileinfo-file-filter", 
FILE_FILTER.getName());
+        config.renameProperty("gethdfsfileinfo-file-exclude-filter", 
FILE_EXCLUDE_FILTER.getName());
+        config.renameProperty("gethdfsfileinfo-ignore-dotted-dirs", 
IGNORE_DOTTED_DIRS.getName());
+        config.renameProperty("gethdfsfileinfo-ignore-dotted-files", 
IGNORE_DOTTED_FILES.getName());
+        config.renameProperty("gethdfsfileinfo-group", GROUPING.getName());
+        config.renameProperty("gethdfsfileinfo-batch-size", 
BATCH_SIZE.getName());
+        config.renameProperty("gethdfsfileinfo-destination", 
DESTINATION.getName());
+    }
 
     /*
      * Walks thru HDFS tree. This method will return null to the main if there 
is no provided path existing.
diff --git 
a/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/ListHDFS.java
 
b/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/ListHDFS.java
index ac500a6658..10e2aa6cc2 100644
--- 
a/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/ListHDFS.java
+++ 
b/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/ListHDFS.java
@@ -38,6 +38,7 @@ import org.apache.nifi.components.ValidationContext;
 import org.apache.nifi.components.ValidationResult;
 import org.apache.nifi.components.state.Scope;
 import org.apache.nifi.components.state.StateMap;
+import org.apache.nifi.migration.PropertyConfiguration;
 import org.apache.nifi.processor.ProcessContext;
 import org.apache.nifi.processor.ProcessSession;
 import org.apache.nifi.processor.Relationship;
@@ -107,8 +108,7 @@ public class ListHDFS extends AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor RECORD_WRITER = new 
PropertyDescriptor.Builder()
-            .name("record-writer")
-            .displayName("Record Writer")
+            .name("Record Writer")
             .description("Specifies the Record Writer to use for creating the 
listing. If not specified, one FlowFile will be created for each "
                     + "entity that is listed. If the Record Writer is 
specified, all entities will be written to a single FlowFile.")
             .required(false)
@@ -124,8 +124,7 @@ public class ListHDFS extends AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor FILE_FILTER_MODE = new 
PropertyDescriptor.Builder()
-            .name("file-filter-mode")
-            .displayName("File Filter Mode")
+            .name("File Filter Mode")
             .description("Determines how the regular expression in  " + 
FILE_FILTER.getDisplayName() + " will be used when retrieving listings.")
             .required(true)
             .allowableValues(FilterMode.class)
@@ -134,8 +133,7 @@ public class ListHDFS extends AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor MINIMUM_FILE_AGE = new 
PropertyDescriptor.Builder()
-            .name("minimum-file-age")
-            .displayName("Minimum File Age")
+            .name("Minimum File Age")
             .description("The minimum age that a file must be in order to be 
pulled; any file younger than this "
                     + "amount of time (based on last modification date) will 
be ignored")
             .required(false)
@@ -143,8 +141,7 @@ public class ListHDFS extends AbstractHadoopProcessor {
             .build();
 
     public static final PropertyDescriptor MAXIMUM_FILE_AGE = new 
PropertyDescriptor.Builder()
-            .name("maximum-file-age")
-            .displayName("Maximum File Age")
+            .name("Maximum File Age")
             .description("The maximum age that a file must be in order to be 
pulled; any file older than this "
                     + "amount of time (based on last modification date) will 
be ignored. Minimum value is 100ms.")
             .required(false)
@@ -319,6 +316,15 @@ public class ListHDFS extends AbstractHadoopProcessor {
 
     }
 
+    @Override
+    public void migrateProperties(PropertyConfiguration config) {
+        super.migrateProperties(config);
+        config.renameProperty("record-writer", RECORD_WRITER.getName());
+        config.renameProperty("file-filter-mode", FILE_FILTER_MODE.getName());
+        config.renameProperty("minimum-file-age", MINIMUM_FILE_AGE.getName());
+        config.renameProperty("maximum-file-age", MAXIMUM_FILE_AGE.getName());
+    }
+
     private PathFilter createPathFilter(final ProcessContext context) {
         final FilterMode filterMode = 
FilterMode.forName(context.getProperty(FILE_FILTER_MODE).getValue());
         final boolean recursive = 
context.getProperty(RECURSE_SUBDIRS).asBoolean();
diff --git 
a/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/PutHDFS.java
 
b/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/PutHDFS.java
index 42ffef97f5..7b9ca0b54c 100644
--- 
a/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/PutHDFS.java
+++ 
b/nifi-extension-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/PutHDFS.java
@@ -56,6 +56,7 @@ import org.apache.nifi.expression.ExpressionLanguageScope;
 import org.apache.nifi.fileresource.service.api.FileResource;
 import org.apache.nifi.flowfile.FlowFile;
 import org.apache.nifi.flowfile.attributes.CoreAttributes;
+import org.apache.nifi.migration.PropertyConfiguration;
 import org.apache.nifi.processor.DataUnit;
 import org.apache.nifi.processor.ProcessContext;
 import org.apache.nifi.processor.ProcessSession;
@@ -162,8 +163,7 @@ public class PutHDFS extends AbstractHadoopProcessor {
             .build();
 
     protected static final PropertyDescriptor WRITING_STRATEGY = new 
PropertyDescriptor.Builder()
-            .name("writing-strategy")
-            .displayName("Writing Strategy")
+            .name("Writing Strategy")
             .description("Defines the approach for writing the FlowFile data.")
             .required(true)
             .defaultValue(WRITE_AND_RENAME_AV)
@@ -554,6 +554,12 @@ public class PutHDFS extends AbstractHadoopProcessor {
         });
     }
 
+    @Override
+    public void migrateProperties(PropertyConfiguration config) {
+        super.migrateProperties(config);
+        config.renameProperty("writing-strategy", WRITING_STRATEGY.getName());
+    }
+
     protected Relationship getSuccessRelationship() {
         return REL_SUCCESS;
     }
diff --git 
a/nifi-extension-bundles/nifi-hashicorp-vault-bundle/nifi-hashicorp-vault-client-service-api/src/main/java/org/apache/nifi/vault/hashicorp/HashiCorpVaultClientService.java
 
b/nifi-extension-bundles/nifi-hashicorp-vault-bundle/nifi-hashicorp-vault-client-service-api/src/main/java/org/apache/nifi/vault/hashicorp/HashiCorpVaultClientService.java
index 3858e37116..c4787c3539 100644
--- 
a/nifi-extension-bundles/nifi-hashicorp-vault-bundle/nifi-hashicorp-vault-client-service-api/src/main/java/org/apache/nifi/vault/hashicorp/HashiCorpVaultClientService.java
+++ 
b/nifi-extension-bundles/nifi-hashicorp-vault-bundle/nifi-hashicorp-vault-client-service-api/src/main/java/org/apache/nifi/vault/hashicorp/HashiCorpVaultClientService.java
@@ -23,6 +23,7 @@ import org.apache.nifi.components.resource.ResourceType;
 import org.apache.nifi.controller.ControllerService;
 import org.apache.nifi.controller.VerifiableControllerService;
 import org.apache.nifi.expression.ExpressionLanguageScope;
+import org.apache.nifi.migration.PropertyConfiguration;
 import org.apache.nifi.processor.util.StandardValidators;
 import org.apache.nifi.ssl.SSLContextService;
 
@@ -37,8 +38,7 @@ public interface HashiCorpVaultClientService extends 
ControllerService, Verifiab
             "Use one or more '.properties' files to configure the client");
 
     PropertyDescriptor CONFIGURATION_STRATEGY = new 
PropertyDescriptor.Builder()
-            .displayName("Configuration Strategy")
-            .name("configuration-strategy")
+            .name("Configuration Strategy")
             .required(true)
             .allowableValues(DIRECT_PROPERTIES, PROPERTIES_FILES)
             .defaultValue(DIRECT_PROPERTIES.getValue())
@@ -46,6 +46,7 @@ public interface HashiCorpVaultClientService extends 
ControllerService, Verifiab
             .build();
 
     PropertyDescriptor VAULT_URI = new PropertyDescriptor.Builder()
+            .name("Vault URI")
             .name("vault.uri")
             .displayName("Vault URI")
             .description("The URI of the HashiCorp Vault server (e.g., 
http://localhost:8200).  Required if not specified in the " +
@@ -57,8 +58,7 @@ public interface HashiCorpVaultClientService extends 
ControllerService, Verifiab
             .build();
 
     PropertyDescriptor VAULT_AUTHENTICATION = new PropertyDescriptor.Builder()
-            .name("vault.authentication")
-            .displayName("Vault Authentication")
+            .name("Vault Authentication")
             .description("Vault authentication method, as described in the 
Spring Vault Environment Configuration documentation " +
                     
"(https://docs.spring.io/spring-vault/docs/2.3.x/reference/html/#vault.core.environment-vault-configuration).")
             .required(true)
@@ -68,8 +68,7 @@ public interface HashiCorpVaultClientService extends 
ControllerService, Verifiab
             .build();
 
     PropertyDescriptor SSL_CONTEXT_SERVICE = new PropertyDescriptor.Builder()
-            .name("vault.ssl.context.service")
-            .displayName("SSL Context Service")
+            .name("SSL Context Service")
             .description("The SSL Context Service used to provide client 
certificate information for TLS/SSL connections to the " +
                     "HashiCorp Vault server.")
             .required(false)
@@ -78,8 +77,7 @@ public interface HashiCorpVaultClientService extends 
ControllerService, Verifiab
             .build();
 
     PropertyDescriptor VAULT_PROPERTIES_FILES = new 
PropertyDescriptor.Builder()
-            .name("vault.properties.files")
-            .displayName("Vault Properties Files")
+            .name("Vault Properties Files")
             .description("A comma-separated list of files containing HashiCorp 
Vault configuration properties, as described in the Spring Vault " +
                     "Environment Configuration documentation 
(https://docs.spring.io/spring-vault/docs/2.3.x/reference/html/#vault.core.environment-vault-configuration).
 " +
                     "All of the Spring property keys and 
authentication-specific property keys are supported.")
@@ -89,8 +87,7 @@ public interface HashiCorpVaultClientService extends 
ControllerService, Verifiab
             .build();
 
     PropertyDescriptor CONNECTION_TIMEOUT = new PropertyDescriptor.Builder()
-            .name("vault.connection.timeout")
-            .displayName("Connection Timeout")
+            .name("Connection Timeout")
             .description("The connection timeout for the HashiCorp Vault 
client")
             .required(true)
             .defaultValue("5 sec")
@@ -98,8 +95,7 @@ public interface HashiCorpVaultClientService extends 
ControllerService, Verifiab
             .build();
 
     PropertyDescriptor READ_TIMEOUT = new PropertyDescriptor.Builder()
-            .name("vault.read.timeout")
-            .displayName("Read Timeout")
+            .name("Read Timeout")
             .description("The read timeout for the HashiCorp Vault client")
             .required(true)
             .defaultValue("15 sec")
@@ -112,4 +108,14 @@ public interface HashiCorpVaultClientService extends 
ControllerService, Verifiab
      */
     HashiCorpVaultCommunicationService getHashiCorpVaultCommunicationService();
 
+    @Override
+    default void migrateProperties(PropertyConfiguration config) {
+        config.renameProperty("configuration-strategy", 
CONFIGURATION_STRATEGY.getName());
+        config.renameProperty("vault.uri", VAULT_URI.getName());
+        config.renameProperty("vault.authentication", 
VAULT_AUTHENTICATION.getName());
+        config.renameProperty("vault.ssl.context.service", 
SSL_CONTEXT_SERVICE.getName());
+        config.renameProperty("vault.properties.files", 
VAULT_PROPERTIES_FILES.getName());
+        config.renameProperty("vault.connection.timeout", 
CONNECTION_TIMEOUT.getName());
+        config.renameProperty("vault.read.timeout", READ_TIMEOUT.getName());
+    }
 }
diff --git 
a/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cacheclient/HazelcastMapCacheClient.java
 
b/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cacheclient/HazelcastMapCacheClient.java
index 143f9a238e..ad069efa30 100644
--- 
a/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cacheclient/HazelcastMapCacheClient.java
+++ 
b/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cacheclient/HazelcastMapCacheClient.java
@@ -30,6 +30,7 @@ import org.apache.nifi.distributed.cache.client.Serializer;
 import org.apache.nifi.expression.ExpressionLanguageScope;
 import org.apache.nifi.hazelcast.services.cache.HazelcastCache;
 import org.apache.nifi.hazelcast.services.cachemanager.HazelcastCacheManager;
+import org.apache.nifi.migration.PropertyConfiguration;
 import org.apache.nifi.processor.util.StandardValidators;
 
 import java.io.ByteArrayOutputStream;
@@ -54,8 +55,7 @@ import java.util.concurrent.TimeUnit;
 public class HazelcastMapCacheClient extends AbstractControllerService 
implements AtomicDistributedMapCacheClient<Long> {
 
     public static final PropertyDescriptor HAZELCAST_CACHE_MANAGER = new 
PropertyDescriptor.Builder()
-            .name("hazelcast-cache-manager")
-            .displayName("Hazelcast Cache Manager")
+            .name("Hazelcast Cache Manager")
             .description("A Hazelcast Cache Manager which manages connections 
to Hazelcast and provides cache instances.")
             .identifiesControllerService(HazelcastCacheManager.class)
             .required(true)
@@ -63,8 +63,7 @@ public class HazelcastMapCacheClient extends 
AbstractControllerService implement
             .build();
 
     public static final PropertyDescriptor HAZELCAST_CACHE_NAME = new 
PropertyDescriptor.Builder()
-            .name("hazelcast-cache-name")
-            .displayName("Hazelcast Cache Name")
+            .name("Hazelcast Cache Name")
             .description("The name of a given cache. A Hazelcast cluster may 
handle multiple independent caches, each identified by a name." +
                     " Clients using caches with the same name are working on 
the same data structure within Hazelcast.")
             .required(true)
@@ -73,8 +72,7 @@ public class HazelcastMapCacheClient extends 
AbstractControllerService implement
             .build();
 
     public static final PropertyDescriptor HAZELCAST_ENTRY_TTL = new 
PropertyDescriptor.Builder()
-            .name("hazelcast-entry-ttl")
-            .displayName("Hazelcast Entry Lifetime")
+            .name("Hazelcast Entry Lifetime")
             .description("Indicates how long the written entries should exist 
in Hazelcast. Setting it to '0 secs' means that the data" +
                     "will exists until its deletion or until the Hazelcast 
server is shut down. Using `EmbeddedHazelcastCacheManager` as" +
                     "cache manager will not provide policies to limit the size 
of the cache.")
@@ -190,6 +188,13 @@ public class HazelcastMapCacheClient extends 
AbstractControllerService implement
         getLogger().debug("Closing {}", getClass().getSimpleName());
     }
 
+    @Override
+    public void migrateProperties(PropertyConfiguration config) {
+        config.renameProperty("hazelcast-cache-manager", 
HAZELCAST_CACHE_MANAGER.getName());
+        config.renameProperty("hazelcast-cache-name", 
HAZELCAST_CACHE_NAME.getName());
+        config.renameProperty("hazelcast-entry-ttl", 
HAZELCAST_ENTRY_TTL.getName());
+    }
+
     @Override
     protected List<PropertyDescriptor> getSupportedPropertyDescriptors() {
         return PROPERTY_DESCRIPTORS;
diff --git 
a/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cachemanager/EmbeddedHazelcastCacheManager.java
 
b/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cachemanager/EmbeddedHazelcastCacheManager.java
index 4578d44234..8a68f7cbec 100644
--- 
a/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cachemanager/EmbeddedHazelcastCacheManager.java
+++ 
b/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cachemanager/EmbeddedHazelcastCacheManager.java
@@ -30,12 +30,12 @@ import org.apache.nifi.components.ValidationResult;
 import org.apache.nifi.context.PropertyContext;
 import org.apache.nifi.controller.ConfigurationContext;
 import org.apache.nifi.expression.ExpressionLanguageScope;
+import org.apache.nifi.migration.PropertyConfiguration;
 import org.apache.nifi.processor.exception.ProcessException;
 import org.apache.nifi.processor.util.StandardValidators;
 
 import java.util.Arrays;
 import java.util.Collection;
-import java.util.Collections;
 import java.util.LinkedList;
 import java.util.List;
 import java.util.UUID;
@@ -63,8 +63,7 @@ public class EmbeddedHazelcastCacheManager extends 
IMapBasedHazelcastCacheManage
             " is allowed. NiFi nodes are not listed will be join to the 
Hazelcast cluster as clients.");
 
     private static final PropertyDescriptor HAZELCAST_PORT = new 
PropertyDescriptor.Builder()
-            .name("hazelcast-port")
-            .displayName("Hazelcast Port")
+            .name("Hazelcast Port")
             .description("Port for the Hazelcast instance to use.")
             .required(true)
             .defaultValue(String.valueOf(DEFAULT_HAZELCAST_PORT))
@@ -73,8 +72,7 @@ public class EmbeddedHazelcastCacheManager extends 
IMapBasedHazelcastCacheManage
             .build();
 
     private static final PropertyDescriptor HAZELCAST_CLUSTERING_STRATEGY = 
new PropertyDescriptor.Builder()
-            .name("hazelcast-clustering-strategy")
-            .displayName("Hazelcast Clustering Strategy")
+            .name("Hazelcast Clustering Strategy")
             .description("Specifies with what strategy the Hazelcast cluster 
should be created.")
             .required(true)
             .allowableValues(CLUSTER_NONE, CLUSTER_ALL_NODES, CLUSTER_EXPLICIT)
@@ -82,8 +80,7 @@ public class EmbeddedHazelcastCacheManager extends 
IMapBasedHazelcastCacheManage
             .build();
 
     private static final PropertyDescriptor HAZELCAST_INSTANCES = new 
PropertyDescriptor.Builder()
-            .name("hazelcast-instances")
-            .displayName("Hazelcast Instances")
+            .name("Hazelcast Instances")
             .description("Only used with \"Explicit\" Clustering Strategy!" +
                     " List of NiFi instance host names which should be part of 
the Hazelcast cluster. Host names are separated by comma." +
                     " The port specified in the \"Hazelcast Port\" property 
will be used as server port." +
@@ -94,15 +91,19 @@ public class EmbeddedHazelcastCacheManager extends 
IMapBasedHazelcastCacheManage
             .expressionLanguageSupported(ExpressionLanguageScope.ENVIRONMENT)
             .build();
 
-    private static final List<PropertyDescriptor> PROPERTY_DESCRIPTORS;
+    private static final List<PropertyDescriptor> PROPERTY_DESCRIPTORS = 
List.of(
+            HAZELCAST_CLUSTER_NAME,
+            HAZELCAST_PORT,
+            HAZELCAST_CLUSTERING_STRATEGY,
+            HAZELCAST_INSTANCES
+    );
 
-    static {
-        PROPERTY_DESCRIPTORS = Collections.unmodifiableList(Arrays.asList(
-                HAZELCAST_CLUSTER_NAME,
-                HAZELCAST_PORT,
-                HAZELCAST_CLUSTERING_STRATEGY,
-                HAZELCAST_INSTANCES
-        ));
+    @Override
+    public void migrateProperties(PropertyConfiguration config) {
+        super.migrateProperties(config);
+        config.renameProperty("hazelcast-port", HAZELCAST_PORT.getName());
+        config.renameProperty("hazelcast-clustering-strategy", 
HAZELCAST_CLUSTERING_STRATEGY.getName());
+        config.renameProperty("hazelcast-instances", 
HAZELCAST_INSTANCES.getName());
     }
 
     @Override
diff --git 
a/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cachemanager/ExternalHazelcastCacheManager.java
 
b/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cachemanager/ExternalHazelcastCacheManager.java
index 19ceeeefeb..6c52988b2f 100644
--- 
a/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cachemanager/ExternalHazelcastCacheManager.java
+++ 
b/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cachemanager/ExternalHazelcastCacheManager.java
@@ -24,12 +24,11 @@ import org.apache.nifi.components.ValidationContext;
 import org.apache.nifi.components.ValidationResult;
 import org.apache.nifi.controller.ConfigurationContext;
 import org.apache.nifi.expression.ExpressionLanguageScope;
+import org.apache.nifi.migration.PropertyConfiguration;
 import org.apache.nifi.processor.util.StandardValidators;
 
-import java.util.ArrayList;
 import java.util.Arrays;
 import java.util.Collection;
-import java.util.Collections;
 import java.util.HashSet;
 import java.util.List;
 import java.util.Set;
@@ -40,8 +39,7 @@ import java.util.concurrent.TimeUnit;
 public class ExternalHazelcastCacheManager extends 
IMapBasedHazelcastCacheManager {
 
     public static final PropertyDescriptor HAZELCAST_SERVER_ADDRESS = new 
PropertyDescriptor.Builder()
-            .name("hazelcast-server-address")
-            .displayName("Hazelcast Server Address")
+            .name("Hazelcast Server Address")
             .description("Addresses of one or more the Hazelcast instances, 
using {host:port} format, separated by comma.")
             .required(true)
             .addValidator(StandardValidators.HOSTNAME_PORT_LIST_VALIDATOR)
@@ -49,8 +47,7 @@ public class ExternalHazelcastCacheManager extends 
IMapBasedHazelcastCacheManage
             .build();
 
     public static final PropertyDescriptor HAZELCAST_RETRY_BACKOFF_INITIAL = 
new PropertyDescriptor.Builder()
-            .name("hazelcast-retry-backoff-initial")
-            .displayName("Hazelcast Initial Backoff")
+            .name("Hazelcast Initial Backoff")
             .description("The amount of time the client waits before it tries 
to reestablish connection for the first time.")
             .addValidator(StandardValidators.TIME_PERIOD_VALIDATOR)
             .required(true)
@@ -58,8 +55,7 @@ public class ExternalHazelcastCacheManager extends 
IMapBasedHazelcastCacheManage
             .build();
 
     public static final PropertyDescriptor HAZELCAST_RETRY_BACKOFF_MAXIMUM = 
new PropertyDescriptor.Builder()
-            .name("hazelcast-retry-backoff-maximum")
-            .displayName("Hazelcast Maximum Backoff")
+            .name("Hazelcast Maximum Backoff")
             .description("The maximum amount of time the client waits before 
it tries to reestablish connection.")
             .addValidator(StandardValidators.TIME_PERIOD_VALIDATOR)
             .required(true)
@@ -67,8 +63,7 @@ public class ExternalHazelcastCacheManager extends 
IMapBasedHazelcastCacheManage
             .build();
 
     public static final PropertyDescriptor HAZELCAST_RETRY_BACKOFF_MULTIPLIER 
= new PropertyDescriptor.Builder()
-            .name("hazelcast-retry-backoff-multiplier")
-            .displayName("Hazelcast Backoff Multiplier")
+            .name("Hazelcast Backoff Multiplier")
             .description("A multiplier by which the wait time is increased 
before each attempt to reestablish connection.")
             .addValidator(StandardValidators.NUMBER_VALIDATOR)
             .required(true)
@@ -76,25 +71,30 @@ public class ExternalHazelcastCacheManager extends 
IMapBasedHazelcastCacheManage
             .build();
 
     public static final PropertyDescriptor HAZELCAST_CONNECTION_TIMEOUT = new 
PropertyDescriptor.Builder()
-            .name("hazelcast-connection-timeout")
-            .displayName("Hazelcast Connection Timeout")
+            .name("Hazelcast Connection Timeout")
             .description("The maximum amount of time the client tries to 
connect or reconnect before giving up.")
             .addValidator(StandardValidators.TIME_PERIOD_VALIDATOR)
             .required(true)
             .defaultValue(DEFAULT_CLIENT_TIMEOUT_MAXIMUM_IN_SEC + " secs")
             .build();
 
-    private static final List<PropertyDescriptor> PROPERTY_DESCRIPTORS;
-
-    static {
-        final List<PropertyDescriptor> properties = new ArrayList<>();
-        properties.add(HAZELCAST_CLUSTER_NAME);
-        properties.add(HAZELCAST_SERVER_ADDRESS);
-        properties.add(HAZELCAST_RETRY_BACKOFF_INITIAL);
-        properties.add(HAZELCAST_RETRY_BACKOFF_MAXIMUM);
-        properties.add(HAZELCAST_RETRY_BACKOFF_MULTIPLIER);
-        properties.add(HAZELCAST_CONNECTION_TIMEOUT);
-        PROPERTY_DESCRIPTORS = Collections.unmodifiableList(properties);
+    private static final List<PropertyDescriptor> PROPERTY_DESCRIPTORS = 
List.of(
+        HAZELCAST_CLUSTER_NAME,
+        HAZELCAST_SERVER_ADDRESS,
+        HAZELCAST_RETRY_BACKOFF_INITIAL,
+        HAZELCAST_RETRY_BACKOFF_MAXIMUM,
+        HAZELCAST_RETRY_BACKOFF_MULTIPLIER,
+        HAZELCAST_CONNECTION_TIMEOUT
+    );
+
+    @Override
+    public void migrateProperties(PropertyConfiguration config) {
+        super.migrateProperties(config);
+        config.renameProperty("hazelcast-server-address", 
HAZELCAST_SERVER_ADDRESS.getName());
+        config.renameProperty("hazelcast-retry-backoff-initial", 
HAZELCAST_RETRY_BACKOFF_INITIAL.getName());
+        config.renameProperty("hazelcast-retry-backoff-maximum", 
HAZELCAST_RETRY_BACKOFF_MAXIMUM.getName());
+        config.renameProperty("hazelcast-retry-backoff-multiplier", 
HAZELCAST_RETRY_BACKOFF_MULTIPLIER.getName());
+        config.renameProperty("hazelcast-connection-timeout", 
HAZELCAST_CONNECTION_TIMEOUT.getName());
     }
 
     @Override
diff --git 
a/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cachemanager/IMapBasedHazelcastCacheManager.java
 
b/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cachemanager/IMapBasedHazelcastCacheManager.java
index fc22c020ef..7800cd2480 100644
--- 
a/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cachemanager/IMapBasedHazelcastCacheManager.java
+++ 
b/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/main/java/org/apache/nifi/hazelcast/services/cachemanager/IMapBasedHazelcastCacheManager.java
@@ -28,6 +28,7 @@ import org.apache.nifi.controller.ConfigurationContext;
 import org.apache.nifi.expression.ExpressionLanguageScope;
 import org.apache.nifi.hazelcast.services.cache.HazelcastCache;
 import org.apache.nifi.hazelcast.services.cache.IMapBasedHazelcastCache;
+import org.apache.nifi.migration.PropertyConfiguration;
 import org.apache.nifi.processor.util.StandardValidators;
 import org.apache.nifi.reporting.InitializationException;
 
@@ -47,8 +48,7 @@ abstract class IMapBasedHazelcastCacheManager extends 
AbstractControllerService
     protected static final double DEFAULT_CLIENT_BACKOFF_MULTIPLIER = 1.5;
 
     public static final PropertyDescriptor HAZELCAST_CLUSTER_NAME = new 
PropertyDescriptor.Builder()
-            .name("hazelcast-cluster-name")
-            .displayName("Hazelcast Cluster Name")
+            .name("Hazelcast Cluster Name")
             .description("Name of the Hazelcast cluster.")
             .defaultValue("nifi") // Hazelcast's default is "dev", "nifi" 
overwrites this.
             .required(true)
@@ -88,6 +88,11 @@ abstract class IMapBasedHazelcastCacheManager extends 
AbstractControllerService
         }
     }
 
+    @Override
+    public void migrateProperties(PropertyConfiguration config) {
+        config.renameProperty("hazelcast-cluster-name", 
HAZELCAST_CLUSTER_NAME.getName());
+    }
+
     protected HazelcastInstance getClientInstance(
             final String clusterName,
             final List<String> serverAddresses,
diff --git 
a/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/test/java/org/apache/nifi/hazelcast/services/cachemanager/TestHazelcastProcessor.java
 
b/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/test/java/org/apache/nifi/hazelcast/services/cachemanager/TestHazelcastProcessor.java
index 76a1474fa9..5797ea99a0 100644
--- 
a/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/test/java/org/apache/nifi/hazelcast/services/cachemanager/TestHazelcastProcessor.java
+++ 
b/nifi-extension-bundles/nifi-hazelcast-bundle/nifi-hazelcast-services/src/test/java/org/apache/nifi/hazelcast/services/cachemanager/TestHazelcastProcessor.java
@@ -48,8 +48,7 @@ class TestHazelcastProcessor extends AbstractProcessor {
     private static final DummyStringSerializer SERIALIZER = new 
DummyStringSerializer();
 
     public static final PropertyDescriptor TEST_HAZELCAST_MAP_CACHE_CLIENT = 
new PropertyDescriptor.Builder()
-            .name("test-hazelcast-map-cache-client")
-            .displayName("Test Hazelcast Map Cache Client")
+            .name("Test Hazelcast Map Cache Client")
             .identifiesControllerService(HazelcastMapCacheClient.class)
             .addValidator(StandardValidators.NON_EMPTY_VALIDATOR)
             .required(true)
diff --git 
a/nifi-extension-bundles/nifi-hl7-bundle/nifi-hl7-processors/src/main/java/org/apache/nifi/processors/hl7/ExtractHL7Attributes.java
 
b/nifi-extension-bundles/nifi-hl7-bundle/nifi-hl7-processors/src/main/java/org/apache/nifi/processors/hl7/ExtractHL7Attributes.java
index 17039ba5ea..1b52a831fa 100644
--- 
a/nifi-extension-bundles/nifi-hl7-bundle/nifi-hl7-processors/src/main/java/org/apache/nifi/processors/hl7/ExtractHL7Attributes.java
+++ 
b/nifi-extension-bundles/nifi-hl7-bundle/nifi-hl7-processors/src/main/java/org/apache/nifi/processors/hl7/ExtractHL7Attributes.java
@@ -55,6 +55,7 @@ import org.apache.nifi.annotation.documentation.Tags;
 import org.apache.nifi.components.PropertyDescriptor;
 import org.apache.nifi.expression.ExpressionLanguageScope;
 import org.apache.nifi.flowfile.FlowFile;
+import org.apache.nifi.migration.PropertyConfiguration;
 import org.apache.nifi.processor.AbstractProcessor;
 import org.apache.nifi.processor.ProcessContext;
 import org.apache.nifi.processor.ProcessSession;
@@ -87,8 +88,7 @@ public class ExtractHL7Attributes extends AbstractProcessor {
             .build();
 
     public static final PropertyDescriptor USE_SEGMENT_NAMES = new 
PropertyDescriptor.Builder()
-            .name("use-segment-names")
-            .displayName("Use Segment Names")
+            .name("Use Segment Names")
             .description("Whether or not to use HL7 segment names in 
attributes")
             .required(true)
             .allowableValues("true", "false")
@@ -97,8 +97,7 @@ public class ExtractHL7Attributes extends AbstractProcessor {
             .build();
 
     public static final PropertyDescriptor PARSE_SEGMENT_FIELDS = new 
PropertyDescriptor.Builder()
-            .name("parse-segment-fields")
-            .displayName("Parse Segment Fields")
+            .name("Parse Segment Fields")
             .description("Whether or not to parse HL7 segment fields into 
attributes")
             .required(true)
             .allowableValues("true", "false")
@@ -107,8 +106,7 @@ public class ExtractHL7Attributes extends AbstractProcessor 
{
             .build();
 
     public static final PropertyDescriptor SKIP_VALIDATION = new 
PropertyDescriptor.Builder()
-            .name("skip-validation")
-            .displayName("Skip Validation")
+            .name("Skip Validation")
             .description("Whether or not to validate HL7 message values")
             .required(true)
             .allowableValues("true", "false")
@@ -117,8 +115,7 @@ public class ExtractHL7Attributes extends AbstractProcessor 
{
             .build();
 
     public static final PropertyDescriptor HL7_INPUT_VERSION = new 
PropertyDescriptor.Builder()
-            .name("hl7-input-version")
-            .displayName("HL7 Input Version")
+            .name("HL7 Input Version")
             .description("The HL7 version to use for parsing and validation")
             .required(true)
             .allowableValues("autodetect", "2.2", "2.3", "2.3.1", "2.4", 
"2.5", "2.5.1", "2.6")
@@ -200,6 +197,14 @@ public class ExtractHL7Attributes extends 
AbstractProcessor {
         session.transfer(flowFile, REL_SUCCESS);
     }
 
+    @Override
+    public void migrateProperties(PropertyConfiguration config) {
+        config.renameProperty("use-segment-names", 
USE_SEGMENT_NAMES.getName());
+        config.renameProperty("parse-segment-fields", 
PARSE_SEGMENT_FIELDS.getName());
+        config.renameProperty("skip-validation", SKIP_VALIDATION.getName());
+        config.renameProperty("hl7-input-version", 
HL7_INPUT_VERSION.getName());
+    }
+
     public static Map<String, String> getAttributes(final Group group, final 
boolean useNames, final boolean parseFields) throws HL7Exception {
         final Map<String, String> attributes = new TreeMap<>();
         if (!isEmpty(group)) {
diff --git 
a/nifi-extension-bundles/nifi-hubspot-bundle/nifi-hubspot-processors/src/main/java/org/apache/nifi/processors/hubspot/GetHubSpot.java
 
b/nifi-extension-bundles/nifi-hubspot-bundle/nifi-hubspot-processors/src/main/java/org/apache/nifi/processors/hubspot/GetHubSpot.java
index 1c350c4f83..6cad840493 100644
--- 
a/nifi-extension-bundles/nifi-hubspot-bundle/nifi-hubspot-processors/src/main/java/org/apache/nifi/processors/hubspot/GetHubSpot.java
+++ 
b/nifi-extension-bundles/nifi-hubspot-bundle/nifi-hubspot-processors/src/main/java/org/apache/nifi/processors/hubspot/GetHubSpot.java
@@ -42,6 +42,7 @@ import org.apache.nifi.components.state.StateMap;
 import org.apache.nifi.expression.ExpressionLanguageScope;
 import org.apache.nifi.flowfile.FlowFile;
 import org.apache.nifi.flowfile.attributes.CoreAttributes;
+import org.apache.nifi.migration.PropertyConfiguration;
 import org.apache.nifi.processor.AbstractProcessor;
 import org.apache.nifi.processor.ProcessContext;
 import org.apache.nifi.processor.ProcessSession;
@@ -87,16 +88,14 @@ import java.util.stream.Collectors;
 public class GetHubSpot extends AbstractProcessor {
 
     static final PropertyDescriptor OBJECT_TYPE = new 
PropertyDescriptor.Builder()
-            .name("object-type")
-            .displayName("Object Type")
+            .name("Object Type")
             .description("The HubSpot Object Type requested")
             .required(true)
             .allowableValues(HubSpotObjectType.class)
             .build();
 
     static final PropertyDescriptor ACCESS_TOKEN = new 
PropertyDescriptor.Builder()
-            .name("access-token")
-            .displayName("Access Token")
+            .name("Access Token")
             .description("Access Token to authenticate requests")
             .required(true)
             .sensitive(true)
@@ -105,8 +104,7 @@ public class GetHubSpot extends AbstractProcessor {
             .build();
 
     static final PropertyDescriptor RESULT_LIMIT = new 
PropertyDescriptor.Builder()
-            .name("result-limit")
-            .displayName("Result Limit")
+            .name("Result Limit")
             .description("The maximum number of results to request for each 
invocation of the Processor")
             .expressionLanguageSupported(ExpressionLanguageScope.ENVIRONMENT)
             .required(false)
@@ -114,8 +112,7 @@ public class GetHubSpot extends AbstractProcessor {
             .build();
 
     static final PropertyDescriptor IS_INCREMENTAL = new 
PropertyDescriptor.Builder()
-            .name("is-incremental")
-            .displayName("Incremental Loading")
+            .name("Incremental Loading")
             .description("The processor can incrementally load the queried 
objects so that each object is queried exactly once." +
                     " For each query, the processor queries objects within a 
time window where the objects were modified between" +
                     " the previous run time and the current time (optionally 
adjusted by the Incremental Delay property).")
@@ -125,8 +122,7 @@ public class GetHubSpot extends AbstractProcessor {
             .build();
 
     static final PropertyDescriptor INCREMENTAL_DELAY = new 
PropertyDescriptor.Builder()
-            .name("incremental-delay")
-            .displayName("Incremental Delay")
+            .name("Incremental Delay")
             .description(("The ending timestamp of the time window will be 
adjusted earlier by the amount configured in this property." +
                     " For example, with a property value of 10 seconds, an 
ending timestamp of 12:30:45 would be changed to 12:30:35." +
                     " Set this property to avoid missing objects when the 
clock of your local machines and HubSpot servers' clock are not in sync" +
@@ -139,8 +135,7 @@ public class GetHubSpot extends AbstractProcessor {
             .build();
 
     static final PropertyDescriptor INCREMENTAL_INITIAL_START_TIME = new 
PropertyDescriptor.Builder()
-            .name("incremental-initial-start-time")
-            .displayName("Incremental Initial Start Time")
+            .name("Incremental Initial Start Time")
             .description("This property specifies the start time that the 
processor applies when running the first request." +
                     " The expected format is a UTC date-time such as 
'2011-12-03T10:15:30Z'")
             .required(false)
@@ -150,8 +145,7 @@ public class GetHubSpot extends AbstractProcessor {
             .build();
 
     static final PropertyDescriptor WEB_CLIENT_SERVICE_PROVIDER = new 
PropertyDescriptor.Builder()
-            .name("web-client-service-provider")
-            .displayName("Web Client Service Provider")
+            .name("Web Client Service Provider")
             .description("Controller service for HTTP client operations")
             .identifiesControllerService(WebClientServiceProvider.class)
             .required(true)
@@ -256,6 +250,17 @@ public class GetHubSpot extends AbstractProcessor {
         }
     }
 
+    @Override
+    public void migrateProperties(PropertyConfiguration config) {
+        config.renameProperty("object-type", OBJECT_TYPE.getName());
+        config.renameProperty("access-token", ACCESS_TOKEN.getName());
+        config.renameProperty("result-limit", RESULT_LIMIT.getName());
+        config.renameProperty("is-incremental", IS_INCREMENTAL.getName());
+        config.renameProperty("incremental-delay", 
INCREMENTAL_DELAY.getName());
+        config.renameProperty("incremental-initial-start-time", 
INCREMENTAL_INITIAL_START_TIME.getName());
+        config.renameProperty("web-client-service-provider", 
WEB_CLIENT_SERVICE_PROVIDER.getName());
+    }
+
     private String getResponseBodyAsString(final ProcessContext context, final 
HttpResponseEntity response, final URI uri) {
         try {
             return IOUtils.toString(response.body(), StandardCharsets.UTF_8);

Reply via email to