This is an automated email from the ASF dual-hosted git repository.

zhangyue19921010 pushed a commit to branch HUDI-8990
in repository https://gitbox.apache.org/repos/asf/hudi.git

commit 60de489d9a925d4f681894fc310081ffc3f27555
Author: YueZhang <[email protected]>
AuthorDate: Mon Mar 17 20:04:40 2025 +0800

    finish hashing_config initial && related UT
---
 .../java/org/apache/hudi/config/HoodieIndexConfig.java     |  6 +++---
 .../apache/hudi/index/bucket/SimpleBucketIndexUtils.java   |  2 +-
 .../common/model/PartitionBucketIndexHashingConfig.java    |  2 --
 .../org/apache/hudi/table/catalog/TestHoodieCatalog.java   | 14 +++++++-------
 4 files changed, 11 insertions(+), 13 deletions(-)

diff --git 
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/config/HoodieIndexConfig.java
 
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/config/HoodieIndexConfig.java
index 626e894fd1a..bfad563a9e7 100644
--- 
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/config/HoodieIndexConfig.java
+++ 
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/config/HoodieIndexConfig.java
@@ -297,9 +297,9 @@ public class HoodieIndexConfig extends HoodieConfig {
       .key("hoodie.bucket.index.partition.expressions")
       .noDefaultValue()
       .markAdvanced()
-      .withDocumentation("Users can use this parameter to specify expression 
and the corresponding bucket " +
-          "numbers (separated by commas).Multiple rules are separated by 
semicolons like " +
-          
"hoodie.bucket.index.partition.expressions=expression1,bucket-number1;expression2,bucket-number2");
+      .withDocumentation("Users can use this parameter to specify expression 
and the corresponding bucket "
+          + "numbers (separated by commas).Multiple rules are separated by 
semicolons like "
+          + 
"hoodie.bucket.index.partition.expressions=expression1,bucket-number1;expression2,bucket-number2");
 
   public static final ConfigProperty<String> BUCKET_INDEX_MAX_NUM_BUCKETS = 
ConfigProperty
       .key("hoodie.bucket.index.max.num.buckets")
diff --git 
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/index/bucket/SimpleBucketIndexUtils.java
 
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/index/bucket/SimpleBucketIndexUtils.java
index 74c8022c9aa..f3a20c6a002 100644
--- 
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/index/bucket/SimpleBucketIndexUtils.java
+++ 
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/index/bucket/SimpleBucketIndexUtils.java
@@ -46,7 +46,7 @@ public class SimpleBucketIndexUtils {
                                               String expressions,
                                               String rule,
                                               int defaultBucketNumber,
-                                              String instant){
+                                              String instant) {
     if (StringUtils.isNullOrEmpty(expressions)) {
       return false;
     }
diff --git 
a/hudi-common/src/main/java/org/apache/hudi/common/model/PartitionBucketIndexHashingConfig.java
 
b/hudi-common/src/main/java/org/apache/hudi/common/model/PartitionBucketIndexHashingConfig.java
index 225629b52ff..46b38896925 100644
--- 
a/hudi-common/src/main/java/org/apache/hudi/common/model/PartitionBucketIndexHashingConfig.java
+++ 
b/hudi-common/src/main/java/org/apache/hudi/common/model/PartitionBucketIndexHashingConfig.java
@@ -38,7 +38,6 @@ public class PartitionBucketIndexHashingConfig implements 
Serializable {
   private final String expressions;
   private final int defaultBucketNumber;
   private final String rule;
-
   private final int version;
   private final String instant;
 
@@ -71,7 +70,6 @@ public class PartitionBucketIndexHashingConfig implements 
Serializable {
     return JsonUtils.getObjectMapper().readValue(jsonStr, clazz);
   }
 
-
   public static PartitionBucketIndexHashingConfig fromBytes(byte[] bytes) 
throws IOException {
     try {
       return fromJsonString(new String(bytes, StandardCharsets.UTF_8), 
PartitionBucketIndexHashingConfig.class);
diff --git 
a/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/table/catalog/TestHoodieCatalog.java
 
b/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/table/catalog/TestHoodieCatalog.java
index 1e518c5ef45..84e982df8f3 100644
--- 
a/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/table/catalog/TestHoodieCatalog.java
+++ 
b/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/table/catalog/TestHoodieCatalog.java
@@ -360,14 +360,14 @@ public class TestHoodieCatalog {
         .createMetaClient(
             new 
HadoopStorageConfiguration(HadoopConfigurations.getHadoopConf(flinkConf)), 
tablePathStr);
     HoodieStorage storage = metaClient.getStorage();
-    StoragePath initialHashing_config =
+    StoragePath initialHashingConfig =
         new StoragePath(metaClient.getHashingMetadataConfigPath(), 
SimpleBucketIndexUtils.INITIAL_HASHING_CONFIG_INSTANT + 
PartitionBucketIndexHashingConfig.HASHING_CONFIG_FILE_SUFFIX);
-    StoragePathInfo info = storage.getPathInfo(initialHashing_config);
-    Option<PartitionBucketIndexHashingConfig> hashing_config = 
SimpleBucketIndexUtils.loadHashingConfig(storage, info);
-    assertTrue(hashing_config.isPresent());
-    assertEquals(hashing_config.get().getDefaultBucketNumber(), 
Integer.parseInt(defaultBucketNumber));
-    assertEquals(hashing_config.get().getRule(), rule);
-    assertEquals(hashing_config.get().getExpressions(), expressions);
+    StoragePathInfo info = storage.getPathInfo(initialHashingConfig);
+    Option<PartitionBucketIndexHashingConfig> hashingConfig = 
SimpleBucketIndexUtils.loadHashingConfig(storage, info);
+    assertTrue(hashingConfig.isPresent());
+    assertEquals(hashingConfig.get().getDefaultBucketNumber(), 
Integer.parseInt(defaultBucketNumber));
+    assertEquals(hashingConfig.get().getRule(), rule);
+    assertEquals(hashingConfig.get().getExpressions(), expressions);
   }
 
   @Test

Reply via email to